Search results for: mobile mapping
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2747

Search results for: mobile mapping

797 Photoleap: An AI-Powered Photo Editing App with Advanced Features and User Satisfaction Analysis

Authors: Joud Basyouni, Rama Zagzoog, Mashael Al Faleh, Jana Alireza

Abstract:

AI is changing many fields and speeding up tasks that used to take a long time. It used to take too long to edit photos. However, many AI-powered apps make photo editing, automatic effects, and animations much easier than other manual editing apps with no AI. The mobile app Photoleap edits photos and creates digital art using AI. Editing photos with text prompts is also becoming a standard these days with the help of apps like Photoleap. Now, users can change backgrounds, add animations, turn text into images, and create scenes with AI. This project report discusses the photo editing app's history and popularity. Photoleap resembles Photoshop, Canva, Photos, and Pixlr. The report includes survey questions to assess Photoleap user satisfaction. The report describes Photoleap's features and functions with screenshots. Photoleap uses AI well. Charts and graphs show Photoleap user ratings and comments from the survey. This project found that most Photoleap users liked how well it worked, was made, and was easy to use. People liked changing photos and adding backgrounds. Users can create stunning photo animations. A few users dislike the app's animations, AI art, and photo effects. The project report discusses the app's pros and cons and offers improvements.

Keywords: artificial intelligence, photoleap, images, background, photo editing

Procedia PDF Downloads 61
796 Ignition Interlock Device for Motorcycles

Authors: Luisito L. Lacatan, Zacha Valerie G. Ancheta, Michelangelo A. Dorado, Lester Joseph M. Ochoa, Anthony Mark G. Tayabas

Abstract:

Ignition Interlock Device or IID is a mechanism installed inside a vehicle which requires the driver to breathe into the device before starting the vehicle. If the IID detects that the alcohol level or blood alcohol content (BAC) is higher than the accepted value, the engine will not start. If the driver is not able to provide a clean breath sample, the IID will log the event, warn the driver, and then start up an alarm. The purpose of the IID is to prevent accidents due to driving under the influence (DUI). With the rise of the two-wheeled vehicle in the Philippines due to its mobility and purchasing power, IIDs are still mainly installed on four-wheeled vehicles. Even though riding the motorcycle when drunk is more dangerous, there are only a small number of installed devices on motorcycles and scooters. The general objective of this study was to develop a system with hardware and software components that would implement IID on motorcycles. The study employed a descriptive method of research. The study also concluded the following: the infrared must have a point-to-point communication, the breathalyzer on the helmet should react to ethanol, the microcontroller on the motorcycle should accept all IR signals from the helmet and interpret it and the GPS shield should have an unobstructed line-of-sight communication with the GPS satellites.

Keywords: blood alcohol content, breathalyser, driving under the influence, global positioning system, global system for mobile communication

Procedia PDF Downloads 326
795 The Biology of Persister Cells and Antibiotic Resistance

Authors: Zikora K. G. Anyaegbunam, Annabel A. Nnawuihe, Ngozi J. Anyaegbunam, Emmanuel A. Eze

Abstract:

The discovery and production of new antibiotics is unavoidable in the fight against drug-resistant bacteria. However, this is only part of the problem; we have never really had medications that could completely eradicate an infection. All pathogens create a limited number of dormant persister cells that are resistant to antibiotic treatment. When the concentration of antibiotics decreases, surviving persisters repopulate the population, resulting in a recurrent chronic infection. Bacterial populations have an alternative survival strategy to withstand harsh conditions or antibiotic exposure, in addition to the well-known methods of antibiotic resistance and biofilm formation. Persister cells are a limited subset of transiently antibiotic-tolerant phenotypic variations capable of surviving high-dose antibiotic therapy. Persisters that flip back to a normal phenotype can restart growth when antibiotic pressure drops, assuring the bacterial population's survival. Persister cells have been found in every major pathogen, and they play a role in antibiotic tolerance in biofilms as well as the recalcitrance of chronic infections. Persister cells has been implicated to play a role in the establishment of antibiotic resistance, according to growing research. Thusthe need to basically elucidate the biology of persisters and how they are linked to antibiotic resistance, and as well it's link to diseases.

Keywords: persister cells, phenotypic variations, repopulation, mobile genetic transfers, antibiotic resistance

Procedia PDF Downloads 209
794 A Methodology to Integrate Data in the Company Based on the Semantic Standard in the Context of Industry 4.0

Authors: Chang Qin, Daham Mustafa, Abderrahmane Khiat, Pierre Bienert, Paulo Zanini

Abstract:

Nowadays, companies are facing lots of challenges in the process of digital transformation, which can be a complex and costly undertaking. Digital transformation involves the collection and analysis of large amounts of data, which can create challenges around data management and governance. Furthermore, it is also challenged to integrate data from multiple systems and technologies. Although with these pains, companies are still pursuing digitalization because by embracing advanced technologies, companies can improve efficiency, quality, decision-making, and customer experience while also creating different business models and revenue streams. In this paper, the issue that data is stored in data silos with different schema and structures is focused. The conventional approaches to addressing this issue involve utilizing data warehousing, data integration tools, data standardization, and business intelligence tools. However, these approaches primarily focus on the grammar and structure of the data and neglect the importance of semantic modeling and semantic standardization, which are essential for achieving data interoperability. In this session, the challenge of data silos in Industry 4.0 is addressed by developing a semantic modeling approach compliant with Asset Administration Shell (AAS) models as an efficient standard for communication in Industry 4.0. The paper highlights how our approach can facilitate the data mapping process and semantic lifting according to existing industry standards such as ECLASS and other industrial dictionaries. It also incorporates the Asset Administration Shell technology to model and map the company’s data and utilize a knowledge graph for data storage and exploration.

Keywords: data interoperability in industry 4.0, digital integration, industrial dictionary, semantic modeling

Procedia PDF Downloads 94
793 Graph Neural Network-Based Classification for Disease Prediction in Health Care Heterogeneous Data Structures of Electronic Health Record

Authors: Raghavi C. Janaswamy

Abstract:

In the healthcare sector, heterogenous data elements such as patients, diagnosis, symptoms, conditions, observation text from physician notes, and prescriptions form the essentials of the Electronic Health Record (EHR). The data in the form of clear text and images are stored or processed in a relational format in most systems. However, the intrinsic structure restrictions and complex joins of relational databases limit the widespread utility. In this regard, the design and development of realistic mapping and deep connections as real-time objects offer unparallel advantages. Herein, a graph neural network-based classification of EHR data has been developed. The patient conditions have been predicted as a node classification task using a graph-based open source EHR data, Synthea Database, stored in Tigergraph. The Synthea DB dataset is leveraged due to its closer representation of the real-time data and being voluminous. The graph model is built from the EHR heterogeneous data using python modules, namely, pyTigerGraph to get nodes and edges from the Tigergraph database, PyTorch to tensorize the nodes and edges, PyTorch-Geometric (PyG) to train the Graph Neural Network (GNN) and adopt the self-supervised learning techniques with the AutoEncoders to generate the node embeddings and eventually perform the node classifications using the node embeddings. The model predicts patient conditions ranging from common to rare situations. The outcome is deemed to open up opportunities for data querying toward better predictions and accuracy.

Keywords: electronic health record, graph neural network, heterogeneous data, prediction

Procedia PDF Downloads 86
792 Development and Validation of a HPLC Method for 6-Gingerol and 6-Shogaol in Joint Pain Relief Gel Containing Ginger (Zingiber officinale)

Authors: Tanwarat Kajsongkram, Saowalux Rotamporn, Sirinat Limbunruang, Sirinan Thubthimthed.

Abstract:

High-Performance Liquid Chromatography (HPLC) method was developed and validated for simultaneous estimation of 6-Gingerol(6G) and 6-Shogaol(6S) in joint pain relief gel containing ginger extract. The chromatographic separation was achieved by using C18 column, 150 x 4.6mm i.d., 5μ Luna, mobile phase containing acetonitrile and water (gradient elution). The flow rate was 1.0 ml/min and the absorbance was monitored at 282 nm. The proposed method was validated in terms of the analytical parameters such as specificity, accuracy, precision, linearity, range, limit of detection (LOD), limit of quantification (LOQ), and determined based on the International Conference on Harmonization (ICH) guidelines. The linearity ranges of 6G and 6S were obtained over 20-60 and 6-18 µg/ml respectively. Good linearity was observed over the above-mentioned range with linear regression equation Y= 11016x- 23778 for 6G and Y = 19276x-19604 for 6S (x is concentration of analytes in μg/ml and Y is peak area). The value of correlation coefficient was found to be 0.9994 for both markers. The limit of detection (LOD) and limit of quantification (LOQ) for 6G were 0.8567 and 2.8555 µg/ml and for 6S were 0.3672 and 1.2238 µg/ml respectively. The recovery range for 6G and 6S were found to be 91.57 to 102.36 % and 84.73 to 92.85 % for all three spiked levels. The RSD values from repeated extractions for 6G and 6S were 3.43 and 3.09% respectively. The validation of developed method on precision, accuracy, specificity, linearity, and range were also performed with well-accepted results.

Keywords: ginger, 6-gingerol, HPLC, 6-shogaol

Procedia PDF Downloads 443
791 MRCP as a Pre-Operative Tool for Predicting Variant Biliary Anatomy in Living Related Liver Donors

Authors: Awais Ahmed, Atif Rana, Haseeb Zia, Maham Jahangir, Rashed Nazir, Faisal Dar

Abstract:

Purpose: Biliary complications represent the most common cause of morbidity in living related liver donor transplantation and detailed preoperative evaluation of biliary anatomic variants is crucial for safe patient selection and improved surgical outcomes. Purpose of this study is to determine the accuracy of preoperative MRCP in predicting biliary variations when compared to intraoperative cholangiography in living related liver donors. Materials and Methods: From 44 potential donors, 40 consecutive living related liver donors (13 females and 28 males) underwent donor hepatectomy at our centre from April 2012 to August 2013. MRCP and IOC of all patients were retrospectively reviewed separately by two radiologists and a transplant surgeon.MRCP was performed on 1.5 Tesla MR magnets using breath-hold heavily T2 weighted radial slab technique. One patient was excluded due to suboptimal MRCP. The accuracy of MRCP for variant biliary anatomy was calculated. Results: MRCP accurately predicted the biliary anatomy in 38 of 39 cases (97 %). Standard biliary anatomy was predicted by MRCP in 25 (64 %) donors (100% sensitivity). Variant biliary anatomy was noted in 14 (36 %) IOCs of which MRCP predicted precise anatomy of 13 variants (93 % sensitivity). The two most common variations were drainage of the RPSD into the LHD (50%) and the triple confluence of the RASD, RPSD and LHD (21%). Conclusion: MRCP is a sensitive imaging tool for precise pre-operative mapping of biliary variations which is critical to surgical decision making in living related liver transplantation.

Keywords: intraoperative cholangiogram, liver transplantation, living related donors, magnetic resonance cholangio-pancreaticogram (MRCP)

Procedia PDF Downloads 397
790 Electrochemical APEX for Genotyping MYH7 Gene: A Low Cost Strategy for Minisequencing of Disease Causing Mutations

Authors: Ahmed M. Debela, Mayreli Ortiz , Ciara K. O´Sullivan

Abstract:

The completion of the human genome Project (HGP) has paved the way for mapping the diversity in the overall genome sequence which helps to understand the genetic causes of inherited diseases and susceptibility to drugs or environmental toxins. Arrayed primer extension (APEX) is a microarray based minisequencing strategy for screening disease causing mutations. It is derived from Sanger DNA sequencing and uses fluorescently dideoxynucleotides (ddNTPs) for termination of a growing DNA strand from a primer with its 3´- end designed immediately upstream of a site where single nucleotide polymorphism (SNP) occurs. The use of DNA polymerase offers a very high accuracy and specificity to APEX which in turn happens to be a method of choice for multiplex SNP detection. Coupling the high specificity of this method with the high sensitivity, low cost and compatibility for miniaturization of electrochemical techniques would offer an excellent platform for detection of mutation as well as sequencing of DNA templates. We are developing an electrochemical APEX for the analysis of SNPs found in the MYH7 gene for group of cardiomyopathy patients. ddNTPs were labeled with four different redox active compounds with four distinct potentials. Thiolated oligonucleotide probes were immobilised on gold and glassy carbon substrates which are followed by hybridisation with complementary target DNA just adjacent to the base to be extended by polymerase. Electrochemical interrogation was performed after the incorporation of the redox labelled dedioxynucleotide. The work involved the synthesis and characterisation of the redox labelled ddNTPs, optimisation and characterisation of surface functionalisation strategies and the nucleotide incorporation assays.

Keywords: array based primer extension, labelled ddNTPs, electrochemical, mutations

Procedia PDF Downloads 246
789 A Method to Compute Efficient 3D Helicopters Flight Trajectories Based On a Motion Polymorph-Primitives Algorithm

Authors: Konstanca Nikolajevic, Nicolas Belanger, David Duvivier, Rabie Ben Atitallah, Abdelhakim Artiba

Abstract:

Finding the optimal 3D path of an aerial vehicle under flight mechanics constraints is a major challenge, especially when the algorithm has to produce real-time results in flight. Kinematics models and Pythagorian Hodograph curves have been widely used in mobile robotics to solve this problematic. The level of difficulty is mainly driven by the number of constraints to be saturated at the same time while minimizing the total length of the path. In this paper, we suggest a pragmatic algorithm capable of saturating at the same time most of dimensioning helicopter 3D trajectories’ constraints like: curvature, curvature derivative, torsion, torsion derivative, climb angle, climb angle derivative, positions. The trajectories generation algorithm is able to generate versatile complex 3D motion primitives feasible by a helicopter with parameterization of the curvature and the climb angle. An upper ”motion primitives’ concatenation” algorithm is presented based. In this article we introduce a new way of designing three-dimensional trajectories based on what we call the ”Dubins gliding symmetry conjecture”. This extremely performing algorithm will be soon integrated to a real-time decisional system dealing with inflight safety issues.

Keywords: robotics, aerial robots, motion primitives, helicopter

Procedia PDF Downloads 616
788 Experimental Study for the Development of a Wireless Communication System in a Solar Central Tower Facility

Authors: Victor H. Benitez, Ramon V. Armas-Flores, Jesus H. Pacheco-Ramirez

Abstract:

Systems transforming solar energy into electrical power have emerged as a viable source of clean, renewable energy. Solar power tower technology is a good example of this type of system, which consists of several mobile mirrors, called heliostats, which reflect the sun's radiation to the same point, located on top of a tower at the center of heliostat field, for collection or transformation into another type of energy. The so-called Hermosillo’s Solar Platform (Plataforma Solar de Hermosillo, PSH, in Spanish) is a facility constituted with several heliostats, its aim and scope is for research purposes. In this paper, the implementation of a wireless communication system based on intelligent nodes is proposed in order to allow the communication and control of the heliostats in PSH. Intelligent nodes transmit information from one point to another, and can perform other actions that allow them to adapt to the conditions and limitations of a field of heliostats, thus achieving effective communication system. After deployment of the nodes in the heliostats, tests were conducted to measure the effectiveness of the communication, and determine the feasibility of using the proposed technologies. The test results were always positive, exceeding expectations held for its operation in the field of heliostats. Therefore, it was possible to validate the efficiency of the wireless communication system to be implemented in PSH, allowing communication and control of the heliostats.

Keywords: heliostat, intelligent node, solar energy, wireless communication

Procedia PDF Downloads 408
787 Impact of Map Generalization in Spatial Analysis

Authors: Lin Li, P. G. R. N. I. Pussella

Abstract:

When representing spatial data and their attributes on different types of maps, the scale plays a key role in the process of map generalization. The process is consisted with two main operators such as selection and omission. Once some data were selected, they would undergo of several geometrical changing processes such as elimination, simplification, smoothing, exaggeration, displacement, aggregation and size reduction. As a result of these operations at different levels of data, the geometry of the spatial features such as length, sinuosity, orientation, perimeter and area would be altered. This would be worst in the case of preparation of small scale maps, since the cartographer has not enough space to represent all the features on the map. What the GIS users do is when they wanted to analyze a set of spatial data; they retrieve a data set and does the analysis part without considering very important characteristics such as the scale, the purpose of the map and the degree of generalization. Further, the GIS users use and compare different maps with different degrees of generalization. Sometimes, GIS users are going beyond the scale of the source map using zoom in facility and violate the basic cartographic rule 'it is not suitable to create a larger scale map using a smaller scale map'. In the study, the effect of map generalization for GIS analysis would be discussed as the main objective. It was used three digital maps with different scales such as 1:10000, 1:50000 and 1:250000 which were prepared by the Survey Department of Sri Lanka, the National Mapping Agency of Sri Lanka. It was used common features which were on above three maps and an overlay analysis was done by repeating the data with different combinations. Road data, River data and Land use data sets were used for the study. A simple model, to find the best place for a wild life park, was used to identify the effects. The results show remarkable effects on different degrees of generalization processes. It can see that different locations with different geometries were received as the outputs from this analysis. The study suggests that there should be reasonable methods to overcome this effect. It can be recommended that, as a solution, it would be very reasonable to take all the data sets into a common scale and do the analysis part.

Keywords: generalization, GIS, scales, spatial analysis

Procedia PDF Downloads 328
786 An Evaluation of Kahoot Application and Its Environment as a Learning Tool

Authors: Muhammad Yasir Babar, Ebrahim Panah

Abstract:

Over the past 20 years, internet has seen continual advancement and with the advent of online technology, various types of web-based games have been developed. Games are frequently being used among different age groups from baby boomers to generation Z. Games are not only used for entertainment but also utilized as a learning approach transmitting education to a level that is more interesting and effective for students. One of the popular web-based education games is Kahoot with growing popularity and usage, which is being used in different fields of studies. However, little knowledge is available on university students’ perception of Kahoot environment and application for learning subjects. Hence, the objective of the current study is to investigate students’ perceptions of Kahoot application and environment as a learning tool. The study employed a survey approach by distributing Google Forms –created questionnaire, with high level of reliability index, to 62 students (11 males and 51 females). The findings show that students have positive attitudes towards Kahoot application and its environment for learning. Regarding Kahoot application, it was indicated that activities created using Kahoot are more interesting for students, Kahoot is useful for collaborative learning, and Kahoot enhances interest in learning lesson. In terms of Kahoot environment, it was found that using this application through mobile is easy for students, its design is simple and useful, Kahoot-created activities can easily be shared, and the application can easily be used on any platform. The findings of the study have implications for instructors, policymakers and curriculum developers.

Keywords: application, environment, Kahoot, learning tool

Procedia PDF Downloads 133
785 A Neuroscience-Based Learning Technique: Framework and Application to STEM

Authors: Dante J. Dorantes-González, Aldrin Balsa-Yepes

Abstract:

Existing learning techniques such as problem-based learning, project-based learning, or case study learning are learning techniques that focus mainly on technical details, but give no specific guidelines on learner’s experience and emotional learning aspects such as arousal salience and valence, being emotional states important factors affecting engagement and retention. Some approaches involving emotion in educational settings, such as social and emotional learning, lack neuroscientific rigorousness and use of specific neurobiological mechanisms. On the other hand, neurobiology approaches lack educational applicability. And educational approaches mainly focus on cognitive aspects and disregard conditioning learning. First, authors start explaining the reasons why it is hard to learn thoughtfully, then they use the method of neurobiological mapping to track the main limbic system functions, such as the reward circuit, and its relations with perception, memories, motivations, sympathetic and parasympathetic reactions, and sensations, as well as the brain cortex. The authors conclude explaining the major finding: The mechanisms of nonconscious learning and the triggers that guarantee long-term memory potentiation. Afterward, the educational framework for practical application and the instructors’ guidelines are established. An implementation example in engineering education is given, namely, the study of tuned-mass dampers for earthquake oscillations attenuation in skyscrapers. This work represents an original learning technique based on nonconscious learning mechanisms to enhance long-term memories that complement existing cognitive learning methods.

Keywords: emotion, emotion-enhanced memory, learning technique, STEM

Procedia PDF Downloads 91
784 Estimation of Lungs Physiological Motion for Patient Undergoing External Lung Irradiation

Authors: Yousif Mohamed Y. Abdallah

Abstract:

This is an experimental study deals with detection, measurement and analysis of the periodic physiological organ motion during external beam radiotherapy; to improve the accuracy of the radiation field placement, and to reduce the exposure of healthy tissue during radiation treatments. The importance of this study is to detect the maximum path of the mobile structures during radiotherapy delivery, to define the planning target volume (PTV) and irradiated volume during both inspiration and expiration period and to verify the target volume. In addition to its role to highlight the importance of the application of Intense Guided Radiotherapy (IGRT) methods in the field of radiotherapy. The results showed (body contour was equally (3.17 + 0.23 mm), for left lung displacement reading (2.56 + 0.99 mm) and right lung is (2.42 + 0.77 mm) which the radiation oncologist to take suitable countermeasures in case of significant errors. In addition, the use of the image registration technique for automatic position control is predicted potential motion. The motion ranged between 2.13 mm and 12.2 mm (low and high). In conclusion, individualized assessment of tumor mobility can improve the accuracy of target areas definition in patients undergo Sterostatic RT for stage I, II and III lung cancer (NSCLC). Definition of the target volume based on a single CT scan with a margin of 10 mm is clearly inappropriate.

Keywords: respiratory motion, external beam radiotherapy, image processing, lung

Procedia PDF Downloads 535
783 Characterization of Penicillin V Acid and Its Related Compounds by HPLC

Authors: Bahdja Guerfi, N. Hadhoum, I. Azouz, M. Bendoumia, S. Bouafia, F. Z. Hadjadj Aoul

Abstract:

Background: 'Penicillin V' is a narrow, bactericidal antibiotic of the beta-lactam family of the naturally occurring penicillin group. It is limited to infections due to the germs defined as sensitive. The objective of this work was to identify and to characterize Penicillin V acid and its related compounds by High-performance liquid chromatography (HPLC). Methods: Firstly phenoxymethylpenicillin was identified by an infrared absorption. The organoleptic characteristics, pH, and determination of water content were also studied. The dosage of Penicillin V acid active substance and the determination of its related compounds were carried on waters HPLC, equipped with a UV detector at 254 nm and Discovery HS C18 column (250 mm X 4.6 mm X 5 µm) which is maintained at room temperature. The flow rate was about 1 ml per min. A mixture of water, acetonitrile and acetic acid (65:35:01) was used as mobile phase for phenoxyacetic acid ‘impurity B' and a mixture of water, acetonitrile and acetic acid (650:150:5.75) for the assay and 4-hydroxypenicillin V 'impurity D'. Results: The identification of Penicillin V acid active substance and the evaluation of its chemical quality showed conformity with USP 35th edition. The Penicillin V acid content in the raw material is equal to 1692.22 UI/mg. The percentage content of phenoxyacetic acid and 4-hydroxypenicillin V was respectively: 0.035% and 0.323%. Conclusion: Through these results, we can conclude that the Penicillin V acid active substance tested is of good physicochemical quality.

Keywords: characterization, HPLC, Penicillin V acid, related substances

Procedia PDF Downloads 278
782 Imaging 255nm Tungsten Thin Film Adhesion with Picosecond Ultrasonics

Authors: A. Abbas, X. Tridon, J. Michelon

Abstract:

In the electronic or in the photovoltaic industries, components are made from wafers which are stacks of thin film layers of a few nanometers to serval micrometers thickness. Early evaluation of the bounding quality between different layers of a wafer is one of the challenges of these industries to avoid dysfunction of their final products. Traditional pump-probe experiments, which have been developed in the 70’s, give a partial solution to this problematic but with a non-negligible drawback. In fact, on one hand, these setups can generate and detect ultra-high ultrasounds frequencies which can be used to evaluate the adhesion quality of wafer layers. But, on the other hand, because of the quiet long acquisition time they need to perform one measurement, these setups remain shut in punctual measurement to evaluate global sample quality. This last point can lead to bad interpretation of the sample quality parameters, especially in the case of inhomogeneous samples. Asynchronous Optical Sampling (ASOPS) systems can perform sample characterization with picosecond acoustics up to 106 times faster than traditional pump-probe setups. This last point allows picosecond ultrasonic to unlock the acoustic imaging field at the nanometric scale to detect inhomogeneities regarding sample mechanical properties. This fact will be illustrated by presenting an image of the measured acoustical reflection coefficients obtained by mapping, with an ASOPS setup, a 255nm thin-film tungsten layer deposited on a silicone substrate. Interpretation of the coefficient reflection in terms of bounding quality adhesion will also be exposed. Origin of zones which exhibit good and bad quality bounding will be discussed.

Keywords: adhesion, picosecond ultrasonics, pump-probe, thin film

Procedia PDF Downloads 159
781 A Machine Learning Pipeline for Real-Time Activity Detection on Low Computational Power Devices for Metaverse Applications

Authors: Amit Kumar, Amanpreet Chander, Ashish Sahani

Abstract:

This paper presents our recent work on real-time human activity detection based on the media pipe pipeline and machine learning algorithms. The proposed system can detect human activities, including running, jumping, squatting, bending to the left or right, and standing still. This is a robust solution for developing a yoga, dance, metaverse, and fitness application that checks for the correction of the pose without having any additional monitor like a personal trainer. MediaPipe solution offers an open-source cross-platform which utilizes a two-step detector-tracker ML pipeline for live detection of key landmarks on our body which can be used for motion data collection. The prediction of real-time poses uses a variety of machine learning techniques and different types of analysis. Without primarily relying on powerful desktop environments for inference, our method achieves real-time performance on the majority of contemporary mobile phones, desktops/laptops, Python, or even the web. Experimental results show that our method outperforms the existing method in terms of accuracy and real-time capability, achieving an accuracy of 99.92% on testing datasets.

Keywords: human activity detection, media pipe, machine learning, metaverse applications

Procedia PDF Downloads 179
780 Land Use Planning Tool to Achieve Land Degradation Neutrality: Tunisia Case Study

Authors: Rafla Attia, Claudio Zucca, Bao Quang Le, Sana Dridi, Thouraya Sahli, Taoufik Hermassi

Abstract:

In Tunisia, landscape change and land degradation are critical issues for landscape conservation, management, and planning. Landscapes are undergoing crucial environmental problems made evident by soil degradation and desertification. Human improper uses of land resources (e.g., unsuitable land uses, unsustainable crop intensification, and poor rangeland management) and climate change are the main factors leading to the landscape transformation and desertification affecting high proportions of the Tunisian lands. Land use planning (LUP) to achieve Land Degradation Neutrality (LDN) must be supported by methodologies and technologies that help identify best solutions and practices and design context-specific sustainable land management (SLM) strategies. Such strategies must include restoration or rehabilitation efforts in areas with high land degradation, as well as prevention of degradation that could be caused by improper land use (LU) and land management (LM). The geoinformatics Land Use Planning for LDN (LUP4LDN) tool has been designed for this purpose. Its aim is to support national and sub-national planners in i) mapping geographic patterns of current land degradation; ii) anticipating further future land degradation expected in areas that are unsustainably managed; and iii) providing an interactive procedure for developing participatory LU-LM transitional scenarios over selected regions of interest and timeframes, visualizing the related expected levels of impacts on ecosystem services via maps and graphs. The tool has been co-developed and piloted with national stakeholders in Tunisia. The piloting implementation assessed how the LUP4LDN tool fits with existing LUP processes and the benefits achieved by using the tool to support land use planning for LDN.

Keywords: land use system, land cover, sustainable land management, land use planning for land degradation neutrality

Procedia PDF Downloads 77
779 Use of Concept Maps as a Tool for Evaluating Students' Understanding of Science

Authors: Aregamalage Sujeewa Vijayanthi Polgampala, Fang Huang

Abstract:

This study explores the genesis and development of concept mapping as a useful tool for science education and its effectiveness as technique for teaching and learning and evaluation for secondary science in schools and the role played by National College of Education science teachers. Concept maps, when carefully employed and executed serves as an integral part of teaching method and measure of effectiveness of teaching and tool for evaluation. Research has shown that science concept maps can have positive influence on student learning and motivation. The success of concept maps played in an instruction class depends on the type of theme selected, the development of learning outcomes, and the flexibility of instruction in providing library unit that is equipped with multimedia equipment where learners can interact. The study was restricted to 6 male and 9 female respondents' teachers in third-year internship pre service science teachers in Gampaha district Sri Lanka. Data were collected through 15 item questionnaire provided to learners and in depth interviews and class observations of 18 science classes. The two generated hypotheses for the study were rejected, while the results revealed that significant difference exists between factors influencing teachers' choice of concept maps, its usefulness and problems hindering the effectiveness of concept maps for teaching and learning process of secondary science in schools. It was examined that concept maps can be used as an effective measure to evaluate students understanding of concepts and misconceptions. Even the teacher trainees could not identify, key concept is on top, and subordinate concepts fall below. It is recommended that pre service science teacher trainees should be provided a thorough training using it as an evaluation instrument.

Keywords: concept maps, evaluation, learning science, misconceptions

Procedia PDF Downloads 274
778 Reliability and Validity for Measurement of Body Composition: A Field Method

Authors: Ahmad Hashim, Zarizi Ab Rahman

Abstract:

Measurement of body composition via a field method has the most popular instruments which are used to estimate the percentage of body fat. Among the instruments used are the Body Mass Index, Bio Impedance Analysis and Skinfold Test. All three of these instruments do not involve high costs, do not require high technical skills, are mobile, save time, and are suitable for use in large populations. Because all three instruments can estimate the percentage of body fat, but it is important to identify the most appropriate instruments and have high reliability. Hence, this study was conducted to determine the reliability and convergent validity of the instruments. A total of 40 students, males and females aged between 13 and 14 years participated in this study. The study found that the test retest and Pearson correlation coefficient of reliability for the three instruments is very high, r = .99. While the inter class reliability also are at high level with r = .99 for Body Mass Index and Bio Impedance Analysis, r = .96 for Skin fold test. Intra class reliability coefficient for these three instruments is too high for Body Mass Index r = .99, Bio Impedance Analysis r = .97, and Skin fold Test r = .90. However, Standard Error of Measurement value for all three instruments indicates the Body Mass Index is the most appropriate instrument with a mean value of .000672 compared with other instruments. The findings show that the Body Mass Index is an instrument which is the most accurate and reliable in estimating body fat percentage for the population studied.

Keywords: reliability, validity, body mass index, bio impedance analysis and skinfold test

Procedia PDF Downloads 335
777 Banking and Accounting Analysis Researches Effect on Environment and Income

Authors: Gerges Samaan Henin Abdalla

Abstract:

Ultra-secured methods of banking services have been introduced to the customer, such as online banking. Banks have begun to consider electronic banking (e-banking) as a way to replace some traditional branch functions by using the Internet as a distribution channel. Some consumers have at least one account at multiple banks and access these accounts through online banking. To check their current net worth, clients need to log into each of their accounts, get detailed information, and work toward consolidation. Not only is it time consuming, but it is also a repeatable activity with a certain frequency. To solve this problem, the concept of account aggregation was added as a solution. Account consolidation in e-banking as a form of electronic banking appears to build a stronger relationship with customers. An account linking service is generally referred to as a service that allows customers to manage their bank accounts held at different institutions via a common online banking platform that places a high priority on security and data protection. Consumers have at least one account at multiple banks and access these accounts through online banking. To check their current net worth, clients need to log into each of their accounts, get detailed information, and work toward consolidation. The article provides an overview of the account aggregation approach in e-banking as a new service in the area of e-banking.

Keywords: compatibility, complexity, mobile banking, observation, risk banking technology, Internet banks, modernization of banks, banks, account aggregation, security, enterprise development

Procedia PDF Downloads 45
776 Study of Structural Behavior and Proton Conductivity of Inorganic Gel Paste Electrolyte at Various Phosphorous to Silicon Ratio by Multiscale Modelling

Authors: P. Haldar, P. Ghosh, S. Ghoshdastidar, K. Kargupta

Abstract:

In polymer electrolyte membrane fuel cells (PEMFC), the membrane electrode assembly (MEA) is consisting of two platinum coated carbon electrodes, sandwiched with one proton conducting phosphoric acid doped polymeric membrane. Due to low mechanical stability, flooding and fuel cell crossover, application of phosphoric acid in polymeric membrane is very critical. Phosphorous and silica based 3D inorganic gel gains the attention in the field of supercapacitors, fuel cells and metal hydrate batteries due to its thermally stable highly proton conductive behavior. Also as a large amount of water molecule and phosphoric acid can easily get trapped in Si-O-Si network cavities, it causes a prevention in the leaching out. In this study, we have performed molecular dynamics (MD) simulation and first principle calculations to understand the structural, electronics and electrochemical and morphological behavior of this inorganic gel at various P to Si ratios. We have used dipole-dipole interactions, H bonding, and van der Waals forces to study the main interactions between the molecules. A 'structure property-performance' mapping is initiated to determine optimum P to Si ratio for best proton conductivity. We have performed the MD simulations at various temperature to understand the temperature dependency on proton conductivity. The observed results will propose a model which fits well with experimental data and other literature values. We have also studied the mechanism behind proton conductivity. And finally we have proposed a structure for the gel paste with optimum P to Si ratio.

Keywords: first principle calculation, molecular dynamics simulation, phosphorous and silica based 3D inorganic gel, polymer electrolyte membrane fuel cells, proton conductivity

Procedia PDF Downloads 129
775 Player Experience: A Research on Cross-Platform Supported Games

Authors: Salih Akkemik

Abstract:

User Experience has a characterized perspective based on two fundamentals: the usage process and the product. Digital games can be considered as a special interactive system. This system has a very specific purpose and this is to make the player feel good while playing. At this point, Player Experience (PX) and User Experience (UX) are similar. UX focuses on the user feels good, PX focuses on the player feels good. The most important difference between the two is the action taken. These are actions of using and playing. In this study, the player experience will be examined primarily. PX may differ on different platforms. Nowadays, companies are releasing the successful and high-income games that they have developed with cross-platform support. Cross-platform is the most common expression that an application can run on different operating systems, in other words, be developed to support different operating systems. In terms of digital games, cross-platform support means that a game can be played on a computer, console or mobile device environment, more specifically, the game developed is designed and programmed to be played in the same way on at least two different platforms, such as Windows, MacOS, Linux, iOS, Android, Orbis OS or Xbox OS. Different platforms also accommodate different player groups, profiles and preferences. This study aims to examine these different player profiles in terms of player experience and to determine the effects of cross-platform support on player experience.

Keywords: cross-platform, digital games, player experience, user experience

Procedia PDF Downloads 206
774 Multi-Impairment Compensation Based Deep Neural Networks for 16-QAM Coherent Optical Orthogonal Frequency Division Multiplexing System

Authors: Ying Han, Yuanxiang Chen, Yongtao Huang, Jia Fu, Kaile Li, Shangjing Lin, Jianguo Yu

Abstract:

In long-haul and high-speed optical transmission system, the orthogonal frequency division multiplexing (OFDM) signal suffers various linear and non-linear impairments. In recent years, researchers have proposed compensation schemes for specific impairment, and the effects are remarkable. However, different impairment compensation algorithms have caused an increase in transmission delay. With the widespread application of deep neural networks (DNN) in communication, multi-impairment compensation based on DNN will be a promising scheme. In this paper, we propose and apply DNN to compensate multi-impairment of 16-QAM coherent optical OFDM signal, thereby improving the performance of the transmission system. The trained DNN models are applied in the offline digital signal processing (DSP) module of the transmission system. The models can optimize the constellation mapping signals at the transmitter and compensate multi-impairment of the OFDM decoded signal at the receiver. Furthermore, the models reduce the peak to average power ratio (PAPR) of the transmitted OFDM signal and the bit error rate (BER) of the received signal. We verify the effectiveness of the proposed scheme for 16-QAM Coherent Optical OFDM signal and demonstrate and analyze transmission performance in different transmission scenarios. The experimental results show that the PAPR and BER of the transmission system are significantly reduced after using the trained DNN. It shows that the DNN with specific loss function and network structure can optimize the transmitted signal and learn the channel feature and compensate for multi-impairment in fiber transmission effectively.

Keywords: coherent optical OFDM, deep neural network, multi-impairment compensation, optical transmission

Procedia PDF Downloads 143
773 A Review on Using Executive Function to Understand the Limited Efficacy of Weight-Loss Interventions

Authors: H. Soltani, Kevin Laugero

Abstract:

Obesity is becoming an increasingly critical issue in the United States due to the steady and substantial increase in prevalence over the last 30 years. Existing interventions have been able to help participants achieve short-term weight loss, but have failed to show long-term results. The complex nature of behavioral change remains one of the most difficult barriers in promoting sustainable weight-loss in overweight individuals. Research suggests that the 'intention-behavior gap' can be explained by a person’s ability to regulate higher-order thinking, or Executive Function (EF). A review of 63 research articles was completed in fall of 2017 to identify the role of EF in regulating eating behavior and to identify whether there is a potential for improving dietary quality by enhancing EF. Results showed that poor EF is positively associated with obesogenic behavior, namely increased consumption of highly palatable foods, eating in the absence of hunger, high saturated fat intake and low fruit and vegetable consumption. Recent research has indicated that interventions targeting an improvement in EF can be successful in helping promote healthy behaviors. Furthermore, interventions of longer duration have a more lasting and versatile effect on weight loss and maintenance. This may present an opportunity for the increasingly ubiquitous use of mobile application technology.

Keywords: eating behavior, executive function, nutrition, obesity, weight-loss

Procedia PDF Downloads 165
772 Food Service Waste Management In Nigeria: Emerging Opportunities And Policy Initiatives For Mitigation

Authors: Victor Oyewumi Ogunbiyi

Abstract:

Food waste is recognised as one of the major global challenges in achieving a sustainable future. Currently, very little is known about the multi-stakeholder approach to food waste management downstream of the supply chain, particularly in the foodservice sector. In order to better understand and explain the complex issues of food waste, a qualitative study was conducted on the generation of food waste in food services (restaurants, catering, canteens, and local food vendors) and policy initiatives to mitigate it from the perspective of the stakeholders. A semi-structured interview approach and observation were used to collect data from some 32 selected stakeholders in Garki, Abuja, Nigeria. Thematic analysis was employed to analyse the data from the qualitative instrument adopted in this study. Results revealed that the attitude of stakeholders, poor environmental hygiene, poor food cooking skills and handling, and lack of communication are the major causes of food waste. This study identified seven policy initiatives: regulations, information and education campaigns, economic instruments, mobile applications, stakeholders’ collaboration, firm internal action, and training. Finally, we link policy initiatives to food waste mitigation to provide a response to the damaging shock of food waste.

Keywords: food waste, foodservices, emerging opportunities, policy initiatives, food waste prevention, multistakeholder. garki district-abuja

Procedia PDF Downloads 81
771 Blueprinting of a Normalized Supply Chain Processes: Results in Implementing Normalized Software Systems

Authors: Bassam Istanbouli

Abstract:

With the technology evolving every day and with the increase in global competition, industries are always under the pressure to be the best. They need to provide good quality products at competitive prices, when and how the customer wants them.  In order to achieve this level of service, products and their respective supply chain processes need to be flexible and evolvable; otherwise changes will be extremely expensive, slow and with many combinatorial effects. Those combinatorial effects impact the whole organizational structure, from a management, financial, documentation, logistics and specially the information system Enterprise Requirement Planning (ERP) perspective. By applying the normalized system concept/theory to segments of the supply chain, we believe minimal effects, especially at the time of launching an organization global software project. The purpose of this paper is to point out that if an organization wants to develop a software from scratch or implement an existing ERP software for their business needs and if their business processes are normalized and modular then most probably this will yield to a normalized and modular software system that can be easily modified when the business evolves. Another important goal of this paper is to increase the awareness regarding the design of the business processes in a software implementation project. If the blueprints created are normalized then the software developers and configurators will use those modular blueprints to map them into modular software. This paper only prepares the ground for further studies;  the above concept will be supported by going through the steps of developing, configuring and/or implementing a software system for an organization by using two methods: The Software Development Lifecycle method (SDLC) and the Accelerated SAP implementation method (ASAP). Both methods start with the customer requirements, then blue printing of its business processes and finally mapping those processes into a software system.  Since those requirements and processes are the starting point of the implementation process, then normalizing those processes will end up in a normalizing software.

Keywords: blueprint, ERP, modular, normalized

Procedia PDF Downloads 139
770 Numerical Investigation of Supertall Buildings and Using Aerodynamic Characteristics to Create New Wind Power Sources

Authors: Mohammad A. Masoumi, Mohammad Zare, Soroush Sabouki

Abstract:

This study investigates the aerodynamic characteristics of supertall buildings to evaluate wind turbine installation at high altitudes. Most recent studies have investigated supertall buildings at a horizontal plane, while a vertical plan could be as important, especially to install wind turbines. A typical square-plan building with a height of 500 m is investigated numerically at horizontal and vertical plans to evaluate wind power generation potentials. The results show good agreement with experimental data and past studies. Then four new geometries are proposed to improvise regions at high altitudes to install wind turbines. Evaluating the simulations shows two regions with high power density, which have the possibility to install wind turbines. Results show that improvised regions to install wind turbines at high altitudes contain significant power density while higher power density is found behind buildings in a far distance. In addition, power density fluctuations behind buildings are investigated, which show decreasing fluctuations by reaching 50 m altitude while altitudes lower than 20 m have the most fluctuations.

Keywords: wind power, supertall building, power density, aerodynamic characteristics, wind turbine mobile, quality assurance, testing, applications

Procedia PDF Downloads 168
769 Vision-Based Collision Avoidance for Unmanned Aerial Vehicles by Recurrent Neural Networks

Authors: Yao-Hong Tsai

Abstract:

Due to the sensor technology, video surveillance has become the main way for security control in every big city in the world. Surveillance is usually used by governments for intelligence gathering, the prevention of crime, the protection of a process, person, group or object, or the investigation of crime. Many surveillance systems based on computer vision technology have been developed in recent years. Moving target tracking is the most common task for Unmanned Aerial Vehicle (UAV) to find and track objects of interest in mobile aerial surveillance for civilian applications. The paper is focused on vision-based collision avoidance for UAVs by recurrent neural networks. First, images from cameras on UAV were fused based on deep convolutional neural network. Then, a recurrent neural network was constructed to obtain high-level image features for object tracking and extracting low-level image features for noise reducing. The system distributed the calculation of the whole system to local and cloud platform to efficiently perform object detection, tracking and collision avoidance based on multiple UAVs. The experiments on several challenging datasets showed that the proposed algorithm outperforms the state-of-the-art methods.

Keywords: unmanned aerial vehicle, object tracking, deep learning, collision avoidance

Procedia PDF Downloads 160
768 Survival and Growth Factors of Korean Start-Ups: Focusing on the Industrial Characteristics

Authors: Hanei Son

Abstract:

Since the beginning of the 2010s, ‘start-up boom’ has continued with the creation of many new enterprises in Korea. Such tendency was led by various changes in society such as emergence and diffusion of smartphones. Especially, the Korean government has been interested in start-ups and entrepreneurship as an alternative engine for Korea's economic growth. With strong support from the government, as a result, many new enterprises have been established for recent years and the Korean government seems to have achieved its goal: expanding the basis of start-ups. However, it is unclear which factors affect the survival and growth of these new enterprises after their creation. Therefore, this study aims to identify which start-ups from early 2010s survived and which factors influenced their survival and growth. The study will strongly focus on which industries the new enterprises were in, as environmental elements are expected to be critical factors for business of start-ups in Korean context. For this purpose, 105 companies which were introduced as high potential start-ups from 2010 to 2012 were considered in the analysis. According to their current status, dead or alive, the start-ups were categorized by their industries and service area. Through this analysis, it was observed that many start-ups that are still in business are in internet or mobile platform businesses and four major sectors. In each group, a representative case has been studied to reveal its survival and growth factors. The results point to the importance of industrial characteristics for the survival and success of Korean startups and offer political implications in which sector and business more potentials for start-ups in Korea lie in.

Keywords: government support for start-ups, industrial characteristics, Korean start-ups, survival of start-ups

Procedia PDF Downloads 186