Search results for: inventory classification
1818 Application of Lean Manufacturing in Brake Shoe Manufacturing Plant: A Case Study
Authors: Anees K. Ahamed, Aakash Kumar R. G., Raj M. Mohan
Abstract:
The main objective is to apply lean tools to identify and eliminate waste in and among the work stations so as to improve the process speed and quality. From the top seven wastes in the lean concept, we consider the movement of materials, defects, and inventory for the improvement since these cause the major impact on the performance measures. The layout was improved to reduce the movement of materials. It also quantifies the reduction in movement among the work stations. Value stream mapping has been used for identification of waste. Cause and effect diagram and 5W analysis are used to identify the reasons for defects and to provide the counter measures. Some cycle time reduction techniques also proposed to improve the productivity. Lean Audit check sheet was also used to identify the current position of the industry and to identify the gap to make the industry Lean.Keywords: cause and effect diagram, cycle time reduction, defects, lean, waste reduction
Procedia PDF Downloads 3851817 Development of an EEG-Based Real-Time Emotion Recognition System on Edge AI
Authors: James Rigor Camacho, Wansu Lim
Abstract:
Over the last few years, the development of new wearable and processing technologies has accelerated in order to harness physiological data such as electroencephalograms (EEGs) for EEG-based applications. EEG has been demonstrated to be a source of emotion recognition signals with the highest classification accuracy among physiological signals. However, when emotion recognition systems are used for real-time classification, the training unit is frequently left to run offline or in the cloud rather than working locally on the edge. That strategy has hampered research, and the full potential of using an edge AI device has yet to be realized. Edge AI devices are computers with high performance that can process complex algorithms. It is capable of collecting, processing, and storing data on its own. It can also analyze and apply complicated algorithms like localization, detection, and recognition on a real-time application, making it a powerful embedded device. The NVIDIA Jetson series, specifically the Jetson Nano device, was used in the implementation. The cEEGrid, which is integrated to the open-source brain computer-interface platform (OpenBCI), is used to collect EEG signals. An EEG-based real-time emotion recognition system on Edge AI is proposed in this paper. To perform graphical spectrogram categorization of EEG signals and to predict emotional states based on input data properties, machine learning-based classifiers were used. Until the emotional state was identified, the EEG signals were analyzed using the K-Nearest Neighbor (KNN) technique, which is a supervised learning system. In EEG signal processing, after each EEG signal has been received in real-time and translated from time to frequency domain, the Fast Fourier Transform (FFT) technique is utilized to observe the frequency bands in each EEG signal. To appropriately show the variance of each EEG frequency band, power density, standard deviation, and mean are calculated and employed. The next stage is to identify the features that have been chosen to predict emotion in EEG data using the K-Nearest Neighbors (KNN) technique. Arousal and valence datasets are used to train the parameters defined by the KNN technique.Because classification and recognition of specific classes, as well as emotion prediction, are conducted both online and locally on the edge, the KNN technique increased the performance of the emotion recognition system on the NVIDIA Jetson Nano. Finally, this implementation aims to bridge the research gap on cost-effective and efficient real-time emotion recognition using a resource constrained hardware device, like the NVIDIA Jetson Nano. On the cutting edge of AI, EEG-based emotion identification can be employed in applications that can rapidly expand the research and implementation industry's use.Keywords: edge AI device, EEG, emotion recognition system, supervised learning algorithm, sensors
Procedia PDF Downloads 1051816 Classification on Statistical Distributions of a Complex N-Body System
Authors: David C. Ni
Abstract:
Contemporary models for N-body systems are based on temporal, two-body, and mass point representation of Newtonian mechanics. Other mainstream models include 2D and 3D Ising models based on local neighborhood the lattice structures. In Quantum mechanics, the theories of collective modes are for superconductivity and for the long-range quantum entanglement. However, these models are still mainly for the specific phenomena with a set of designated parameters. We are therefore motivated to develop a new construction directly from the complex-variable N-body systems based on the extended Blaschke functions (EBF), which represent a non-temporal and nonlinear extension of Lorentz transformation on the complex plane – the normalized momentum spaces. A point on the complex plane represents a normalized state of particle momentums observed from a reference frame in the theory of special relativity. There are only two key parameters, normalized momentum and nonlinearity for modelling. An algorithm similar to Jenkins-Traub method is adopted for solving EBF iteratively. Through iteration, the solution sets show a form of σ + i [-t, t], where σ and t are the real numbers, and the [-t, t] shows various distributions, such as 1-peak, 2-peak, and 3-peak etc. distributions and some of them are analog to the canonical distributions. The results of the numerical analysis demonstrate continuum-to-discreteness transitions, evolutional invariance of distributions, phase transitions with conjugate symmetry, etc., which manifest the construction as a potential candidate for the unification of statistics. We hereby classify the observed distributions on the finite convergent domains. Continuous and discrete distributions both exist and are predictable for given partitions in different regions of parameter-pair. We further compare these distributions with canonical distributions and address the impacts on the existing applications.Keywords: blaschke, lorentz transformation, complex variables, continuous, discrete, canonical, classification
Procedia PDF Downloads 3091815 An Autonomous Passive Acoustic System for Detection, Tracking and Classification of Motorboats in Portofino Sea
Authors: A. Casale, J. Alessi, C. N. Bianchi, G. Bozzini, M. Brunoldi, V. Cappanera, P. Corvisiero, G. Fanciulli, D. Grosso, N. Magnoli, A. Mandich, C. Melchiorre, C. Morri, P. Povero, N. Stasi, M. Taiuti, G. Viano, M. Wurtz
Abstract:
This work describes a real-time algorithm for detecting, tracking and classifying single motorboats, developed using the acoustic data recorded by a hydrophone array within the framework of EU LIFE + project ARION (LIFE09NAT/IT/000190). The project aims to improve the conservation status of bottlenose dolphins through a real-time simultaneous monitoring of their population and surface ship traffic. A Passive Acoustic Monitoring (PAM) system is installed on two autonomous permanent marine buoys, located close to the boundaries of the Marine Protected Area (MPA) of Portofino (Ligurian Sea- Italy). Detecting surface ships is also a necessity in many other sensible areas, such as wind farms, oil platforms, and harbours. A PAM system could be an effective alternative to the usual monitoring systems, as radar or active sonar, for localizing unauthorized ship presence or illegal activities, with the advantage of not revealing its presence. Each ARION buoy consists of a particular type of structure, named meda elastica (elastic beacon) composed of a main pole, about 30-meter length, emerging for 7 meters, anchored to a mooring of 30 tons at 90 m depth by an anti-twist steel wire. Each buoy is equipped with a floating element and a hydrophone tetrahedron array, whose raw data are send via a Wi-Fi bridge to a ground station where real-time analysis is performed. Bottlenose dolphin detection algorithm and ship monitoring algorithm are operating in parallel and in real time. Three modules were developed and commissioned for ship monitoring. The first is the detection algorithm, based on Time Difference Of Arrival (TDOA) measurements, i.e., the evaluation of angular direction of the target respect to each buoy and the triangulation for obtaining the target position. The second is the tracking algorithm, based on a Kalman filter, i.e., the estimate of the real course and speed of the target through a predictor filter. At last, the classification algorithm is based on the DEMON method, i.e., the extraction of the acoustic signature of single vessels. The following results were obtained; the detection algorithm succeeded in evaluating the bearing angle with respect to each buoy and the position of the target, with an uncertainty of 2 degrees and a maximum range of 2.5 km. The tracking algorithm succeeded in reconstructing the real vessel courses and estimating the speed with an accuracy of 20% respect to the Automatic Identification System (AIS) signals. The classification algorithm succeeded in isolating the acoustic signature of single vessels, demonstrating its temporal stability and the consistency of both buoys results. As reference, the results were compared with the Hilbert transform of single channel signals. The algorithm for tracking multiple targets is ready to be developed, thanks to the modularity of the single ship algorithm: the classification module will enumerate and identify all targets present in the study area; for each of them, the detection module and the tracking module will be applied to monitor their course.Keywords: acoustic-noise, bottlenose-dolphin, hydrophone, motorboat
Procedia PDF Downloads 1731814 Multivariate Data Analysis for Automatic Atrial Fibrillation Detection
Authors: Zouhair Haddi, Stephane Delliaux, Jean-Francois Pons, Ismail Kechaf, Jean-Claude De Haro, Mustapha Ouladsine
Abstract:
Atrial fibrillation (AF) has been considered as the most common cardiac arrhythmia, and a major public health burden associated with significant morbidity and mortality. Nowadays, telemedical approaches targeting cardiac outpatients situate AF among the most challenged medical issues. The automatic, early, and fast AF detection is still a major concern for the healthcare professional. Several algorithms based on univariate analysis have been developed to detect atrial fibrillation. However, the published results do not show satisfactory classification accuracy. This work was aimed at resolving this shortcoming by proposing multivariate data analysis methods for automatic AF detection. Four publicly-accessible sets of clinical data (AF Termination Challenge Database, MIT-BIH AF, Normal Sinus Rhythm RR Interval Database, and MIT-BIH Normal Sinus Rhythm Databases) were used for assessment. All time series were segmented in 1 min RR intervals window and then four specific features were calculated. Two pattern recognition methods, i.e., Principal Component Analysis (PCA) and Learning Vector Quantization (LVQ) neural network were used to develop classification models. PCA, as a feature reduction method, was employed to find important features to discriminate between AF and Normal Sinus Rhythm. Despite its very simple structure, the results show that the LVQ model performs better on the analyzed databases than do existing algorithms, with high sensitivity and specificity (99.19% and 99.39%, respectively). The proposed AF detection holds several interesting properties, and can be implemented with just a few arithmetical operations which make it a suitable choice for telecare applications.Keywords: atrial fibrillation, multivariate data analysis, automatic detection, telemedicine
Procedia PDF Downloads 2671813 Factors Associated with Suicidal Ideation among Undergraduate College Students
Authors: Samantha Vennice G. Sarcia
Abstract:
A person dies every 40 seconds throughout the world due to suicide-related behaviors. Suicidal ideation is a strong precursor to suicide completion. It is one of the major health challenges faced by the world today thus, it is highly substantial. The present study investigated the influence of personality traits and socio-demographic characteristics in predicting suicidal ideation. Using the Suicide Behaviors Questionnaire-Revised and the Big Five Inventory, the degree of suicidal ideation and the associated personality traits were identified. Out of 194 students from the allied health courses, the findings suggest that the college students are at-risk and have passive thoughts about suicide. Using multiple regression analysis, there was an identified significant relationship among the factors associated with suicidal ideation, particularly the number of persons in the household, living arrangement, attendance in church activities, agreeableness, conscientiousness, and neuroticism. Findings can help in the development of campus-based suicide prevention programs.Keywords: depression, personality traits, suicidal ideation, suicide
Procedia PDF Downloads 2251812 Life Cycle Assessment in Road Pavements: A Literature Review and the Potential Use in Brazil
Authors: B. V. Santos, M. T. M. Carvalho, J. H. S. Rêgo
Abstract:
The article presents a literature review on recent advances related to studies of the environmental impact of road pavements, with reference to the concepts of Life Cycle Assessment (LCA). An introduction with the main motivations for the development of the research is presented, with a current overview of the Brazilian transport infrastructure and the projections for the road mode for the coming years, and the possibility of using the referred methodology by the road sector in Brazil. The article explores the origin of LCA in road pavements and the details linked to its implementation from the perspective of the four main phases of the study (goal and scope definition, inventory analysis, impact assessment, and interpretation). Finally, the main advances and deficiencies observed in the selected studies are gathered, with the proposition of research fields that can be explored in future national or international studies of LCA of road pavements.Keywords: Brazil, life cycle assessment, road pavements, sustainable
Procedia PDF Downloads 791811 Emotion Detection in Twitter Messages Using Combination of Long Short-Term Memory and Convolutional Deep Neural Networks
Authors: Bahareh Golchin, Nooshin Riahi
Abstract:
One of the most significant issues as attended a lot in recent years is that of recognizing the sentiments and emotions in social media texts. The analysis of sentiments and emotions is intended to recognize the conceptual information such as the opinions, feelings, attitudes and emotions of people towards the products, services, organizations, people, topics, events and features in the written text. These indicate the greatness of the problem space. In the real world, businesses and organizations are always looking for tools to gather ideas, emotions, and directions of people about their products, services, or events related to their own. This article uses the Twitter social network, one of the most popular social networks with about 420 million active users, to extract data. Using this social network, users can share their information and opinions about personal issues, policies, products, events, etc. It can be used with appropriate classification of emotional states due to the availability of its data. In this study, supervised learning and deep neural network algorithms are used to classify the emotional states of Twitter users. The use of deep learning methods to increase the learning capacity of the model is an advantage due to the large amount of available data. Tweets collected on various topics are classified into four classes using a combination of two Bidirectional Long Short Term Memory network and a Convolutional network. The results obtained from this study with an average accuracy of 93%, show good results extracted from the proposed framework and improved accuracy compared to previous work.Keywords: emotion classification, sentiment analysis, social networks, deep neural networks
Procedia PDF Downloads 1361810 Analysis of Information Sharing and Capacity Constraint on Backlog Bullwhip Effect in Two Level Supply Chain
Authors: Matloub Hussaina
Abstract:
This paper investigates the impact of information sharing and capacity constraints on backlog bullwhip effect of Automatic Pipe Line Inventory and Order Based Production Control System (APIOBPCS). System dynamic simulation using iThink Software has been applied. It has been found that smooth ordering by Tier 1 can be achieved when Tier 1 has medium capacity constraints. Simulation experiments also show that information sharing helps to reduce 50% of backlog bullwhip effect in capacitated supply chains. This knowledge is of value per se, giving supply chain operations managers and designers a practical way in to controlling the backlog bullwhip effect. Future work should investigate the total cost implications of capacity constraints and safety stocks in multi-echelon supply chain.Keywords: supply chain dynamics, information sharing, capacity constraints, simulation, APIOBPCS
Procedia PDF Downloads 3181809 Locus of Control, Metacognitive Knowledge, Metacognitive Regulation, and Student Performance in an Introductory Economics Course
Authors: Ahmad A. Kader
Abstract:
In the principles of Microeconomics course taught during the Fall Semester 2019, 158out of 179 students participated in the completion of two questionnaires and a survey describing their demographic and academic profiles. The two questionnaires include the 29 items of the Rotter Locus of Control Scale and the 52 items of the Schraw andDennisonMetacognitive Awareness Scale. The 52 items consist of 17 items describing knowledge of cognition and 37 items describing the regulation of cognition. The paper is intended to show the combined influence of locus of control, metacognitive knowledge, and metacognitive regulation on student performance. The survey covers variables that have been tested and recognized in economic education literature, which include GPA, gender, age, course level, race, student classification, whether the course was required or elective, employments, whether a high school economic course was taken, and attendance. Regression results show that of the economic education variables, GPA, classification, whether the course was required or elective, and attendance are the only significant variables in their influence on student grade. Of the educational psychology variables, the regression results show that the locus of control variable has a negative and significant effect, while the metacognitive knowledge variable has a positive and significant effect on student grade. Also, the adjusted R square value increased markedly with the addition of the locus of control, metacognitive knowledge, and metacognitive regulation variables to the regression equation. The t test results also show that students who are internally oriented and are high on the metacognitive knowledge scale significantly outperform students who are externally oriented and are low on the metacognitive knowledge scale. The implication of these results for educators is discussed in the paper.Keywords: locus of control, metacognitive knowledge, metacognitive regulation, student performance, economic education
Procedia PDF Downloads 1201808 Optimal Policies in a Two-Level Supply Chain with Defective Product and Price Dependent Demand
Authors: Samira Mohabbatdar, Abbas Ahmadi, Mohsen S. Sajadieh
Abstract:
This paper deals with a two-level supply chain consisted of one manufacturer and one retailer for single-type product. The demand function of the customers depends on price. We consider an integrated production inventory system where the manufacturer processes raw materials in order to deliver finished product with imperfect quality to the retailer. Then retailer inspects the products and after that delivers perfect products to customers. The proposed model is based on the joint total profit of both the manufacturer and the retailer, and it determines the optimal ordering lot-size, number of shipment and selling price of the retailer. A numerical example is provided to analyse and illustrate the behaviour and application of the model. Finally, sensitivity analysis of the key parameters are presented to test feasibility of the model.Keywords: supply chain, pricing policy, defective quality, joint economic lot sizing
Procedia PDF Downloads 3371807 Classification of Coughing and Breathing Activities Using Wearable and a Light-Weight DL Model
Authors: Subham Ghosh, Arnab Nandi
Abstract:
Background: The proliferation of Wireless Body Area Networks (WBAN) and Internet of Things (IoT) applications demonstrates the potential for continuous monitoring of physical changes in the body. These technologies are vital for health monitoring tasks, such as identifying coughing and breathing activities, which are necessary for disease diagnosis and management. Monitoring activities such as coughing and deep breathing can provide valuable insights into a variety of medical issues. Wearable radio-based antenna sensors, which are lightweight and easy to incorporate into clothing or portable goods, provide continuous monitoring. This mobility gives it a substantial advantage over stationary environmental sensors like as cameras and radar, which are constrained to certain places. Furthermore, using compressive techniques provides benefits such as reduced data transmission speeds and memory needs. These wearable sensors offer more advanced and diverse health monitoring capabilities. Methodology: This study analyzes the feasibility of using a semi-flexible antenna operating at 2.4 GHz (ISM band) and positioned around the neck and near the mouth to identify three activities: coughing, deep breathing, and idleness. Vector network analyzer (VNA) is used to collect time-varying complex reflection coefficient data from perturbed antenna nearfield. The reflection coefficient (S11) conveys nuanced information caused by simultaneous variations in the nearfield radiation of three activities across time. The signatures are sparsely represented with gaussian windowed Gabor spectrograms. The Gabor spectrogram is used as a sparse representation approach, which reassigns the ridges of the spectrogram images to improve their resolution and focus on essential components. The antenna is biocompatible in terms of specific absorption rate (SAR). The sparsely represented Gabor spectrogram pictures are fed into a lightweight deep learning (DL) model for feature extraction and classification. Two antenna locations are investigated in order to determine the most effective localization for three different activities. Findings: Cross-validation techniques were used on data from both locations. Due to the complex form of the recorded S11, separate analyzes and assessments were performed on the magnitude, phase, and their combination. The combination of magnitude and phase fared better than the separate analyses. Various sliding window sizes, ranging from 1 to 5 seconds, were tested to find the best window for activity classification. It was discovered that a neck-mounted design was effective at detecting the three unique behaviors.Keywords: activity recognition, antenna, deep-learning, time-frequency
Procedia PDF Downloads 91806 An Open Source Advertisement System
Authors: Pushkar Umaranikar, Chris Pollett
Abstract:
An online advertisement system and its implementation for the Yioop open source search engine are presented. This system supports both selling advertisements and displaying them within search results. The selling of advertisements is done using a system to auction off daily impressions for keyword searches. This is an open, ascending price auction system in which all accepted bids will receive a fraction of the auctioned day’s impressions. New bids in our system are required to be at least one half of the sum of all previous bids ensuring the number of accepted bids is logarithmic in the total ad spend on a keyword for a day. The mechanics of creating an advertisement, attaching keywords to it, and adding it to an advertisement inventory are described. The algorithm used to go from accepted bids for a keyword to which ads are displayed at search time is also presented. We discuss properties of our system and compare it to existing auction systems and systems for selling online advertisements.Keywords: online markets, online ad system, online auctions, search engines
Procedia PDF Downloads 3261805 Revolutionizing Financial Forecasts: Enhancing Predictions with Graph Convolutional Networks (GCN) - Long Short-Term Memory (LSTM) Fusion
Authors: Ali Kazemi
Abstract:
Those within the volatile and interconnected international economic markets, appropriately predicting market trends, hold substantial fees for traders and financial establishments. Traditional device mastering strategies have made full-size strides in forecasting marketplace movements; however, monetary data's complicated and networked nature calls for extra sophisticated processes. This observation offers a groundbreaking method for monetary marketplace prediction that leverages the synergistic capability of Graph Convolutional Networks (GCNs) and Long Short-Term Memory (LSTM) networks. Our suggested algorithm is meticulously designed to forecast the traits of inventory market indices and cryptocurrency costs, utilizing a comprehensive dataset spanning from January 1, 2015, to December 31, 2023. This era, marked by sizable volatility and transformation in financial markets, affords a solid basis for schooling and checking out our predictive version. Our algorithm integrates diverse facts to construct a dynamic economic graph that correctly reflects market intricacies. We meticulously collect opening, closing, and high and low costs daily for key inventory marketplace indices (e.g., S&P 500, NASDAQ) and widespread cryptocurrencies (e.g., Bitcoin, Ethereum), ensuring a holistic view of marketplace traits. Daily trading volumes are also incorporated to seize marketplace pastime and liquidity, providing critical insights into the market's shopping for and selling dynamics. Furthermore, recognizing the profound influence of the monetary surroundings on financial markets, we integrate critical macroeconomic signs with hobby fees, inflation rates, GDP increase, and unemployment costs into our model. Our GCN algorithm is adept at learning the relational patterns amongst specific financial devices represented as nodes in a comprehensive market graph. Edges in this graph encapsulate the relationships based totally on co-movement styles and sentiment correlations, enabling our version to grasp the complicated community of influences governing marketplace moves. Complementing this, our LSTM algorithm is trained on sequences of the spatial-temporal illustration discovered through the GCN, enriched with historic fee and extent records. This lets the LSTM seize and expect temporal marketplace developments accurately. Inside the complete assessment of our GCN-LSTM algorithm across the inventory marketplace and cryptocurrency datasets, the version confirmed advanced predictive accuracy and profitability compared to conventional and opportunity machine learning to know benchmarks. Specifically, the model performed a Mean Absolute Error (MAE) of 0.85%, indicating high precision in predicting day-by-day charge movements. The RMSE was recorded at 1.2%, underscoring the model's effectiveness in minimizing tremendous prediction mistakes, which is vital in volatile markets. Furthermore, when assessing the model's predictive performance on directional market movements, it achieved an accuracy rate of 78%, significantly outperforming the benchmark models, averaging an accuracy of 65%. This high degree of accuracy is instrumental for techniques that predict the course of price moves. This study showcases the efficacy of mixing graph-based totally and sequential deep learning knowledge in economic marketplace prediction and highlights the fee of a comprehensive, records-pushed evaluation framework. Our findings promise to revolutionize investment techniques and hazard management practices, offering investors and economic analysts a powerful device to navigate the complexities of cutting-edge economic markets.Keywords: financial market prediction, graph convolutional networks (GCNs), long short-term memory (LSTM), cryptocurrency forecasting
Procedia PDF Downloads 651804 Relationship Between Family Factors and Tendency to Addiction
Authors: Farzaneh Golshekoh
Abstract:
The aim of this study was to examine the relationship between religious beliefs, family responsibility and emotional atmosphere with a tendency to addiction in high school female students in Ahwaz. The sample consisted of 250 students who were selected by cluster random sampling from among all high school female students in Ahvaz. Measuring tools were Iranian tendency towards addiction (IAPS), responsibility California Psychological Inventory (CPI), emotional family atmosphere (AFC) and religious beliefs. The simple correlation coefficient at α=0/05 showed that there is a significant negative relationship between religious beliefs, family responsibility and emotional atmosphere with a tendency to abuse female students. The regression analysis showed that the variables of the emotional atmosphere of the family and religious beliefs as predictors of female students have a tendency to addiction.Keywords: emotional atmosphere, family responsibility, religious beliefs, tendency to addiction
Procedia PDF Downloads 4351803 Investigating the Morphological Patterns of Lip Prints and Their Effectiveness in Individualization and Gender Determination in Pakistani Population
Authors: Makhdoom Saad Wasim Ghouri, Muneeba Butt, Mohammad Ashraf Tahir, Rashid Bhatti, Akbar Ali, Abdul Rehman, Abdul Basit, Muzzamel Rehman, Shahbaz Aslam, Farakh Mansoor, Ahmad Fayyaz, Hadia Siddiqui
Abstract:
Lip print analysis (Cheiloscopy) is the new emerging technique that might be the guardian angel in establishing the personal identity. Cheiloscopy is basically the study of elevations and depressions present on the external surface of the lips. In our study, 600 lip prints samples were taken (300 males and 300 females). Lip prints of each individual were divided into four quadrants and the upper middle portion. For general classification, middle part of the lower lip almost 10 mm wide would be taken into consideration. After analysis of lip-prints, our results show that lip prints are the unique and permanent character of every individual. No two lip print was matched with each other even of the identical twins. Our study reveals that there is equal distribution of lip print patterns among all the four quadrants of lips and the upper middle portion; these distributions were statistically analyzed by applying chi-square test which shows the significant results. In general classification, 5 lip print types/patterns were studied, Type 1 (Vertical lines), Type 2 (Branched pattern), Type 3 (Intersected pattern), Type 4 (Reticular pattern) and Type 5 (Undetermined). Type 1 and Type 2 were found to be the most frequent patterns in female population, while Type 3 and Type 4 most commonly found in male population. These results were also analyzed by applying Chi-square test, and the results show significance statistically. Thus, establishing sex determination on the basis of lip print types among the gender. Type 5 was the least common pattern among genders.Keywords: cheiloscopy, distribution, quadrants, sex determination
Procedia PDF Downloads 2971802 On the Relation between λ-Symmetries and μ-Symmetries of Partial Differential Equations
Authors: Teoman Ozer, Ozlem Orhan
Abstract:
This study deals with symmetry group properties and conservation laws of partial differential equations. We give a geometrical interpretation of notion of μ-prolongations of vector fields and of the related concept of μ-symmetry for partial differential equations. We show that these are in providing symmetry reduction of partial differential equations and systems and invariant solutions.Keywords: λ-symmetry, μ-symmetry, classification, invariant solution
Procedia PDF Downloads 3191801 Prevalence of Malocclusion and Assessment of Orthodontic Treatment Needs in Malay Transfusion-Dependent Thalassemia Patients
Authors: Mohamed H. Kosba, Heba A. Ibrahim, H. Rozita
Abstract:
Statement of the Problem: The life expectancy for transfusion-dependent thalassemia patients has increased dramatically with iron-chelation therapy and other modern management modalities. In these patients, the most dominant maxillofacial manifestations are protrusion of zygomatic bones and premaxilla due to the hyperplasia of bone marrow. The purpose of this study is to determine the prevalence of malocclusion and orthodontic treatment needs according to the Dental Aesthetic Index (DAI) among Malay transfusion-dependent thalassemia patients. Orientation: This is a cross-sectional study consist of 43 Malay transfusion-dependent thalassemia patients, 22 males, and 19 females with the mean age of 15.9 years old (SD 3.58). The subjects were selected randomly from patients attending Paediatrics and Internal Medicine Clinic at Hospital USM and Hospital Sultana Bahiyah. The subjects were assessed for malocclusion according to Angle’s classification, and orthodontic treatment needs using DAI. The results show that 22 of the subjects (51.1%) have class II malocclusion, 12 subjects (28%) have class І, while 9 subjects (20.9%) have class Ⅲ. The assessment of orthodontic treatment needs to reveal 22 cases (51.1%) fall in the normal/minor needs category, 12 subjects (28%) fall in the severe and very severe category, while 9 subjects (20.9%) fall in the definite category. Conclusion & Significance: Half of Malay transfusion-dependent thalassemia patients have Class Ⅱmalocclusion. About 28% had malocclusion and required orthodontic treatment. This research shows that Malay transfusion-dependent thalassemia may require orthodontic management; earlier intervention to reduce the complexity of the treatment later, suggesting functional appliance as a suitable treatment option for them, a twin block appliance together with headgear to restrict maxillary growth suggested for management. The current protocol implemented by the Malaysian Ministry of Health for the management of these patients seems to be sufficient since the result shows that about 28% require orthodontic treatment need, according to DAI.Keywords: prevalence, DAI, thalassaemia, angle classification
Procedia PDF Downloads 1431800 Remote Sensing and GIS Integration for Paddy Production Estimation in Bali Province, Indonesia
Authors: Sarono, Hamim Zaky Hadibasyir, dan Ridho Kurniawan
Abstract:
Estimation of paddy production is one of the areas that can be examined using the techniques of remote sensing and geographic information systems (GIS) in the field of agriculture. The purpose of this research is to know the amount of the paddy production estimation and how remote sensing and geographic information systems (GIS) are able to perform analysis of paddy production estimation in Tegalallang and Payangan Sub district, Bali Province, Indonesia. The method used is the method of land suitability. This method associates a physical parameters which are to be embodied in the smallest unit of a mapping that represents a mapping unit in a particular field and connecting with its field productivity. Analysis of estimated production using standard land suitability from FAO using matching technique. The parameters used to create the land unit is slope (FAO), climate classification (Oldeman), landform (Prapto Suharsono), and soil type. Land use map consist of paddy and non paddy field information obtained from Geo-eye 1 imagery using visual interpretation technique. Landsat image of the Data used for the interpretation of the landform, the classification of the slopes obtained from high point identification with method of interpolation spline, whereas climate data, soil, use secondary data originating from institutions-related institutions. The results of this research indicate Tegallalang and Payangan Districts in known wetland suitability consists of S1 (very suitable) covering an area of 2884,7 ha with the productivity of 5 tons/ha and S2 (suitable) covering an area of 482,9 ha with the productivity of 3 tons/ha. The sum of paddy production estimation as a results in both districts are 31.744, 3 tons in one year.Keywords: production estimation, paddy, remote sensing, geography information system, land suitability
Procedia PDF Downloads 3411799 A Laser Instrument Rapid-E+ for Real-Time Measurements of Airborne Bioaerosols Such as Bacteria, Fungi, and Pollen
Authors: Minghui Zhang, Sirine Fkaier, Sabri Fernana, Svetlana Kiseleva, Denis Kiselev
Abstract:
The real-time identification of bacteria and fungi is difficult because they emit much weaker signals than pollen. In 2020, Plair developed Rapid-E+, which extends abilities of Rapid-E to detect smaller bioaerosols such as bacteria and fungal spores with diameters down to 0.3 µm, while keeping the similar or even better capability for measurements of large bioaerosols like pollen. Rapid-E+ enables simultaneous measurements of (1) time-resolved, polarization and angle dependent Mie scattering patterns, (2) fluorescence spectra resolved in 16 channels, and (3) fluorescence lifetime of individual particles. Moreover, (4) it provides 2D Mie scattering images which give the full information on particle morphology. The parameters of every single bioaerosol aspired into the instrument are subsequently analysed by machine learning. Firstly, pure species of microbes, e.g., Bacillus subtilis (a species of bacteria), and Penicillium chrysogenum (a species of fungal spores), were aerosolized in a bioaerosol chamber for Rapid-E+ training. Afterwards, we tested microbes under different concentrations. We used several steps of data analysis to classify and identify microbes. All single particles were analysed by the parameters of light scattering and fluorescence in the following steps. (1) They were treated with a smart filter block to get rid of non-microbes. (2) By classification algorithm, we verified the filtered particles were microbes based on the calibration data. (3) The probability threshold (defined by the user) step provides the probability of being microbes ranging from 0 to 100%. We demonstrate how Rapid-E+ identified simultaneously microbes based on the results of Bacillus subtilis (bacteria) and Penicillium chrysogenum (fungal spores). By using machine learning, Rapid-E+ achieved identification precision of 99% against the background. The further classification suggests the precision of 87% and 89% for Bacillus subtilis and Penicillium chrysogenum, respectively. The developed algorithm was subsequently used to evaluate the performance of microbe classification and quantification in real-time. The bacteria and fungi were aerosolized again in the chamber with different concentrations. Rapid-E+ can classify different types of microbes and then quantify them in real-time. Rapid-E+ enables classifying different types of microbes and quantifying them in real-time. Rapid-E+ can identify pollen down to species with similar or even better performance than the previous version (Rapid-E). Therefore, Rapid-E+ is an all-in-one instrument which classifies and quantifies not only pollen, but also bacteria and fungi. Based on the machine learning platform, the user can further develop proprietary algorithms for specific microbes (e.g., virus aerosols) and other aerosols (e.g., combustion-related particles that contain polycyclic aromatic hydrocarbons).Keywords: bioaerosols, laser-induced fluorescence, Mie-scattering, microorganisms
Procedia PDF Downloads 901798 Influence of Geologic and Geotechnical Dataset Resolution on Regional Liquefaction Assessment of the Lower Wairau Plains
Authors: Omer Altaf, Liam Wotherspoon, Rolando Orense
Abstract:
The Wairau Plains are located in the northeast of the South Island of New Zealand, with alluvial deposits of fine-grained silts and sands combined with low-lying topography suggesting the presence of liquefiable deposits over significant portions of the region. Liquefaction manifestations were observed in past earthquakes, including the 1848 Marlborough and 1855 Wairarapa earthquakes, and more recently during the 2013 Lake Grassmere and 2016 Kaikōura earthquakes. Therefore, a good understanding of the deposits that may be susceptible to liquefaction is important for land use planning in the region and to allow developers and asset owners to appropriately address their risk. For this purpose, multiple approaches have been employed to develop regional-scale maps showing the liquefaction vulnerability categories for the region. After applying semi-qualitative criteria linked to geologic age and deposit type, the higher resolution surface mapping of geomorphologic characteristics encompassing the Wairau River and the Opaoa River was used for screening. A detailed basin geologic model developed for groundwater modelling was analysed to provide a higher level of resolution than the surface-geology based classification. This is used to identify the thickness of near-surface gravel deposits, providing an improved understanding of the presence or lack of potentially non-liquefiable crust deposits. This paper describes the methodology adopted for this project and focuses on the influence of geomorphic characteristics and analysis of the detailed geologic basin model on the liquefaction classification of the Lower Wairau Plains.Keywords: liquefaction, earthquake, cone penetration test, mapping, liquefaction-induced damage
Procedia PDF Downloads 1761797 Pruning Algorithm for the Minimum Rule Reduct Generation
Authors: Sahin Emrah Amrahov, Fatih Aybar, Serhat Dogan
Abstract:
In this paper we consider the rule reduct generation problem. Rule Reduct Generation (RG) and Modified Rule Generation (MRG) algorithms, that are used to solve this problem, are well-known. Alternative to these algorithms, we develop Pruning Rule Generation (PRG) algorithm. We compare the PRG algorithm with RG and MRG.Keywords: rough sets, decision rules, rule induction, classification
Procedia PDF Downloads 5281796 Intergenerational Class Mobility in Greece: A Cross-Cohort Analysis with Evidence from European Union-Statistics on Income and Living Conditions
Authors: G. Stamatopoulou, M. Symeonaki, C. Michalopoulou
Abstract:
In this work, we study the intergenerational social mobility in Greece, in order to provide up-to-date evidence on the changes in the mobility patterns throughout the years. An analysis for both men and women aged between 25-64 years old is carried out. Three main research objectives are addressed. First, we aim to examine the relationship between the socio-economic status of parents and their children. Secondly, we investigate the evolution of the mobility patterns between different birth cohorts. Finally, the role of education is explored in shaping the mobility patterns. For the analysis, we draw data on both parental and individuals' social outcomes from different national databases. The social class of origins and destination is measured according to the European Socio-Economic Classification (ESeC), while the respondents' educational attainment is coded into categories based on the International Standard Classification of Education (ISCED). Applying the Markov transition probability theory, and a range of measures and models, this work focuses on the magnitude and the direction of the movements that take place in the Greek labour market, as well as the level of social fluidity. Three-way mobility tables are presented, where the transition probabilities between the classes of destination and origins are calculated for different cohorts. Additionally, a range of absolute and relative mobility rates, as well as distance measures, are presented. The study covers a large time span beginning in 1940 until 1995, shedding light on the effects of the national institutional processes on the social movements of individuals. Given the evidence on the mobility patterns of the most recent birth cohorts, we also investigate the possible effects of the 2008 economic crisis.Keywords: cohort analysis, education, Greece, intergenerational mobility, social class
Procedia PDF Downloads 1291795 Relationship between Codependency, Perceived Social Support, and Depression in Mothers of Children with Intellectual Disability
Authors: Sajed Yaghoubnezhad, Mina Karimi, Seyede Marjan Modirkhazeni
Abstract:
The goal of this research was to study the relationship between codependency, perceived social support and depression in mothers of children with intellectual disability (ID). The correlational method was used in this study. The research population is comprised of mothers of educable children with ID in the age range of 25 to 61 years. From among this, a sample of 251 individuals, in the multistage cluster sampling method, was selected from educational districts in Tehran, who responded to the Spann-Fischer Codependency Scale (SFCDS), the Social Support Questionnaire and the Beck Depression Inventory (BDI). The findings of this study indicate that among mothers of children with ID depression has a positive and significant correlation with codependency (P<0.01, r=0.4) and a negative and significant correlation with the total score of social support (P<0.01, r=-0.34). Moreover, the results of stepwise multiple regression analysis showed that codependency is allocated a higher variance than social support in explaining depression (R2=0.023).Keywords: codependency, social support, depression, mothers of children with ID
Procedia PDF Downloads 3681794 A Bayesian Classification System for Facilitating an Institutional Risk Profile Definition
Authors: Roman Graf, Sergiu Gordea, Heather M. Ryan
Abstract:
This paper presents an approach for easy creation and classification of institutional risk profiles supporting endangerment analysis of file formats. The main contribution of this work is the employment of data mining techniques to support set up of the most important risk factors. Subsequently, risk profiles employ risk factors classifier and associated configurations to support digital preservation experts with a semi-automatic estimation of endangerment group for file format risk profiles. Our goal is to make use of an expert knowledge base, accuired through a digital preservation survey in order to detect preservation risks for a particular institution. Another contribution is support for visualisation of risk factors for a requried dimension for analysis. Using the naive Bayes method, the decision support system recommends to an expert the matching risk profile group for the previously selected institutional risk profile. The proposed methods improve the visibility of risk factor values and the quality of a digital preservation process. The presented approach is designed to facilitate decision making for the preservation of digital content in libraries and archives using domain expert knowledge and values of file format risk profiles. To facilitate decision-making, the aggregated information about the risk factors is presented as a multidimensional vector. The goal is to visualise particular dimensions of this vector for analysis by an expert and to define its profile group. The sample risk profile calculation and the visualisation of some risk factor dimensions is presented in the evaluation section.Keywords: linked open data, information integration, digital libraries, data mining
Procedia PDF Downloads 4261793 Food Insecurity Assessment, Consumption Pattern and Implications of Integrated Food Security Phase Classification: Evidence from Sudan
Authors: Ahmed A. A. Fadol, Guangji Tong, Wlaa Mohamed
Abstract:
This paper provides a comprehensive analysis of food insecurity in Sudan, focusing on consumption patterns and their implications, employing the Integrated Food Security Phase Classification (IPC) assessment framework. Years of conflict and economic instability have driven large segments of the population in Sudan into crisis levels of acute food insecurity according to the (IPC). A substantial number of people are estimated to currently face emergency conditions, with an additional sizeable portion categorized under less severe but still extreme hunger levels. In this study, we explore the multifaceted nature of food insecurity in Sudan, considering its historical, political, economic, and social dimensions. An analysis of consumption patterns and trends was conducted, taking into account cultural influences, dietary shifts, and demographic changes. Furthermore, we employ logistic regression and random forest analysis to identify significant independent variables influencing food security status in Sudan. Random forest clearly outperforms logistic regression in terms of area under curve (AUC), accuracy, precision and recall. Forward projections of the IPC for Sudan estimate that 15 million individuals are anticipated to face Crisis level (IPC Phase 3) or worse acute food insecurity conditions between October 2023 and February 2024. Of this, 60% are concentrated in Greater Darfur, Greater Kordofan, and Khartoum State, with Greater Darfur alone representing 29% of this total. These findings emphasize the urgent need for both short-term humanitarian aid and long-term strategies to address Sudan's deepening food insecurity crisis.Keywords: food insecurity, consumption patterns, logistic regression, random forest analysis
Procedia PDF Downloads 731792 Comparati̇ve Study of Pi̇xel and Object-Based Image Classificati̇on Techni̇ques for Extracti̇on of Land Use/Land Cover Informati̇on
Authors: Mahesh Kumar Jat, Manisha Choudhary
Abstract:
Rapid population and economic growth resulted in changes in large-scale land use land cover (LULC) changes. Changes in the biophysical properties of the Earth's surface and its impact on climate are of primary concern nowadays. Different approaches, ranging from location-based relationships or modelling earth surface - atmospheric interaction through modelling techniques like surface energy balance (SEB) have been used in the recent past to examine the relationship between changes in Earth surface land cover and climatic characteristics like temperature and precipitation. A remote sensing-based model i.e., Surface Energy Balance Algorithm for Land (SEBAL), has been used to estimate the surface heat fluxes over Mahi Bajaj Sagar catchment (India) from 2001 to 2020. Landsat ETM and OLI satellite data are used to model the SEB of the area. Changes in observed precipitation and temperature, obtained from India Meteorological Department (IMD) have been correlated with changes in surface heat fluxes to understand the relative contributions of LULC change in changing these climatic variables. Results indicate a noticeable impact of LULC changes on climatic variables, which are aligned with respective changes in SEB components. Results suggest that precipitation increases at a rate of 20 mm/year. The maximum and minimum temperature decreases and increases at 0.007 ℃ /year and 0.02 ℃ /year, respectively. The average temperature increases at 0.009 ℃ /year. Changes in latent heat flux and sensible heat flux positively correlate with precipitation and temperature, respectively. Variation in surface heat fluxes influences the climate parameters and is an adequate reason for climate change. So, SEB modelling is helpful to understand the LULC change and its impact on climate.Keywords: remote sensing, GIS, object based, classification
Procedia PDF Downloads 1301791 An Analysis on Clustering Based Gene Selection and Classification for Gene Expression Data
Authors: K. Sathishkumar, V. Thiagarasu
Abstract:
Due to recent advances in DNA microarray technology, it is now feasible to obtain gene expression profiles of tissue samples at relatively low costs. Many scientists around the world use the advantage of this gene profiling to characterize complex biological circumstances and diseases. Microarray techniques that are used in genome-wide gene expression and genome mutation analysis help scientists and physicians in understanding of the pathophysiological mechanisms, in diagnoses and prognoses, and choosing treatment plans. DNA microarray technology has now made it possible to simultaneously monitor the expression levels of thousands of genes during important biological processes and across collections of related samples. Elucidating the patterns hidden in gene expression data offers a tremendous opportunity for an enhanced understanding of functional genomics. However, the large number of genes and the complexity of biological networks greatly increase the challenges of comprehending and interpreting the resulting mass of data, which often consists of millions of measurements. A first step toward addressing this challenge is the use of clustering techniques, which is essential in the data mining process to reveal natural structures and identify interesting patterns in the underlying data. This work presents an analysis of several clustering algorithms proposed to deals with the gene expression data effectively. The existing clustering algorithms like Support Vector Machine (SVM), K-means algorithm and evolutionary algorithm etc. are analyzed thoroughly to identify the advantages and limitations. The performance evaluation of the existing algorithms is carried out to determine the best approach. In order to improve the classification performance of the best approach in terms of Accuracy, Convergence Behavior and processing time, a hybrid clustering based optimization approach has been proposed.Keywords: microarray technology, gene expression data, clustering, gene Selection
Procedia PDF Downloads 3231790 Power Quality Modeling Using Recognition Learning Methods for Waveform Disturbances
Authors: Sang-Keun Moon, Hong-Rok Lim, Jin-O Kim
Abstract:
This paper presents a Power Quality (PQ) modeling and filtering processes for the distribution system disturbances using recognition learning methods. Typical PQ waveforms with mathematical applications and gathered field data are applied to the proposed models. The objective of this paper is analyzing PQ data with respect to monitoring, discriminating, and evaluating the waveform of power disturbances to ensure the system preventative system failure protections and complex system problem estimations. Examined signal filtering techniques are used for the field waveform noises and feature extractions. Using extraction and learning classification techniques, the efficiency was verified for the recognition of the PQ disturbances with focusing on interactive modeling methods in this paper. The waveform of selected 8 disturbances is modeled with randomized parameters of IEEE 1159 PQ ranges. The range, parameters, and weights are updated regarding field waveform obtained. Along with voltages, currents have same process to obtain the waveform features as the voltage apart from some of ratings and filters. Changing loads are causing the distortion in the voltage waveform due to the drawing of the different patterns of current variation. In the conclusion, PQ disturbances in the voltage and current waveforms indicate different types of patterns of variations and disturbance, and a modified technique based on the symmetrical components in time domain was proposed in this paper for the PQ disturbances detection and then classification. Our method is based on the fact that obtained waveforms from suggested trigger conditions contain potential information for abnormality detections. The extracted features are sequentially applied to estimation and recognition learning modules for further studies.Keywords: power quality recognition, PQ modeling, waveform feature extraction, disturbance trigger condition, PQ signal filtering
Procedia PDF Downloads 1861789 Airport Pavement Crack Measurement Systems and Crack Density for Pavement Evaluation
Authors: Ali Ashtiani, Hamid Shirazi
Abstract:
This paper reviews the status of existing practice and research related to measuring pavement cracking and using crack density as a pavement surface evaluation protocol. Crack density for pavement evaluation is currently not widely used within the airport community and its use by the highway community is limited. However, surface cracking is a distress that is closely monitored by airport staff and significantly influences the development of maintenance, rehabilitation and reconstruction plans for airport pavements. Therefore crack density has the potential to become an important indicator of pavement condition if the type, severity and extent of surface cracking can be accurately measured. A pavement distress survey is an essential component of any pavement assessment. Manual crack surveying has been widely used for decades to measure pavement performance. However, the accuracy and precision of manual surveys can vary depending upon the surveyor and performing surveys may disrupt normal operations. Given the variability of manual surveys, this method has shown inconsistencies in distress classification and measurement. This can potentially impact the planning for pavement maintenance, rehabilitation and reconstruction and the associated funding strategies. A substantial effort has been devoted for the past 20 years to reduce the human intervention and the error associated with it by moving toward automated distress collection methods. The automated methods refer to the systems that identify, classify and quantify pavement distresses through processes that require no or very minimal human intervention. This principally involves the use of a digital recognition software to analyze and characterize pavement distresses. The lack of established protocols for measurement and classification of pavement cracks captured using digital images is a challenge to developing a reliable automated system for distress assessment. Variations in types and severity of distresses, different pavement surface textures and colors and presence of pavement joints and edges all complicate automated image processing and crack measurement and classification. This paper summarizes the commercially available systems and technologies for automated pavement distress evaluation. A comprehensive automated pavement distress survey involves collection, interpretation, and processing of the surface images to identify the type, quantity and severity of the surface distresses. The outputs can be used to quantitatively calculate the crack density. The systems for automated distress survey using digital images reviewed in this paper can assist the airport industry in the development of a pavement evaluation protocol based on crack density. Analysis of automated distress survey data can lead to a crack density index. This index can be used as a means of assessing pavement condition and to predict pavement performance. This can be used by airport owners to determine the type of pavement maintenance and rehabilitation in a more consistent way.Keywords: airport pavement management, crack density, pavement evaluation, pavement management
Procedia PDF Downloads 185