Search results for: real time data processing
38533 Machine Learning Facing Behavioral Noise Problem in an Imbalanced Data Using One Side Behavioral Noise Reduction: Application to a Fraud Detection
Authors: Salma El Hajjami, Jamal Malki, Alain Bouju, Mohammed Berrada
Abstract:
With the expansion of machine learning and data mining in the context of Big Data analytics, the common problem that affects data is class imbalance. It refers to an imbalanced distribution of instances belonging to each class. This problem is present in many real world applications such as fraud detection, network intrusion detection, medical diagnostics, etc. In these cases, data instances labeled negatively are significantly more numerous than the instances labeled positively. When this difference is too large, the learning system may face difficulty when tackling this problem, since it is initially designed to work in relatively balanced class distribution scenarios. Another important problem, which usually accompanies these imbalanced data, is the overlapping instances between the two classes. It is commonly referred to as noise or overlapping data. In this article, we propose an approach called: One Side Behavioral Noise Reduction (OSBNR). This approach presents a way to deal with the problem of class imbalance in the presence of a high noise level. OSBNR is based on two steps. Firstly, a cluster analysis is applied to groups similar instances from the minority class into several behavior clusters. Secondly, we select and eliminate the instances of the majority class, considered as behavioral noise, which overlap with behavior clusters of the minority class. The results of experiments carried out on a representative public dataset confirm that the proposed approach is efficient for the treatment of class imbalances in the presence of noise.Keywords: machine learning, imbalanced data, data mining, big data
Procedia PDF Downloads 13038532 A Smartphone-Based Real-Time Activity Recognition and Fall Detection System
Authors: Manutchanok Jongprasithporn, Rawiphorn Srivilai, Paweena Pongsopha
Abstract:
Fall is the most serious accident leading to increased unintentional injuries and mortality. Falls are not only the cause of suffering and functional impairments to the individuals, but also the cause of increasing medical cost and days away from work. The early detection of falls could be an advantage to reduce fall-related injuries and consequences of falls. Smartphones, embedded accelerometer, have become a common device in everyday life due to decreasing technology cost. This paper explores a physical activity monitoring and fall detection application in smartphones which is a non-invasive biomedical device to determine physical activities and fall event. The combination of application and sensors could perform as a biomedical sensor to monitor physical activities and recognize a fall. We have chosen Android-based smartphone in this study since android operating system is an open-source and no cost. Moreover, android phone users become a majority of Thai’s smartphone users. We developed Thai 3 Axis (TH3AX) as a physical activities and fall detection application which included command, manual, results in Thai language. The smartphone was attached to right hip of 10 young, healthy adult subjects (5 males, 5 females; aged< 35y) to collect accelerometer and gyroscope data during performing physical activities (e.g., walking, running, sitting, and lying down) and falling to determine threshold for each activity. Dependent variables are including accelerometer data (acceleration, peak acceleration, average resultant acceleration, and time between peak acceleration). A repeated measures ANOVA was performed to test whether there are any differences between DVs’ means. Statistical analyses were considered significant at p<0.05. After finding threshold, the results were used as training data for a predictive model of activity recognition. In the future, accuracies of activity recognition will be performed to assess the overall performance of the classifier. Moreover, to help improve the quality of life, our system will be implemented with patients and elderly people who need intensive care in hospitals and nursing homes in Thailand.Keywords: activity recognition, accelerometer, fall, gyroscope, smartphone
Procedia PDF Downloads 69238531 Ensemble Methods in Machine Learning: An Algorithmic Approach to Derive Distinctive Behaviors of Criminal Activity Applied to the Poaching Domain
Authors: Zachary Blanks, Solomon Sonya
Abstract:
Poaching presents a serious threat to endangered animal species, environment conservations, and human life. Additionally, some poaching activity has even been linked to supplying funds to support terrorist networks elsewhere around the world. Consequently, agencies dedicated to protecting wildlife habitats have a near intractable task of adequately patrolling an entire area (spanning several thousand kilometers) given limited resources, funds, and personnel at their disposal. Thus, agencies need predictive tools that are both high-performing and easily implementable by the user to help in learning how the significant features (e.g. animal population densities, topography, behavior patterns of the criminals within the area, etc) interact with each other in hopes of abating poaching. This research develops a classification model using machine learning algorithms to aid in forecasting future attacks that is both easy to train and performs well when compared to other models. In this research, we demonstrate how data imputation methods (specifically predictive mean matching, gradient boosting, and random forest multiple imputation) can be applied to analyze data and create significant predictions across a varied data set. Specifically, we apply these methods to improve the accuracy of adopted prediction models (Logistic Regression, Support Vector Machine, etc). Finally, we assess the performance of the model and the accuracy of our data imputation methods by learning on a real-world data set constituting four years of imputed data and testing on one year of non-imputed data. This paper provides three main contributions. First, we extend work done by the Teamcore and CREATE (Center for Risk and Economic Analysis of Terrorism Events) research group at the University of Southern California (USC) working in conjunction with the Department of Homeland Security to apply game theory and machine learning algorithms to develop more efficient ways of reducing poaching. This research introduces ensemble methods (Random Forests and Stochastic Gradient Boosting) and applies it to real-world poaching data gathered from the Ugandan rain forest park rangers. Next, we consider the effect of data imputation on both the performance of various algorithms and the general accuracy of the method itself when applied to a dependent variable where a large number of observations are missing. Third, we provide an alternate approach to predict the probability of observing poaching both by season and by month. The results from this research are very promising. We conclude that by using Stochastic Gradient Boosting to predict observations for non-commercial poaching by season, we are able to produce statistically equivalent results while being orders of magnitude faster in computation time and complexity. Additionally, when predicting potential poaching incidents by individual month vice entire seasons, boosting techniques produce a mean area under the curve increase of approximately 3% relative to previous prediction schedules by entire seasons.Keywords: ensemble methods, imputation, machine learning, random forests, statistical analysis, stochastic gradient boosting, wildlife protection
Procedia PDF Downloads 29238530 Clinical Training Simulation Experience of Medical Sector Students
Authors: Tahsien Mohamed Okasha
Abstract:
Simulation is one of the emerging educational strategies that depend on the creation of scenarios to imitate what could happen in real life. At the time of COVID, we faced big obstacles in medical education, specially the clinical part and how we could apply it, the simulation was the golden key. Simulation is a very important tool of education for medical sector students, through creating a safe, changeable, quiet environment with less anxiety level for students to practice and to have repeated trials on their competencies. That impacts the level of practice, achievement, and the way of acting in real situations and experiences. A blind Random sample of students from different specialties and colleges who came and finished their training in an integrated environment was collected and tested, and the responses were graded from (1-5). The results revealed that 77% of the studied subjects agreed that dealing and interacting with different medical sector candidates in the same place was beneficial. 77% of the studied subjects agreed that simulations were challenging in thinking and decision-making skills .75% agreed that using high-fidelity manikins was helpful. 75% agree .76% agreed that working in a safe, prepared environment is helpful for realistic situations.Keywords: simulation, clinical training, education, medical sector students
Procedia PDF Downloads 3138529 Sequential Data Assimilation with High-Frequency (HF) Radar Surface Current
Authors: Lei Ren, Michael Hartnett, Stephen Nash
Abstract:
The abundant measured surface current from HF radar system in coastal area is assimilated into model to improve the modeling forecasting ability. A simple sequential data assimilation scheme, Direct Insertion (DI), is applied to update model forecast states. The influence of Direct Insertion data assimilation over time is analyzed at one reference point. Vector maps of surface current from models are compared with HF radar measurements. Root-Mean-Squared-Error (RMSE) between modeling results and HF radar measurements is calculated during the last four days with no data assimilation.Keywords: data assimilation, CODAR, HF radar, surface current, direct insertion
Procedia PDF Downloads 57538528 Experiences of Timing Analysis of Parallel Embedded Software
Authors: Muhammad Waqar Aziz, Syed Abdul Baqi Shah
Abstract:
The execution time analysis is fundamental to the successful design and execution of real-time embedded software. In such analysis, the Worst-Case Execution Time (WCET) of a program is a key measure, on the basis of which system tasks are scheduled. The WCET analysis of embedded software is also needed for system understanding and to guarantee its behavior. WCET analysis can be performed statically (without executing the program) or dynamically (through measurement). Traditionally, research on the WCET analysis assumes sequential code running on single-core platforms. However, as computation is steadily moving towards using a combination of parallel programs and multi-core hardware, new challenges in WCET analysis need to be addressed. In this article, we report our experiences of performing the WCET analysis of Parallel Embedded Software (PES) running on multi-core platform. The primary purpose was to investigate how WCET estimates of PES can be computed statically, and how they can be derived dynamically. Our experiences, as reported in this article, include the challenges we faced, possible suggestions to these challenges and the workarounds that were developed. This article also provides observations on the benefits and drawbacks of deriving the WCET estimates using the said methods and provides useful recommendations for further research in this area.Keywords: embedded software, worst-case execution-time analysis, static flow analysis, measurement-based analysis, parallel computing
Procedia PDF Downloads 32438527 Methodology for the Multi-Objective Analysis of Data Sets in Freight Delivery
Authors: Dale Dzemydiene, Aurelija Burinskiene, Arunas Miliauskas, Kristina Ciziuniene
Abstract:
Data flow and the purpose of reporting the data are different and dependent on business needs. Different parameters are reported and transferred regularly during freight delivery. This business practices form the dataset constructed for each time point and contain all required information for freight moving decisions. As a significant amount of these data is used for various purposes, an integrating methodological approach must be developed to respond to the indicated problem. The proposed methodology contains several steps: (1) collecting context data sets and data validation; (2) multi-objective analysis for optimizing freight transfer services. For data validation, the study involves Grubbs outliers analysis, particularly for data cleaning and the identification of statistical significance of data reporting event cases. The Grubbs test is often used as it measures one external value at a time exceeding the boundaries of standard normal distribution. In the study area, the test was not widely applied by authors, except when the Grubbs test for outlier detection was used to identify outsiders in fuel consumption data. In the study, the authors applied the method with a confidence level of 99%. For the multi-objective analysis, the authors would like to select the forms of construction of the genetic algorithms, which have more possibilities to extract the best solution. For freight delivery management, the schemas of genetic algorithms' structure are used as a more effective technique. Due to that, the adaptable genetic algorithm is applied for the description of choosing process of the effective transportation corridor. In this study, the multi-objective genetic algorithm methods are used to optimize the data evaluation and select the appropriate transport corridor. The authors suggest a methodology for the multi-objective analysis, which evaluates collected context data sets and uses this evaluation to determine a delivery corridor for freight transfer service in the multi-modal transportation network. In the multi-objective analysis, authors include safety components, the number of accidents a year, and freight delivery time in the multi-modal transportation network. The proposed methodology has practical value in the management of multi-modal transportation processes.Keywords: multi-objective, analysis, data flow, freight delivery, methodology
Procedia PDF Downloads 18038526 Effect of Local Processing Techniques on the Nutrients and Anti-Nutrients Content of Bitter Cassava (Manihot Esculenta Crantz)
Authors: J. S. Alakali, A. R. Ismaila, T. G. Atume
Abstract:
The effects of local processing techniques on the nutrients and anti-nutrients content of bitter cassava were investigated. Raw bitter cassava tubers were boiled, sundried, roasted, fried to produce Kuese, partially fermented and sun dried to produce Alubo, fermented by submersion to produce Akpu and fermented by solid state to produce yellow and white gari. These locally processed cassava products were subjected to proximate, mineral analysis and anti-nutrient analysis using standard methods. The result of the proximate analysis showed that, raw bitter cassava is composed of 1.85% ash, 20.38% moisture, 4.11% crude fibre, 1.03% crude protein, 0.66% lipids and 71.88% total carbohydrate. For the mineral analysis, the raw bitter cassava tuber contained 32.00% Calcium, 12.55% Magnesium, 1.38% Iron and 80.17% Phosphorous. Even though all processing techniques significantly increased the mineral content, fermentation had higher mineral increment effect. The anti-nutrients analysis showed that the raw tuber contained 98.16mg/100g cyanide, 44.00mg/100g oxalate 304.20mg/100g phytate and 73.00mg/100g saponin. In general all the processing techniques showed a significant reduction of the phytate, oxalate and saponin content of the cassava. However, only fermentation, sun drying and gasification were able to reduce the cyanide content of bitter cassava below the safe level (10mg/100g) recommended by Standard Organization of Nigeria. Yellow gari(with the addition of palm oil) showed low cyanide content (1.10 mg/100g) than white gari (3.51 mg/100g). Processing methods involving fermentation reduce cyanide and other anti-nutrients in the cassava to levels that are safe for consumption and should be widely practiced.Keywords: bitter cassava, local processing, fermentation, anti-nutrient.
Procedia PDF Downloads 30438525 Microarray Data Visualization and Preprocessing Using R and Bioconductor
Authors: Ruchi Yadav, Shivani Pandey, Prachi Srivastava
Abstract:
Microarrays provide a rich source of data on the molecular working of cells. Each microarray reports on the abundance of tens of thousands of mRNAs. Virtually every human disease is being studied using microarrays with the hope of finding the molecular mechanisms of disease. Bioinformatics analysis plays an important part of processing the information embedded in large-scale expression profiling studies and for laying the foundation for biological interpretation. A basic, yet challenging task in the analysis of microarray gene expression data is the identification of changes in gene expression that are associated with particular biological conditions. Careful statistical design and analysis are essential to improve the efficiency and reliability of microarray experiments throughout the data acquisition and analysis process. One of the most popular platforms for microarray analysis is Bioconductor, an open source and open development software project based on the R programming language. This paper describes specific procedures for conducting quality assessment, visualization and preprocessing of Affymetrix Gene Chip and also details the different bioconductor packages used to analyze affymetrix microarray data and describe the analysis and outcome of each plots.Keywords: microarray analysis, R language, affymetrix visualization, bioconductor
Procedia PDF Downloads 48038524 An Event-Related Potential Study of Individual Differences in Word Recognition: The Evidence from Morphological Knowledge of Sino-Korean Prefixes
Authors: Jinwon Kang, Seonghak Jo, Joohee Ahn, Junghye Choi, Sun-Young Lee
Abstract:
A morphological priming has proved its importance by showing that segmentation occurs in morphemes when visual words are recognized within a noticeably short time. Regarding Sino-Korean prefixes, this study conducted an experiment on visual masked priming tasks with 57 ms stimulus-onset asynchrony (SOA) to see how individual differences in the amount of morphological knowledge affect morphological priming. The relationship between the prime and target words were classified as morphological (e.g., 미개척 migaecheog [unexplored] – 미해결 mihaegyel [unresolved]), semantical (e.g., 친환경 chinhwangyeong [eco-friendly]) – 무공해 mugonghae [no-pollution]), and orthographical (e.g., 미용실 miyongsil [beauty shop] – 미확보 mihwagbo [uncertainty]) conditions. We then compared the priming by configuring irrelevant paired stimuli for each condition’s control group. As a result, in the behavioral data, we observed facilitatory priming from a group with high morphological knowledge only under the morphological condition. In contrast, a group with low morphological knowledge showed the priming only under the orthographic condition. In the event-related potential (ERP) data, the group with high morphological knowledge presented the N250 only under the morphological condition. The findings of this study imply that individual differences in morphological knowledge in Korean may have a significant influence on the segmental processing of Korean word recognition.Keywords: ERP, individual differences, morphological priming, sino-Korean prefixes
Procedia PDF Downloads 21538523 Deep-Learning Coupled with Pragmatic Categorization Method to Classify the Urban Environment of the Developing World
Authors: Qianwei Cheng, A. K. M. Mahbubur Rahman, Anis Sarker, Abu Bakar Siddik Nayem, Ovi Paul, Amin Ahsan Ali, M. Ashraful Amin, Ryosuke Shibasaki, Moinul Zaber
Abstract:
Thomas Friedman, in his famous book, argued that the world in this 21st century is flat and will continue to be flatter. This is attributed to rapid globalization and the interdependence of humanity that engendered tremendous in-flow of human migration towards the urban spaces. In order to keep the urban environment sustainable, policy makers need to plan based on extensive analysis of the urban environment. With the advent of high definition satellite images, high resolution data, computational methods such as deep neural network analysis, and hardware capable of high-speed analysis; urban planning is seeing a paradigm shift. Legacy data on urban environments are now being complemented with high-volume, high-frequency data. However, the first step of understanding urban space lies in useful categorization of the space that is usable for data collection, analysis, and visualization. In this paper, we propose a pragmatic categorization method that is readily usable for machine analysis and show applicability of the methodology on a developing world setting. Categorization to plan sustainable urban spaces should encompass the buildings and their surroundings. However, the state-of-the-art is mostly dominated by classification of building structures, building types, etc. and largely represents the developed world. Hence, these methods and models are not sufficient for developing countries such as Bangladesh, where the surrounding environment is crucial for the categorization. Moreover, these categorizations propose small-scale classifications, which give limited information, have poor scalability and are slow to compute in real time. Our proposed method is divided into two steps-categorization and automation. We categorize the urban area in terms of informal and formal spaces and take the surrounding environment into account. 50 km × 50 km Google Earth image of Dhaka, Bangladesh was visually annotated and categorized by an expert and consequently a map was drawn. The categorization is based broadly on two dimensions-the state of urbanization and the architectural form of urban environment. Consequently, the urban space is divided into four categories: 1) highly informal area; 2) moderately informal area; 3) moderately formal area; and 4) highly formal area. In total, sixteen sub-categories were identified. For semantic segmentation and automatic categorization, Google’s DeeplabV3+ model was used. The model uses Atrous convolution operation to analyze different layers of texture and shape. This allows us to enlarge the field of view of the filters to incorporate larger context. Image encompassing 70% of the urban space was used to train the model, and the remaining 30% was used for testing and validation. The model is able to segment with 75% accuracy and 60% Mean Intersection over Union (mIoU). In this paper, we propose a pragmatic categorization method that is readily applicable for automatic use in both developing and developed world context. The method can be augmented for real-time socio-economic comparative analysis among cities. It can be an essential tool for the policy makers to plan future sustainable urban spaces.Keywords: semantic segmentation, urban environment, deep learning, urban building, classification
Procedia PDF Downloads 19138522 Multimodal Employee Attendance Management System
Authors: Khaled Mohammed
Abstract:
This paper presents novel face recognition and identification approaches for the real-time attendance management problem in large companies/factories and government institutions. The proposed uses the Minimum Ratio (MR) approach for employee identification. Capturing the authentic face variability from a sequence of video frames has been considered for the recognition of faces and resulted in system robustness against the variability of facial features. Experimental results indicated an improvement in the performance of the proposed system compared to the Previous approaches at a rate between 2% to 5%. In addition, it decreased the time two times if compared with the Previous techniques, such as Extreme Learning Machine (ELM) & Multi-Scale Structural Similarity index (MS-SSIM). Finally, it achieved an accuracy of 99%.Keywords: attendance management system, face detection and recognition, live face recognition, minimum ratio
Procedia PDF Downloads 15538521 Isolation and Classification of Red Blood Cells in Anemic Microscopic Images
Authors: Jameela Ali Alkrimi, Abdul Rahim Ahmad, Azizah Suliman, Loay E. George
Abstract:
Red blood cells (RBCs) are among the most commonly and intensively studied type of blood cells in cell biology. The lack of RBCs is a condition characterized by lower than normal hemoglobin level; this condition is referred to as 'anemia'. In this study, a software was developed to isolate RBCs by using a machine learning approach to classify anemic RBCs in microscopic images. Several features of RBCs were extracted using image processing algorithms, including principal component analysis (PCA). With the proposed method, RBCs were isolated in 34 second from an image containing 18 to 27 cells. We also proposed that PCA could be performed to increase the speed and efficiency of classification. Our classifier algorithm yielded accuracy rates of 100%, 99.99%, and 96.50% for K-nearest neighbor (K-NN) algorithm, support vector machine (SVM), and neural network ANN, respectively. Classification was evaluated in highly sensitivity, specificity, and kappa statistical parameters. In conclusion, the classification results were obtained for a short time period with more efficient when PCA was used.Keywords: red blood cells, pre-processing image algorithms, classification algorithms, principal component analysis PCA, confusion matrix, kappa statistical parameters, ROC
Procedia PDF Downloads 40538520 The Food Industry in Nigeria: Development and Quality Assurance
Authors: Agi Sunday, Agih Ukuru Agih
Abstract:
In Nigeria, the food processing sector is dominated by small and medium enterprises, as well as multinational food companies. Quality standards are usually related to improving the safety of food products suitable for consumption in accordance to specifications by food regulatory bodies. These standards are essential elements for local and international businesses which contribute to economic progress through industrial development and trade. This review takes a critical look on the Nigerian food industry development in terms of quality standards that are necessary to be given consideration in the production of food and also ways of improving food production in Nigeria through the use of Total Quality Management (TQM) technique and the use of computerized systems to produce high quality and high value products while at the same time reducing production time and cost.Keywords: food industry, quality assurance, Nigeria, TQM, computerized systems
Procedia PDF Downloads 45538519 Barriers towards Effective Participation in Physically Oriented Leisure Time Activities: A Case Study of Federal College of Education, Pankshin Plateau State, Nigeria
Authors: Mulak Moses Yokdi
Abstract:
Correct use of leisure time has suffered neglect in our society and the people ignorantly think that the trend does not matter. The researcher felt concerned about the issue and went on to find out why using FCE, Pankshin workers as a case study. Four hypotheses were used, considering such variables as leadership, traditional activities, stress due to work pressure and time constraint. The participants selected for the study were one hundred and ten members of FCE, Pankshin staff. A self-developed questionnaire was the instrument used. Chi-square (x2) was employed to test the hypotheses at P = 0.005; df = 3. The statistics of percentages was also used to describe the situation as implicated by the data. The results showed that all hypotheses were significant (P = 0.05). It was concluded that the four variables were impediments to effective participation in physically oriented leisure time activities among the FCE, Staff. Based on the findings, it was recommended that the FCE should get good leadership, create good awareness for people to understand why they should be effectively involved in physically oriented leisure time activities.Keywords: barriers, effective participation, leisure time, physically oriented, work pressure, time constraint
Procedia PDF Downloads 36838518 Effect of Temperature and Time on the Yield of Silica from Rice Husk Ash
Authors: Mohammed Adamu Musa, Shehu Saminu Babba
Abstract:
The technological trend towards waste utilization and cost reduction in industrial processing has attracted use of Rice Husk as a value added material. Both rice husk (RH) and Rice Husk Ash (RHA) has been found suitable for wide range of domestic as well as industrial applications. Therefore, the purpose of this research is to produce high grade sodium silicate from rice husk ash by considering the effect of temperature and time of heating as the process variables. The experiment was performed by heating the rice husk at temperatures 500 °C, 600 °C, 700 °C and 800 °C and time 60min, 90min, 120min and 150min were used to obtain the ash. 1.0M of aqueous sodium hydroxide solution was used to dissolve the silicate from the ash, which contained crude sodium silicate. In addition, the ash was neutralized by adding 5M of HCL until the pH reached 3.5 to give silica gel. At 6000C and 120mins, 94.23% silica was obtained from the RHA. At higher temperatures (700 °C and 800 °C) the percentage yield of silica reduced due to surface melting and carbon fixation in the lattice caused by presence of potassium. For this research, 600 °C is considered to be the optimum temperature for silica production from RHA. Silica produced from RHA can generate aggregate value and can be used in areas such as pulp and paper, plastic and rubber reinforcement industries.Keywords: burning, rice husk, rice husk ash, silica, silica gel, temperature
Procedia PDF Downloads 24338517 Unbalanced Mean-Time and Buffer Effects in Lines Suffering Breakdown
Authors: Sabry Shaaban, Tom McNamara, Sarah Hudson
Abstract:
This article studies the performance of unpaced serial production lines that are subject to breakdown and are imbalanced in terms of both of their processing time means (MTs) and buffer storage capacities (BCs). Simulation results show that the best pattern in terms of throughput is a balanced line with respect to average buffer level; the best configuration is a monotone decreasing MT order, together with an ascending BC arrangement. Statistical analysis shows that BC, patterns of MT and BC imbalance, line length and degree of imbalance all contribute significantly to performance. Results show that unbalanced lines cope well with unreliability.Keywords: unreliable unpaced serial lines, simulation, unequal mean operation times, uneven buffer capacities, patterns of imbalance, throughput, average buffer level
Procedia PDF Downloads 47338516 Reducing the Computational Overhead of Metaheuristics Parameterization with Exploratory Landscape Analysis
Authors: Iannick Gagnon, Alain April
Abstract:
The performance of a metaheuristic on a given problem class depends on the class itself and the choice of parameters. Parameter tuning is the most time-consuming phase of the optimization process after the main calculations and it often nullifies the speed advantage of metaheuristics over traditional optimization algorithms. Several off-the-shelf parameter tuning algorithms are available, but when the objective function is expensive to evaluate, these can be prohibitively expensive to use. This paper presents a surrogate-like method for finding adequate parameters using fitness landscape analysis on simple benchmark functions and real-world objective functions. The result is a simple compound similarity metric based on the empirical correlation coefficient and a measure of convexity. It is then used to find the best benchmark functions to serve as surrogates. The near-optimal parameter set is then found using fractional factorial design. The real-world problem of NACA airfoil lift coefficient maximization is used as a preliminary proof of concept. The overall aim of this research is to reduce the computational overhead of metaheuristics parameterization.Keywords: metaheuristics, stochastic optimization, particle swarm optimization, exploratory landscape analysis
Procedia PDF Downloads 15338515 Comparison of Applicability of Time Series Forecasting Models VAR, ARCH and ARMA in Management Science: Study Based on Empirical Analysis of Time Series Techniques
Authors: Muhammad Tariq, Hammad Tahir, Fawwad Mahmood Butt
Abstract:
Purpose: This study attempts to examine the best forecasting methodologies in the time series. The time series forecasting models such as VAR, ARCH and the ARMA are considered for the analysis. Methodology: The Bench Marks or the parameters such as Adjusted R square, F-stats, Durban Watson, and Direction of the roots have been critically and empirically analyzed. The empirical analysis consists of time series data of Consumer Price Index and Closing Stock Price. Findings: The results show that the VAR model performed better in comparison to other models. Both the reliability and significance of VAR model is highly appreciable. In contrary to it, the ARCH model showed very poor results for forecasting. However, the results of ARMA model appeared double standards i.e. the AR roots showed that model is stationary and that of MA roots showed that the model is invertible. Therefore, the forecasting would remain doubtful if it made on the bases of ARMA model. It has been concluded that VAR model provides best forecasting results. Practical Implications: This paper provides empirical evidences for the application of time series forecasting model. This paper therefore provides the base for the application of best time series forecasting model.Keywords: forecasting, time series, auto regression, ARCH, ARMA
Procedia PDF Downloads 34838514 Unsupervised Learning and Similarity Comparison of Water Mass Characteristics with Gaussian Mixture Model for Visualizing Ocean Data
Authors: Jian-Heng Wu, Bor-Shen Lin
Abstract:
The temperature-salinity relationship is one of the most important characteristics used for identifying water masses in marine research. Temperature-salinity characteristics, however, may change dynamically with respect to the geographic location and is quite sensitive to the depth at the same location. When depth is taken into consideration, however, it is not easy to compare the characteristics of different water masses efficiently for a wide range of areas of the ocean. In this paper, the Gaussian mixture model was proposed to analyze the temperature-salinity-depth characteristics of water masses, based on which comparison between water masses may be conducted. Gaussian mixture model could model the distribution of a random vector and is formulated as the weighting sum for a set of multivariate normal distributions. The temperature-salinity-depth data for different locations are first used to train a set of Gaussian mixture models individually. The distance between two Gaussian mixture models can then be defined as the weighting sum of pairwise Bhattacharyya distances among the Gaussian distributions. Consequently, the distance between two water masses may be measured fast, which allows the automatic and efficient comparison of the water masses for a wide range area. The proposed approach not only can approximate the distribution of temperature, salinity, and depth directly without the prior knowledge for assuming the regression family, but may restrict the complexity by controlling the number of mixtures when the amounts of samples are unevenly distributed. In addition, it is critical for knowledge discovery in marine research to represent, manage and share the temperature-salinity-depth characteristics flexibly and responsively. The proposed approach has been applied to a real-time visualization system of ocean data, which may facilitate the comparison of water masses by aggregating the data without degrading the discriminating capabilities. This system provides an interface for querying geographic locations with similar temperature-salinity-depth characteristics interactively and for tracking specific patterns of water masses, such as the Kuroshio near Taiwan or those in the South China Sea.Keywords: water mass, Gaussian mixture model, data visualization, system framework
Procedia PDF Downloads 14538513 Increasing a Computer Performance by Overclocking Central Processing Unit (CPU)
Authors: Witthaya Mekhum, Wutthikorn Malikong
Abstract:
The objective of this study is to investigate the increasing desktop computer performance after overclocking central processing unit or CPU by running a computer component at a higher clock rate (more clock cycles per second) than it was designed at the rate of 0.1 GHz for each level or 100 MHz starting at 4000 GHz-4500 GHz. The computer performance is tested for each level with 4 programs, i.e. Hyper PI ver. 0.99b, Cinebench R15, LinX ver.0.6.4 and WinRAR . After the CPU overclock, the computer performance increased. When overclocking CPU at 29% the computer performance tested by Hyper PI ver. 0.99b increased by 10.03% and when tested by Cinebench R15 the performance increased by 20.05% and when tested by LinX Program the performance increased by 16.61%. However, the performance increased only 8.14% when tested with Winrar program. The computer performance did not increase according to the overclock rate because the computer consists of many components such as Random Access Memory or RAM, Hard disk Drive, Motherboard and Display Card, etc.Keywords: overclock, performance, central processing unit, computer
Procedia PDF Downloads 28338512 Optimizing CNC Production Line Efficiency Using NSGA-II: Adaptive Layout and Operational Sequence for Enhanced Manufacturing Flexibility
Authors: Yi-Ling Chen, Dung-Ying Lin
Abstract:
In the manufacturing process, computer numerical control (CNC) machining plays a crucial role. CNC enables precise machinery control through computer programs, achieving automation in the production process and significantly enhancing production efficiency. However, traditional CNC production lines often require manual intervention for loading and unloading operations, which limits the production line's operational efficiency and production capacity. Additionally, existing CNC automation systems frequently lack sufficient intelligence and fail to achieve optimal configuration efficiency, resulting in the need for substantial time to reconfigure production lines when producing different products, thereby impacting overall production efficiency. Using the NSGA-II algorithm, we generate production line layout configurations that consider field constraints and select robotic arm specifications from an arm list. This allows us to calculate loading and unloading times for each job order, perform demand allocation, and assign processing sequences. The NSGA-II algorithm is further employed to determine the optimal processing sequence, with the aim of minimizing demand completion time and maximizing average machine utilization. These objectives are used to evaluate the performance of each layout, ultimately determining the optimal layout configuration. By employing this method, it enhance the configuration efficiency of CNC production lines and establish an adaptive capability that allows the production line to respond promptly to changes in demand. This will minimize production losses caused by the need to reconfigure the layout, ensuring that the CNC production line can maintain optimal efficiency even when adjustments are required due to fluctuating demands.Keywords: evolutionary algorithms, multi-objective optimization, pareto optimality, layout optimization, operations sequence
Procedia PDF Downloads 2138511 Gene Expression Profiling of Iron-Related Genes of Pasteurella multocida Serotype A Strain PMTB2.1
Authors: Shagufta Jabeen, Faez Jesse Firdaus Abdullah, Zunita Zakaria, Nurulfiza Mat Isa, Yung Chie Tan, Wai Yan Yee, Abdul Rahman Omar
Abstract:
Pasteurella multocida is associated with acute, as well as, chronic infections in avian and bovine such as pasteurellosis and hemorrhagic septicemia (HS) in cattle and buffaloes. Iron is one of the most important nutrients for pathogenic bacteria including Pasteurella and acts as a cofactor or prosthetic group in several essential enzymes and is needed for amino acid, pyrimidine, and DNA biosynthesis. In our recent study, we showed that 2% of Pasteurella multocida serotype A strain PMTB2.1 encode for iron regulating genes (Accession number CP007205.1). Genome sequencing of other Pasteurella multocida serotypes namely PM70 and HB01 also indicated up to 2.5% of the respective genome encode for iron regulating genes, suggesting that Pasteurella multocida genome comprises of multiple systems for iron uptake. Since P. multocida PMTB2.1 has more than 40 CDs out of 2097 CDs (approximately 2%), encode for iron-regulated. The gene expression profiling of four iron-regulating genes namely fbpb, yfea, fece and fur were characterized under iron-restricted environment. The P. multocida strain PMTB2.1 was grown in broth with and without iron chelating agent and samples were collected at different time points. Relative mRNA expression profile of these genes was determined using Taqman probe based real-time PCR assay. The data analysis, normalization with two house-keeping genes and the quantification of fold changes were carried out using Bio-Rad CFX manager software version 3.1. Results of this study reflect that iron reduced environment has significant effect on expression profile of iron regulating genes (p < 0.05) when compared to control (normal broth) and all evaluated genes act differently with response to iron reduction in media. The highest relative fold change of fece gene was observed at early stage of treatment indicating that PMTB2.1 may utilize its periplasmic protein at early stage to acquire iron. Furthermore, down-regulation expression of fece with the elevated expression of other genes at later time points suggests that PMTB2.1 control their iron requirements in response to iron availability by down-regulating the expression of iron proteins. Moreover, significantly high relative fold change (p ≤ 0.05) of fbpb gene is probably associated with the ability of P. multocida to directly use host iron complex such as hem, hemoglobin. In addition, the significant increase (p ≤ 0.05) in fbpb and yfea expressions also reflects the utilization of multiple iron systems in P. multocida strain PMTB2.1. The findings of this study are very much important as relative scarcity of free iron within hosts creates a major barrier to microbial growth inside host and utilization of outer-membrane proteins system in iron acquisition probably occurred at early stage of infection with P. multocida. In conclusion, the presence and utilization of multiple iron system in P. multocida strain PMTB2.1 revealed the importance of iron in the survival of P. multocida.Keywords: iron-related genes, real-time PCR, gene expression profiling, fold changes
Procedia PDF Downloads 46038510 Extending Smart City Infrastructure to Cover Natural Disasters
Authors: Nina Dasari, Satvik Dasari
Abstract:
Smart city solutions are being developed across the globe to transform urban areas. However, the infrastructure enablement for alerting natural disasters such as floods and wildfires is deficient. This paper discusses an innovative device that could be used as part of the smart city initiative to detect and provide alerts in case of floods at road crossings and wildfires. An Internet of Things (IoT) smart city node was designed, tested, and deployed with collaboration from the City of Austin. The end to end solution includes a 3G enabled IoT device, flood and fire sensors, cloud, a mobile app, and IoT analytics. The real-time data was collected and analyzed using IoT analytics to refine the solution for the past year. The results demonstrate that the proposed solution is reliable and provides accurate results. This low-cost solution is viable, and it can replace the current solution which costs tens of thousands of dollars.Keywords: analytics, internet of things, natural disasters, smart city
Procedia PDF Downloads 22438509 Privacy Concerns and Law Enforcement Data Collection to Tackle Domestic and Sexual Violence
Authors: Francesca Radice
Abstract:
Domestic and sexual violence provokes, on average in Australia, one female death per week due to intimate violence behaviours. 83% of couples meet online, and intercepting domestic and sexual violence at this level would be beneficial. It has been observed that violent or coercive behaviour has been apparent from initial conversations on dating apps like Tinder. Child pornography, stalking, and coercive control are some criminal offences from dating apps, including women murdered after finding partners through Tinder. Police databases and predictive policing are novel approaches taken to prevent crime before harm is done. This research will investigate how police databases can be used in a privacy-preserving way to characterise users in terms of their potential for violent crime. Using the COPS database of NSW Police, we will explore how the past criminal record can be interpreted to yield a category of potential danger for each dating app user. It is up to the judgement of each subscriber on what degree of the potential danger they are prepared to enter into. Sentiment analysis is an area where research into natural language processing has made great progress over the last decade. This research will investigate how sentiment analysis can be used to interpret interchanges between dating app users to detect manipulative or coercive sentiments. These can be used to alert law enforcement if continued for a defined number of communications. One of the potential problems of this approach is the potential prejudice a categorisation can cause. Another drawback is the possibility of misinterpreting communications and involving law enforcement without reason. The approach will be thoroughly tested with cross-checks by human readers who verify both the level of danger predicted by the interpretation of the criminal record and the sentiment detected from personal messages. Even if only a few violent crimes can be prevented, the approach will have a tangible value for real people.Keywords: sentiment analysis, data mining, predictive policing, virtual manipulation
Procedia PDF Downloads 7838508 Optical and Double Folding Model Analysis for Alpha Particles Elastically Scattered from 9Be and 11B Nuclei at Different Energies
Authors: Ahmed H. Amer, A. Amar, Sh. Hamada, I. I. Bondouk, F. A. El-Hussiny
Abstract:
Elastic scattering of α-particles from 9Be and 11B nuclei at different alpha energies have been analyzed. Optical model parameters (OMPs) of α-particles elastic scattering by these nuclei at different energies have been obtained. In the present calculations, the real part of the optical potential are derived by folding of nucleon-nucleon (NN) interaction into nuclear matter density distribution of the projectile and target nuclei using computer code FRESCO. A density-dependent version of the M3Y interaction (CDM3Y6), which is based on the G-matrix elements of the Paris NN potential, has been used. Volumetric integrals of the real and imaginary potential depth (JR, JW) have been calculated and found to be energy dependent. Good agreement between the experimental data and the theoretical predictions in the whole angular range. In double folding (DF) calculations, the obtained normalization coefficient Nr is in the range 0.70–1.32.Keywords: elastic scattering, optical model, double folding model, density distribution
Procedia PDF Downloads 29038507 Modeling Food Popularity Dependencies Using Social Media Data
Authors: DEVASHISH KHULBE, MANU PATHAK
Abstract:
The rise in popularity of major social media platforms have enabled people to share photos and textual information about their daily life. One of the popular topics about which information is shared is food. Since a lot of media about food are attributed to particular locations and restaurants, information like spatio-temporal popularity of various cuisines can be analyzed. Tracking the popularity of food types and retail locations across space and time can also be useful for business owners and restaurant investors. In this work, we present an approach using off-the shelf machine learning techniques to identify trends and popularity of cuisine types in an area using geo-tagged data from social media, Google images and Yelp. After adjusting for time, we use the Kernel Density Estimation to get hot spots across the location and model the dependencies among food cuisines popularity using Bayesian Networks. We consider the Manhattan borough of New York City as the location for our analyses but the approach can be used for any area with social media data and information about retail businesses.Keywords: Web Mining, Geographic Information Systems, Business popularity, Spatial Data Analyses
Procedia PDF Downloads 11638506 Assessing Project Performance through Work Sampling and Earned Value Analysis
Authors: Shobha Ramalingam
Abstract:
The majority of the infrastructure projects are affected by time overrun, resulting in project delays and subsequently cost overruns. Time overrun may vary from a few months to as high as five or more years, placing the project viability at risk. One of the probable reasons noted in the literature for this outcome in projects is due to poor productivity. Researchers contend that productivity in construction has only marginally increased over the years. While studies in the literature have extensively focused on time and cost parameters in projects, there are limited studies that integrate time and cost with productivity to assess project performance. To this end, a study was conducted to understand the project delay factors concerning cost, time and productivity. A case-study approach was adopted to collect rich data from a nuclear power plant project site for two months through observation, interviews and document review. The data were analyzed using three different approaches for a comprehensive understanding. Foremost, a root-cause analysis was performed on the data using Ishikawa’s fish-bone diagram technique to identify the various factors impacting the delay concerning time. Based on it, a questionnaire was designed and circulated to concerned executives, including project engineers and contractors to determine the frequency of occurrence of the delay, which was then compiled and presented to the management for a possible solution to mitigate. Second, a productivity analysis was performed on select activities, including rebar bending and concreting through a time-motion study to analyze product performance. Third, data on cost of construction for three years allowed analyzing the cost performance using earned value management technique. All three techniques allowed to systematically and comprehensively identify the key factors that deter project performance and productivity loss in the construction of the nuclear power plant project. The findings showed that improper planning and coordination between multiple trades, concurrent operations, improper workforce and material management, fatigue due to overtime were some of the key factors that led to delays and poor productivity. The findings are expected to act as a stepping stone for further research and have implications for practitioners.Keywords: earned value analysis, time performance, project costs, project delays, construction productivity
Procedia PDF Downloads 9738505 Experimental Investigation and Constitutive Modeling of Volume Strain under Uniaxial Strain Rate Jump Test in HDPE
Authors: Rida B. Arieby, Hameed N. Hameed
Abstract:
In this work, tensile tests on high density polyethylene have been carried out under various constant strain rate and strain rate jump tests. The dependency of the true stress and specially the variation of volume strain have been investigated, the volume strain due to the phenomena of damage was determined in real time during the tests by an optical extensometer called Videotraction. A modified constitutive equations, including strain rate and damage effects, are proposed, such a model is based on a non-equilibrium thermodynamic approach called (DNLR). The ability of the model to predict the complex nonlinear response of this polymer is examined by comparing the model simulation with the available experimental data, which demonstrate that this model can represent the deformation behavior of the polymer reasonably well.Keywords: strain rate jump tests, volume strain, high density polyethylene, large strain, thermodynamics approach
Procedia PDF Downloads 25838504 Traffic Analysis and Prediction Using Closed-Circuit Television Systems
Authors: Aragorn Joaquin Pineda Dela Cruz
Abstract:
Road traffic congestion is continually deteriorating in Hong Kong. The largest contributing factor is the increase in vehicle fleet size, resulting in higher competition over the utilisation of road space. This study proposes a project that can process closed-circuit television images and videos to provide real-time traffic detection and prediction capabilities. Specifically, a deep-learning model involving computer vision techniques for video and image-based vehicle counting, then a separate model to detect and predict traffic congestion levels based on said data. State-of-the-art object detection models such as You Only Look Once and Faster Region-based Convolutional Neural Networks are tested and compared on closed-circuit television data from various major roads in Hong Kong. It is then used for training in long short-term memory networks to be able to predict traffic conditions in the near future, in an effort to provide more precise and quicker overviews of current and future traffic conditions relative to current solutions such as navigation apps.Keywords: intelligent transportation system, vehicle detection, traffic analysis, deep learning, machine learning, computer vision, traffic prediction
Procedia PDF Downloads 102