Search results for: parallel algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3079

Search results for: parallel algorithms

1369 Sensitivity Analysis Optimization of a Horizontal Axis Wind Turbine from Its Aerodynamic Profiles

Authors: Kevin Molina, Daniel Ortega, Manuel Martinez, Andres Gonzalez-Estrada, William Pinto

Abstract:

Due to the increasing environmental impact, the wind energy is getting strong. This research studied the relationship between the power produced by a horizontal axis wind turbine (HAWT) and the aerodynamic profiles used for its construction. The analysis is studied using the Computational Fluid Dynamic (CFD), presenting the parallel between the energy generated by a turbine designed with selected profiles and another one optimized. For the study, a selection process was carried out from profile NACA 6 digits recommended by the National Renewable Energy Laboratory (NREL) for the construction of this type of turbines. The selection was taken into account different characteristics of the wind (speed and density) and the profiles (aerodynamic coefficients Cl and Cd to different Reynolds and incidence angles). From the selected profiles, was carried out a sensitivity analysis optimization process between its geometry and the aerodynamic forces that are induced on it. The 3D model of the turbines was realized using the Blade Element Momentum method (BEM) and both profiles. The flow fields on the turbines were simulated, obtaining the forces induced on the blade, the torques produced and an increase of 3% in power due to the optimized profiles. Therefore, the results show that the sensitivity analysis optimization process can assist to increment the wind turbine power.

Keywords: blade element momentum, blade, fluid structure interaction, horizontal axis wind turbine, profile design

Procedia PDF Downloads 253
1368 Detection of Cardiac Arrhythmia Using Principal Component Analysis and Xgboost Model

Authors: Sujay Kotwale, Ramasubba Reddy M.

Abstract:

Electrocardiogram (ECG) is a non-invasive technique used to study and analyze various heart diseases. Cardiac arrhythmia is a serious heart disease which leads to death of the patients, when left untreated. An early-time detection of cardiac arrhythmia would help the doctors to do proper treatment of the heart. In the past, various algorithms and machine learning (ML) models were used to early-time detection of cardiac arrhythmia, but few of them have achieved better results. In order to improve the performance, this paper implements principal component analysis (PCA) along with XGBoost model. The PCA was implemented to the raw ECG signals which suppress redundancy information and extracted significant features. The obtained significant ECG features were fed into XGBoost model and the performance of the model was evaluated. In order to valid the proposed technique, raw ECG signals obtained from standard MIT-BIH database were employed for the analysis. The result shows that the performance of proposed method is superior to the several state-of-the-arts techniques.

Keywords: cardiac arrhythmia, electrocardiogram, principal component analysis, XGBoost

Procedia PDF Downloads 111
1367 Online Learning Versus Face to Face Learning: A Sentiment Analysis on General Education Mathematics in the Modern World of University of San Carlos School of Arts and Sciences Students Using Natural Language Processing

Authors: Derek Brandon G. Yu, Clyde Vincent O. Pilapil, Christine F. Peña

Abstract:

College students of Cebu province have been indoors since March 2020, and a challenge encountered is the sudden shift from face to face to online learning and with the lack of empirical data on online learning on Higher Education Institutions (HEIs) in the Philippines. Sentiments on face to face and online learning will be collected from University of San Carlos (USC), School of Arts and Sciences (SAS) students regarding Mathematics in the Modern World (MMW), a General Education (GE) course. Natural Language Processing with machine learning algorithms will be used to classify the sentiments of the students. Results of the research study are the themes identified through topic modelling and the overall sentiments of the students in USC SAS

Keywords: natural language processing, online learning, sentiment analysis, topic modelling

Procedia PDF Downloads 234
1366 The Intersection/Union Region Computation for Drosophila Brain Images Using Encoding Schemes Based on Multi-Core CPUs

Authors: Ming-Yang Guo, Cheng-Xian Wu, Wei-Xiang Chen, Chun-Yuan Lin, Yen-Jen Lin, Ann-Shyn Chiang

Abstract:

With more and more Drosophila Driver and Neuron images, it is an important work to find the similarity relationships among them as the functional inference. There is a general problem that how to find a Drosophila Driver image, which can cover a set of Drosophila Driver/Neuron images. In order to solve this problem, the intersection/union region for a set of images should be computed at first, then a comparison work is used to calculate the similarities between the region and other images. In this paper, three encoding schemes, namely Integer, Boolean, Decimal, are proposed to encode each image as a one-dimensional structure. Then, the intersection/union region from these images can be computed by using the compare operations, Boolean operators and lookup table method. Finally, the comparison work is done as the union region computation, and the similarity score can be calculated by the definition of Tanimoto coefficient. The above methods for the region computation are also implemented in the multi-core CPUs environment with the OpenMP. From the experimental results, in the encoding phase, the performance by the Boolean scheme is the best than that by others; in the region computation phase, the performance by Decimal is the best when the number of images is large. The speedup ratio can achieve 12 based on 16 CPUs. This work was supported by the Ministry of Science and Technology under the grant MOST 106-2221-E-182-070.

Keywords: Drosophila driver image, Drosophila neuron images, intersection/union computation, parallel processing, OpenMP

Procedia PDF Downloads 232
1365 Proteomic Analysis of 2,4-Epibrassinolide Alleviating Low Temperature Stress in Rice Seedling Leaves

Authors: Jiang Xu, Daoping Wang, Qun Li, Yinghong Pan

Abstract:

2,4-Epibrassinolide (EBR), which is a kind of plant hormone Brassinosteroids (BRs), is widely studied and applied in the global scale but the proteomic characteristics of EBR alleviating low temperature stress in rice seedling leaves are still not clear. In this study, seeding rice of Nipponbare were treated with EBR and distilled water, then stressed at 4℃ or 26 ℃, and analyzed by mass spectrometry analysis, verified by parallel reaction monitoring technique (PRM). The results showed that 5778 proteins were identified in total and 4834 proteins were identified with quantitative information. Among them, 401 up-regulated and 220 down-regulated proteins may be related to EBR alleviating low temperature stress in rice seedling leaves. The molecular functions of most of up-regulated proteins are RNA binding and hydrolase activity and are mainly enriched in the pathways of carbon metabolism, folic acid synthesis, and amino acid biosynthesis. The down-regulated proteins are mainly related to catalytic activity and oxidoreductase activity and are mainly enriched in the pathways of limonene and pinene degradation, riboflavin metabolism, porphyrin and chlorophyll metabolism, and other metabolic pathways. PRM validation and literature analysis showed that NADP-malic acidase, peroxidase, 3-phosphoglycerate dehydrogenase, enolase, glyceraldehyde-3- phosphate dehydrogenase and pyruvate kinase are closely related to the effect of EBR on low temperature stress. These results also suggested that BRs could relieve the effect of low temperature stress on rice seed germination in many ways.

Keywords: 2, 4-Epibrassinolid, low temperature stress, proteomic analysis, rice

Procedia PDF Downloads 154
1364 An Activity Based Trajectory Search Approach

Authors: Mohamed Mahmoud Hasan, Hoda M. O. Mokhtar

Abstract:

With the gigantic increment in portable applications use and the spread of positioning and location-aware technologies that we are seeing today, new procedures and methodologies for location-based strategies are required. Location recommendation is one of the highly demanded location-aware applications uniquely with the wide accessibility of social network applications that are location-aware including Facebook check-ins, Foursquare, and others. In this paper, we aim to present a new methodology for location recommendation. The proposed approach coordinates customary spatial traits alongside other essential components including shortest distance, and user interests. We also present another idea namely, "activity trajectory" that represents trajectory that fulfills the set of activities that the user is intrigued to do. The approach dispatched acquaints the related distance value to select trajectory(ies) with minimum cost value (distance) and spatial-area to prune unneeded directions. The proposed calculation utilizes the idea of movement direction to prescribe most comparable N-trajectory(ies) that matches the client's required action design with least voyaging separation. To upgrade the execution of the proposed approach, parallel handling is applied through the employment of a MapReduce based approach. Experiments taking into account genuine information sets were built up and tested for assessing the proposed approach. The exhibited tests indicate how the proposed approach beets different strategies giving better precision and run time.

Keywords: location based recommendation, map-reduce, recommendation system, trajectory search

Procedia PDF Downloads 218
1363 An Equivalent Circuit Model Approach for Battery Pack Simulation in a Hybrid Electric Vehicle System Powertrain

Authors: Suchitra Sivakumar, Hajime Shingyouchi, Toshinori Okajima, Kyohei Yamaguchi, Jin Kusaka

Abstract:

The progressing need for powertrain electrification calls for more accurate and reliable simulation models. A battery pack serves as the most vital component for energy storage in an electrified powertrain. Hybrid electric vehicles (HEV) do not behave the same way as they age, and there are several environmental factors that account for the degradation of the battery on a system level. Therefore, in this work, a battery model was proposed to study the state of charge (SOC) variation and the internal dynamic changes that contribute to aging and performance degradation in HEV batteries. An equivalent circuit battery model (ECM) is built using MATLAB Simulink to investigate the output characteristics of the lithium-ion battery. The ECM comprises of circuit elements like a voltage source, a series resistor and a parallel RC network connected in series. A parameter estimation study is conducted on the ECM to study the dependencies of the circuit elements with the state of charge (SOC) and the terminal voltage of the battery. The battery model is extended to simulate the temperature dependence of the individual battery cell and the battery pack with the environment. The temperature dependence model accounts for the heat loss due to internal resistance build up in the battery pack during charging, discharging, and due to atmospheric temperature. The model was validated for a lithium-ion battery pack with an independent drive cycle showing a voltage accuracy of 4% and SOC accuracy of about 2%.

Keywords: battery model, hybrid electric vehicle, lithium-ion battery, thermal model

Procedia PDF Downloads 285
1362 Enhancing Code Security with AI-Powered Vulnerability Detection

Authors: Zzibu Mark Brian

Abstract:

As software systems become increasingly complex, ensuring code security is a growing concern. Traditional vulnerability detection methods often rely on manual code reviews or static analysis tools, which can be time-consuming and prone to errors. This paper presents a distinct approach to enhancing code security by leveraging artificial intelligence (AI) and machine learning (ML) techniques. Our proposed system utilizes a combination of natural language processing (NLP) and deep learning algorithms to identify and classify vulnerabilities in real-world codebases. By analyzing vast amounts of open-source code data, our AI-powered tool learns to recognize patterns and anomalies indicative of security weaknesses. We evaluated our system on a dataset of over 10,000 open-source projects, achieving an accuracy rate of 92% in detecting known vulnerabilities. Furthermore, our tool identified previously unknown vulnerabilities in popular libraries and frameworks, demonstrating its potential for improving software security.

Keywords: AI, machine language, cord security, machine leaning

Procedia PDF Downloads 18
1361 Development of a Computer Vision System for the Blind and Visually Impaired Person

Authors: Rodrigo C. Belleza, Jr., Roselyn A. Maaño, Karl Patrick E. Camota, Darwin Kim Q. Bulawan

Abstract:

Eyes are an essential and conspicuous organ of the human body. Human eyes are outward and inward portals of the body that allows to see the outside world and provides glimpses into ones inner thoughts and feelings. Inevitable blindness and visual impairments may result from eye-related disease, trauma, or congenital or degenerative conditions that cannot be corrected by conventional means. The study emphasizes innovative tools that will serve as an aid to the blind and visually impaired (VI) individuals. The researchers fabricated a prototype that utilizes the Microsoft Kinect for Windows and Arduino microcontroller board. The prototype facilitates advanced gesture recognition, voice recognition, obstacle detection and indoor environment navigation. Open Computer Vision (OpenCV) performs image analysis, and gesture tracking to transform Kinect data to the desired output. A computer vision technology device provides greater accessibility for those with vision impairments.

Keywords: algorithms, blind, computer vision, embedded systems, image analysis

Procedia PDF Downloads 312
1360 Developing New Algorithm and Its Application on Optimal Control of Pumps in Water Distribution Network

Authors: R. Rajabpour, N. Talebbeydokhti, M. H. Ahmadi

Abstract:

In recent years, new techniques for solving complex problems in engineering are proposed. One of these techniques is JPSO algorithm. With innovative changes in the nature of the jump algorithm JPSO, it is possible to construct a graph-based solution with a new algorithm called G-JPSO. In this paper, a new algorithm to solve the optimal control problem Fletcher-Powell and optimal control of pumps in water distribution network was evaluated. Optimal control of pumps comprise of optimum timetable operation (status on and off) for each of the pumps at the desired time interval. Maximum number of status on and off for each pumps imposed to the objective function as another constraint. To determine the optimal operation of pumps, a model-based optimization-simulation algorithm was developed based on G-JPSO and JPSO algorithms. The proposed algorithm results were compared well with the ant colony algorithm, genetic and JPSO results. This shows the robustness of proposed algorithm in finding near optimum solutions with reasonable computational cost.

Keywords: G-JPSO, operation, optimization, pumping station, water distribution networks

Procedia PDF Downloads 396
1359 Simple Multipath Compensation for Frequency Modulated Signals: A Case of Radio Frequency vs. Quadrature Baseband

Authors: Lusungu Ndovi

Abstract:

Radio propagation from point-to-point is affected by the physical channel in many ways. A signal arriving at a destination travels through a number of different paths which are referred to as multi-paths. Research in this area of wireless communications has progressed well over the years with the research taking different angles of focus. By this is meant that some researchers focus on ways of reducing or eluding Multipath effects whilst others focus on ways of mitigating the effects of Multipath through compensation schemes. Baseband processing is seen as one field of signal processing that is cardinal to the advancement of software-defined radio technology. This has led to wide research into the carrying out certain algorithms at baseband. This paper considers compensating for Multipath for Frequency Modulated signals. The compensation process is carried out at Radio frequency (RF) and at Quadrature baseband (QBB) and the results are compared. Simulations are carried out using MatLab so as to show the benefits of working at lower QBB frequencies than at RF.

Keywords: quadrature baseband, qadio frequency, qultipath compensation, frequency qodulation, signal processing

Procedia PDF Downloads 476
1358 CPU Architecture Based on Static Hardware Scheduler Engine and Multiple Pipeline Registers

Authors: Ionel Zagan, Vasile Gheorghita Gaitan

Abstract:

The development of CPUs and of real-time systems based on them made it possible to use time at increasingly low resolutions. Together with the scheduling methods and algorithms, time organizing has been improved so as to respond positively to the need for optimization and to the way in which the CPU is used. This presentation contains both a detailed theoretical description and the results obtained from research on improving the performances of the nMPRA (Multi Pipeline Register Architecture) processor by implementing specific functions in hardware. The proposed CPU architecture has been developed, simulated and validated by using the FPGA Virtex-7 circuit, via a SoC project. Although the nMPRA processor hardware structure with five pipeline stages is very complex, the present paper presents and analyzes the tests dedicated to the implementation of the CPU and of the memory on-chip for instructions and data. In order to practically implement and test the entire SoC project, various tests have been performed. These tests have been performed in order to verify the drivers for peripherals and the boot module named Bootloader.

Keywords: hardware scheduler, nMPRA processor, real-time systems, scheduling methods

Procedia PDF Downloads 264
1357 Impact of Climate Change on Energy Consumption of the Residential Building Stock in Turkey

Authors: Sadik Yigit

Abstract:

The energy consumed in the buildings constitutes a large portion of the total energy consumption in the world. In this study, it was aimed to measure the impact of climate change on the energy consumption of residential building stock by analyzing a typical mid-rise residential building in four different climate regions of Turkey. An integrated system was developed using the "Distribution Evolutionary Algorithms in Python" tool and Energy Plus. By using the developed integrated system, the energy performance of the typical residential building was analyzed under the effect of different climate change scenarios. The results indicated that predicted overheating will be experienced in the future, which will significantly increase the cooling energy loads of the buildings. In addition, design solutions to improve the future energy performance of the buildings were proposed, considering budget constraints. The results of the study will guide researchers studying in this area of research and designers in the sector in finding climate change resilient design solutions.

Keywords: energy_efficient, residential buildings, climate change, energyplus

Procedia PDF Downloads 96
1356 Analysis of Fault Tolerance on Grid Computing in Real Time Approach

Authors: Parampal Kaur, Deepak Aggarwal

Abstract:

In the computational Grid, fault tolerance is an imperative issue to be considered during job scheduling. Due to the widespread use of resources, systems are highly prone to errors and failures. Hence, fault tolerance plays a key role in the grid to avoid the problem of unreliability. Scheduling the task to the appropriate resource is a vital requirement in computational Grid. The fittest resource scheduling algorithm searches for the appropriate resource based on the job requirements, in contrary to the general scheduling algorithms where jobs are scheduled to the resources with best performance factor. The proposed method is to improve the fault tolerance of the fittest resource scheduling algorithm by scheduling the job in coordination with job replication when the resource has low reliability. Based on the reliability index of the resource, the resource is identified as critical. The tasks are scheduled based on the criticality of the resources. Results show that the execution time of the tasks is comparatively reduced with the proposed algorithm using real-time approach rather than a simulator.

Keywords: computational grid, fault tolerance, task replication, job scheduling

Procedia PDF Downloads 430
1355 Application of Harris Hawks Optimization Metaheuristic Algorithm and Random Forest Machine Learning Method for Long-Term Production Scheduling Problem under Uncertainty in Open-Pit Mines

Authors: Kamyar Tolouei, Ehsan Moosavi

Abstract:

In open-pit mines, the long-term production scheduling optimization problem (LTPSOP) is a complicated problem that contains constraints, large datasets, and uncertainties. Uncertainty in the output is caused by several geological, economic, or technical factors. Due to its dimensions and NP-hard nature, it is usually difficult to find an ideal solution to the LTPSOP. The optimal schedule generally restricts the ore, metal, and waste tonnages, average grades, and cash flows of each period. Past decades have witnessed important measurements of long-term production scheduling and optimal algorithms since researchers have become highly cognizant of the issue. In fact, it is not possible to consider LTPSOP as a well-solved problem. Traditional production scheduling methods in open-pit mines apply an estimated orebody model to produce optimal schedules. The smoothing result of some geostatistical estimation procedures causes most of the mine schedules and production predictions to be unrealistic and imperfect. With the expansion of simulation procedures, the risks from grade uncertainty in ore reserves can be evaluated and organized through a set of equally probable orebody realizations. In this paper, to synthesize grade uncertainty into the strategic mine schedule, a stochastic integer programming framework is presented to LTPSOP. The objective function of the model is to maximize the net present value and minimize the risk of deviation from the production targets considering grade uncertainty simultaneously while satisfying all technical constraints and operational requirements. Instead of applying one estimated orebody model as input to optimize the production schedule, a set of equally probable orebody realizations are applied to synthesize grade uncertainty in the strategic mine schedule and to produce a more profitable and risk-based production schedule. A mixture of metaheuristic procedures and mathematical methods paves the way to achieve an appropriate solution. This paper introduced a hybrid model between the augmented Lagrangian relaxation (ALR) method and the metaheuristic algorithm, the Harris Hawks optimization (HHO), to solve the LTPSOP under grade uncertainty conditions. In this study, the HHO is experienced to update Lagrange coefficients. Besides, a machine learning method called Random Forest is applied to estimate gold grade in a mineral deposit. The Monte Carlo method is used as the simulation method with 20 realizations. The results specify that the progressive versions have been considerably developed in comparison with the traditional methods. The outcomes were also compared with the ALR-genetic algorithm and ALR-sub-gradient. To indicate the applicability of the model, a case study on an open-pit gold mining operation is implemented. The framework displays the capability to minimize risk and improvement in the expected net present value and financial profitability for LTPSOP. The framework could control geological risk more effectively than the traditional procedure considering grade uncertainty in the hybrid model framework.

Keywords: grade uncertainty, metaheuristic algorithms, open-pit mine, production scheduling optimization

Procedia PDF Downloads 97
1354 Dissecting Big Trajectory Data to Analyse Road Network Travel Efficiency

Authors: Rania Alshikhe, Vinita Jindal

Abstract:

Digital innovation has played a crucial role in managing smart transportation. For this, big trajectory data collected from traveling vehicles, such as taxis through installed global positioning system (GPS)-enabled devices can be utilized. It offers an unprecedented opportunity to trace the movements of vehicles in fine spatiotemporal granularity. This paper aims to explore big trajectory data to measure the travel efficiency of road networks using the proposed statistical travel efficiency measure (STEM) across an entire city. Further, it identifies the cause of low travel efficiency by proposed least square approximation network-based causality exploration (LANCE). Finally, the resulting data analysis reveals the causes of low travel efficiency, along with the road segments that need to be optimized to improve the traffic conditions and thus minimize the average travel time from given point A to point B in the road network. Obtained results show that our proposed approach outperforms the baseline algorithms for measuring the travel efficiency of the road network.

Keywords: GPS trajectory, road network, taxi trips, digital map, big data, STEM, LANCE

Procedia PDF Downloads 153
1353 Hierarchical Tree Long Short-Term Memory for Sentence Representations

Authors: Xiuying Wang, Changliang Li, Bo Xu

Abstract:

A fixed-length feature vector is required for many machine learning algorithms in NLP field. Word embeddings have been very successful at learning lexical information. However, they cannot capture the compositional meaning of sentences, which prevents them from a deeper understanding of language. In this paper, we introduce a novel hierarchical tree long short-term memory (HTLSTM) model that learns vector representations for sentences of arbitrary syntactic type and length. We propose to split one sentence into three hierarchies: short phrase, long phrase and full sentence level. The HTLSTM model gives our algorithm the potential to fully consider the hierarchical information and long-term dependencies of language. We design the experiments on both English and Chinese corpus to evaluate our model on sentiment analysis task. And the results show that our model outperforms several existing state of the art approaches significantly.

Keywords: deep learning, hierarchical tree long short-term memory, sentence representation, sentiment analysis

Procedia PDF Downloads 347
1352 Spatial Data Mining by Decision Trees

Authors: Sihem Oujdi, Hafida Belbachir

Abstract:

Existing methods of data mining cannot be applied on spatial data because they require spatial specificity consideration, as spatial relationships. This paper focuses on the classification with decision trees, which are one of the data mining techniques. We propose an extension of the C4.5 algorithm for spatial data, based on two different approaches Join materialization and Querying on the fly the different tables. Similar works have been done on these two main approaches, the first - Join materialization - favors the processing time in spite of memory space, whereas the second - Querying on the fly different tables- promotes memory space despite of the processing time. The modified C4.5 algorithm requires three entries tables: a target table, a neighbor table, and a spatial index join that contains the possible spatial relationship among the objects in the target table and those in the neighbor table. Thus, the proposed algorithms are applied to a spatial data pattern in the accidentology domain. A comparative study of our approach with other works of classification by spatial decision trees will be detailed.

Keywords: C4.5 algorithm, decision trees, S-CART, spatial data mining

Procedia PDF Downloads 607
1351 Evaluation and New Modeling Improvement of Water Quality

Authors: Sebahat Seker

Abstract:

Since there is a parallel connection between drinking water quality and public health, studies on drinking and domestic water are of vital importance. Ardahan Province is one of the provinces located in the Northeast Anatolian Region, where animal husbandry and agriculture are carried out economically. City mains water uses underground spring water as a source and is chlorinated and given to the city center by gravity. However, mains water cannot be used outside the central district of the city, and the majority of the people meet their drinking and utility water needs from the wells they have opened individually. The water element, which is vital for all living things, is the most important substance that sustains life for humans. Under normal conditions, a healthy person consumes approximately 1.8-2 liters of water. The quality and use of potable water is one of the most important issues in terms of health. The quality parameters of drinking and utility water have been revealed by the scientific world. Scientific studies on drinking water quality in the world and its impact on public health are among the most popular topics. Although our country is surrounded by water on three sides, potable water resources are very few. In the Eastern Anatolia Region, it is difficult for the public to access drinking and utility water due to the difficult conditions both climatically and geographically. In this study, samples taken from drinking and utility water at certain intervals from the stations determined, and water quality parameters will be determined. The fact that such a study has not been carried out in the region before and the knowledge of the local people about water quality is very important in terms of its original and widespread effect.

Keywords: water quality, modelling, evaluation, northeastern anatolia

Procedia PDF Downloads 196
1350 An Optimization Model for Waste Management in Demolition Works

Authors: Eva Queheille, Franck Taillandier, Nadia Saiyouri

Abstract:

Waste management has become a major issue in demolition works, because of its environmental impact (energy consumption, resource consumption, pollution…). However, improving waste management requires to take also into account the overall demolition process and to consider demolition main objectives (e.g. cost, delay). Establishing a strategy with these conflicting objectives (economic and environment) remains complex. In order to provide a decision-support for demolition companies, a multi-objective optimization model was developed. In this model, a demolition strategy is computed from a set of 80 decision variables (worker team composition, machines, treatment for each type of waste, choice of treatment platform…), which impacts the demolition objectives. The model has experimented on a real-case study (demolition of several buildings in France). To process the optimization, different optimization algorithms (NSGA2, MOPSO, DBEA…) were tested. Results allow the engineer in charge of this case, to build a sustainable demolition strategy without affecting cost or delay.

Keywords: deconstruction, life cycle assessment, multi-objective optimization, waste management

Procedia PDF Downloads 146
1349 Benchmarking of Pentesting Tools

Authors: Esteban Alejandro Armas Vega, Ana Lucila Sandoval Orozco, Luis Javier García Villalba

Abstract:

The benchmarking of tools for dynamic analysis of vulnerabilities in web applications is something that is done periodically, because these tools from time to time update their knowledge base and search algorithms, in order to improve their accuracy. Unfortunately, the vast majority of these evaluations are made by software enthusiasts who publish their results on blogs or on non-academic websites and always with the same evaluation methodology. Similarly, academics who have carried out this type of analysis from a scientific approach, the majority, make their analysis within the same methodology as well the empirical authors. This paper is based on the interest of finding answers to questions that many users of this type of tools have been asking over the years, such as, to know if the tool truly test and evaluate every vulnerability that it ensures do, or if the tool, really, deliver a real report of all the vulnerabilities tested and exploited. This kind of questions have also motivated previous work but without real answers. The aim of this paper is to show results that truly answer, at least on the tested tools, all those unanswered questions. All the results have been obtained by changing the common model of benchmarking used for all those previous works.

Keywords: cybersecurity, IDS, security, web scanners, web vulnerabilities

Procedia PDF Downloads 314
1348 Predicting Machine-Down of Woodworking Industrial Machines

Authors: Matteo Calabrese, Martin Cimmino, Dimos Kapetis, Martina Manfrin, Donato Concilio, Giuseppe Toscano, Giovanni Ciandrini, Giancarlo Paccapeli, Gianluca Giarratana, Marco Siciliano, Andrea Forlani, Alberto Carrotta

Abstract:

In this paper we describe a machine learning methodology for Predictive Maintenance (PdM) applied on woodworking industrial machines. PdM is a prominent strategy consisting of all the operational techniques and actions required to ensure machine availability and to prevent a machine-down failure. One of the challenges with PdM approach is to design and develop of an embedded smart system to enable the health status of the machine. The proposed approach allows screening simultaneously multiple connected machines, thus providing real-time monitoring that can be adopted with maintenance management. This is achieved by applying temporal feature engineering techniques and training an ensemble of classification algorithms to predict Remaining Useful Lifetime of woodworking machines. The effectiveness of the methodology is demonstrated by testing an independent sample of additional woodworking machines without presenting machine down event.

Keywords: predictive maintenance, machine learning, connected machines, artificial intelligence

Procedia PDF Downloads 220
1347 Development of a Psychometric Testing Instrument Using Algorithms and Combinatorics to Yield Coupled Parameters and Multiple Geometric Arrays in Large Information Grids

Authors: Laith F. Gulli, Nicole M. Mallory

Abstract:

The undertaking to develop a psychometric instrument is monumental. Understanding the relationship between variables and events is important in structural and exploratory design of psychometric instruments. Considering this, we describe a method used to group, pair and combine multiple Philosophical Assumption statements that assisted in development of a 13 item psychometric screening instrument. We abbreviated our Philosophical Assumptions (PA)s and added parameters, which were then condensed and mathematically modeled in a specific process. This model produced clusters of combinatorics which was utilized in design and development for 1) information retrieval and categorization 2) item development and 3) estimation of interactions among variables and likelihood of events. The psychometric screening instrument measured Knowledge, Assessment (education) and Beliefs (KAB) of New Addictions Research (NAR), which we called KABNAR. We obtained an overall internal consistency for the seven Likert belief items as measured by Cronbach’s α of .81 in the final study of 40 Clinicians, calculated by SPSS 14.0.1 for Windows. We constructed the instrument to begin with demographic items (degree/addictions certifications) for identification of target populations that practiced within Outpatient Substance Abuse Counseling (OSAC) settings. We then devised education items, beliefs items (seven items) and a modifiable “barrier from learning” item that consisted of six “choose any” choices. We also conceptualized a close relationship between identifying various degrees and certifications held by Outpatient Substance Abuse Therapists (OSAT) (the demographics domain) and all aspects of their education related to EB-NAR (past and present education and desired future training). We placed a descriptive (PA)1tx in both demographic and education domains to trace relationships of therapist education within these two domains. The two perceptions domains B1/b1 and B2/b2 represented different but interrelated perceptions from the therapist perspective. The belief items measured therapist perceptions concerning EB-NAR and therapist perceptions using EB-NAR during the beginning of outpatient addictions counseling. The (PA)s were written in simple words and descriptively accurate and concise. We then devised a list of parameters and appropriately matched them to each PA and devised descriptive parametric (PA)s in a domain categorized information grid. Descriptive parametric (PA)s were reduced to simple mathematical symbols. This made it easy to utilize parametric (PA)s into algorithms, combinatorics and clusters to develop larger information grids. By using matching combinatorics we took paired demographic and education domains with a subscript of 1 and matched them to the column with each B domain with subscript 1. Our algorithmic matching formed larger information grids with organized clusters in columns and rows. We repeated the process using different demographic, education and belief domains and devised multiple information grids with different parametric clusters and geometric arrays. We found benefit combining clusters by different geometric arrays, which enabled us to trace parametric variables and concepts. We were able to understand potential differences between dependent and independent variables and trace relationships of maximum likelihoods.

Keywords: psychometric, parametric, domains, grids, therapists

Procedia PDF Downloads 273
1346 DNA Based Identification of Insect Vectors for Zoonotic Diseases From District Faisalabad, Pakistan

Authors: Zain Ul Abdin, Mirza Aizaz Asim, Rao Sohail Ahmad Khan, Luqman Amrao, Fiaz Hussain, Hasooba Hira, Saqi Kosar Abbas

Abstract:

The success of Integrated vector management programmes mainly depends on the correct identification of insect vector species involved in vector borne diseases. Based on molecular data the most important insect species involved as vectors for Zoonotic diseases in Pakistan were identified. The precise and accurate identification of such type of organism is only possible through molecular based techniques like “DNA barcoding”. Morphological species identification in insects at any life stage, is very challenging, therefore, DNA barcoding was used as a tool for rapid and accurate species identification in a wide variety of taxa across the globe and parallel studies revealed that DNA barcoding data can be effectively used in resolving taxonomic ambiguities, detection of cryptic diversity, invasion biology, description of new species etc. A comprehensive survey was carried out for the collection of insects (both adult and immature stages) in district Faisalabad, Pakistan and their DNA was extracted and mitochondrial cytochrome oxidase subunit I (COI-59) barcode sequences was used for molecular identification of immature and adult life stage.This preliminary research work opens new frontiers for developing sustainable insect vectors management programmes for saving lives of mankind from fatal diseases.

Keywords: zoonotic diseases, cytochrome oxidase, and insect vectors, CO1

Procedia PDF Downloads 161
1345 Method for Evaluating the Monetary Value of a Customized Version of the Digital Twin for the Additive Manufacturing

Authors: Fabio Oettl, Sebastian Hoerbrand, Tobias Wittmeir, Johannes Schilp

Abstract:

By combining the additive manufacturing (AM)- process with digital concepts, like the digital twin (DT) or the downsized and basing concept of the digital part file (DPF), the competitiveness of additive manufacturing is enhanced and new use cases like decentral production are enabled. But in literature, one can´t find any quantitative approach for valuing the usage of a DT or DPF in AM. Out of this fact, such an approach will be developed within this paper in order to further promote or dissuade the usage of these concepts. The focus is set on the production as an early lifecycle phase, which means that the AM-production process gets analyzed regarding the potential advantages of using DPF in AM. These advantages are transferred to a monetary value with this approach. By calculating the costs of the DPF, an overall monetary value is a result. Thereon a tool, based on a simulation environment is constructed, where the algorithms are transformed into a program. The results of applying this tool show that an overall value of 20,81 € for the DPF can be realized for one special use case. For the future application of the DPF there is the recommendation to integrate especially sustainable information because out of this, a higher value of the DPF can be expected.

Keywords: additive manufacturing, digital concept costs, digital part file, digital twin, monetary value estimation

Procedia PDF Downloads 195
1344 WebAppShield: An Approach Exploiting Machine Learning to Detect SQLi Attacks in an Application Layer in Run-time

Authors: Ahmed Abdulla Ashlam, Atta Badii, Frederic Stahl

Abstract:

In recent years, SQL injection attacks have been identified as being prevalent against web applications. They affect network security and user data, which leads to a considerable loss of money and data every year. This paper presents the use of classification algorithms in machine learning using a method to classify the login data filtering inputs into "SQLi" or "Non-SQLi,” thus increasing the reliability and accuracy of results in terms of deciding whether an operation is an attack or a valid operation. A method Web-App auto-generated twin data structure replication. Shielding against SQLi attacks (WebAppShield) that verifies all users and prevents attackers (SQLi attacks) from entering and or accessing the database, which the machine learning module predicts as "Non-SQLi" has been developed. A special login form has been developed with a special instance of data validation; this verification process secures the web application from its early stages. The system has been tested and validated, up to 99% of SQLi attacks have been prevented.

Keywords: SQL injection, attacks, web application, accuracy, database

Procedia PDF Downloads 144
1343 Modified Naive Bayes-Based Prediction Modeling for Crop Yield Prediction

Authors: Kefaya Qaddoum

Abstract:

Most of greenhouse growers desire a determined amount of yields in order to accurately meet market requirements. The purpose of this paper is to model a simple but often satisfactory supervised classification method. The original naive Bayes have a serious weakness, which is producing redundant predictors. In this paper, utilized regularization technique was used to obtain a computationally efficient classifier based on naive Bayes. The suggested construction, utilized L1-penalty, is capable of clearing redundant predictors, where a modification of the LARS algorithm is devised to solve this problem, making this method applicable to a wide range of data. In the experimental section, a study conducted to examine the effect of redundant and irrelevant predictors, and test the method on WSG data set for tomato yields, where there are many more predictors than data, and the urge need to predict weekly yield is the goal of this approach. Finally, the modified approach is compared with several naive Bayes variants and other classification algorithms (SVM and kNN), and is shown to be fairly good.

Keywords: tomato yield prediction, naive Bayes, redundancy, WSG

Procedia PDF Downloads 227
1342 A Fast Silhouette Detection Algorithm for Shadow Volumes in Augmented Reality

Authors: Hoshang Kolivand, Mahyar Kolivand, Mohd Shahrizal Sunar, Mohd Azhar M. Arsad

Abstract:

Real-time shadow generation in virtual environments and Augmented Reality (AR) was always a hot topic in the last three decades. Lots of calculation for shadow generation among AR needs a fast algorithm to overcome this issue and to be capable of implementing in any real-time rendering. In this paper, a silhouette detection algorithm is presented to generate shadows for AR systems. Δ+ algorithm is presented based on extending edges of occluders to recognize which edges are silhouettes in the case of real-time rendering. An accurate comparison between the proposed algorithm and current algorithms in silhouette detection is done to show the reduction calculation by presented algorithm. The algorithm is tested in both virtual environments and AR systems. We think that this algorithm has the potential to be a fundamental algorithm for shadow generation in all complex environments.

Keywords: silhouette detection, shadow volumes, real-time shadows, rendering, augmented reality

Procedia PDF Downloads 434
1341 Estimating Gait Parameter from Digital RGB Camera Using Real Time AlphaPose Learning Architecture

Authors: Murad Almadani, Khalil Abu-Hantash, Xinyu Wang, Herbert Jelinek, Kinda Khalaf

Abstract:

Gait analysis is used by healthcare professionals as a tool to gain a better understanding of the movement impairment and track progress. In most circumstances, monitoring patients in their real-life environments with low-cost equipment such as cameras and wearable sensors is more important. Inertial sensors, on the other hand, cannot provide enough information on angular dynamics. This research offers a method for tracking 2D joint coordinates using cutting-edge vision algorithms and a single RGB camera. We provide an end-to-end comprehensive deep learning pipeline for marker-less gait parameter estimation, which, to our knowledge, has never been done before. To make our pipeline function in real-time for real-world applications, we leverage the AlphaPose human posture prediction model and a deep learning transformer. We tested our approach on the well-known GPJATK dataset, which produces promising results.

Keywords: gait analysis, human pose estimation, deep learning, real time gait estimation, AlphaPose, transformer

Procedia PDF Downloads 112
1340 Health and Safety Risk Assesment with Electromagnetic Field Exposure for Call Center Workers

Authors: Dilsad Akal

Abstract:

Aim: Companies communicate with each other and with their costumers via call centers. Call centers are defined as stressful because of their uncertain working hours, inadequate relief time, performance based system and heavy workload. In literature, this sector is defined as risky as mining sector by means of health and safety. The aim of this research is to enlight the relatively dark area. Subject and Methods: The collection of data for this study completed during April-May 2015 for the two selected call centers in different parts of Turkey. The applied question mostly investigated the health conditions of call center workers. Electromagnetic field measurements were completed at the same time with applying the question poll. The ratio of employee accessibility noted as 73% for the first call center and 87% for the second. Results: The results of electromagnetic field measurements were as between 371 V/m-32 V/m for the first location and between 370 V/m-61 V/m for the second. The general complaints of the employees for both workplaces can be counted as; inadequate relief time, inadequate air conditioning, disturbance, poor thermal conditions, inadequate or extreme lighting. Furthermore, musculoskeletal discomfort, stress, ear and eye discomfort are main health problems of employees. Conclusion: The measured values and the responses to the question poll were found parallel with the other similar research results in literature. At the end of this survey, a risk map of workplace was prepared in terms of safety and health at work in general and some suggestions for resolution were provided.

Keywords: call center, health and safety, electromagnetic field, risk map

Procedia PDF Downloads 177