Search results for: AI algorithm internal audit
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6284

Search results for: AI algorithm internal audit

4694 Training a Neural Network Using Input Dropout with Aggressive Reweighting (IDAR) on Datasets with Many Useless Features

Authors: Stylianos Kampakis

Abstract:

This paper presents a new algorithm for neural networks called “Input Dropout with Aggressive Re-weighting” (IDAR) aimed specifically at datasets with many useless features. IDAR combines two techniques (dropout of input neurons and aggressive re weighting) in order to eliminate the influence of noisy features. The technique can be seen as a generalization of dropout. The algorithm is tested on two different benchmark data sets: a noisy version of the iris dataset and the MADELON data set. Its performance is compared against three other popular techniques for dealing with useless features: L2 regularization, LASSO and random forests. The results demonstrate that IDAR can be an effective technique for handling data sets with many useless features.

Keywords: neural networks, feature selection, regularization, aggressive reweighting

Procedia PDF Downloads 455
4693 EEG Signal Processing Methods to Differentiate Mental States

Authors: Sun H. Hwang, Young E. Lee, Yunhan Ga, Gilwon Yoon

Abstract:

EEG is a very complex signal with noises and other bio-potential interferences. EOG is the most distinct interfering signal when EEG signals are measured and analyzed. It is very important how to process raw EEG signals in order to obtain useful information. In this study, the EEG signal processing techniques such as EOG filtering and outlier removal were examined to minimize unwanted EOG signals and other noises. The two different mental states of resting and focusing were examined through EEG analysis. A focused state was induced by letting subjects to watch a red dot on the white screen. EEG data for 32 healthy subjects were measured. EEG data after 60-Hz notch filtering were processed by a commercially available EOG filtering and our presented algorithm based on the removal of outliers. The ratio of beta wave to theta wave was used as a parameter for determining the degree of focusing. The results show that our algorithm was more appropriate than the existing EOG filtering.

Keywords: EEG, focus, mental state, outlier, signal processing

Procedia PDF Downloads 283
4692 The Challenges to Information Communication Technology Integration in Mathematics Teaching and Learning

Authors: George Onomah

Abstract:

Background: The integration of information communication technology (ICT) in Mathematics education faces notable challenges, which this study aimed to dissect and understand. Objectives: The primary goal was to assess the internal and external factors affecting the adoption of ICT by in-service Mathematics teachers. Internal factors examined included teachers' pedagogical beliefs, prior teaching experience, attitudes towards computers, and proficiency with technology. External factors included the availability of technological resources, the level of ICT training received, the sufficiency of allocated time for technology use, and the institutional culture within educational environments. Methods: A descriptive survey design was employed to methodically investigate these factors. Data collection was carried out using a five-point Likert scale questionnaire, administered to a carefully selected sample of 100 in-service Mathematics teachers through a combination of purposive and convenience sampling techniques. Findings: Results from multiple regression analysis revealed a significant underutilization of ICT in Mathematics teaching, highlighting a pronounced deficiency in current classroom practices. Recommendations: The findings suggest an urgent need for educational department heads to implement regular and comprehensive ICT training programs aimed at enhancing teachers' technological capabilities and promoting the integration of ICT in Mathematics teaching methodologies.

Keywords: ICT, Mathematics, integration, barriers

Procedia PDF Downloads 40
4691 Using Genetic Algorithm to Organize Sustainable Urban Landscape in Historical Part of City

Authors: Shahab Mirzaean Mahabadi, Elham Ebrahimi

Abstract:

The urban development process in the historical urban context has predominately witnessed two main approaches: the first is the Preservation and conservation of the urban fabric and its value, and the second approach is urban renewal and redevelopment. The latter is generally supported by political and economic aspirations. These two approaches conflict evidently. The authors go through the history of urban planning in order to review the historical development of the mentioned approaches. In this article, various values which are inherent in the historical fabric of a city are illustrated by emphasizing on cultural identity and activity. In the following, it is tried to find an optimized plan which maximizes economic development and minimizes change in historical-cultural sites simultaneously. In the proposed model, regarding the decision maker’s intention, and the variety of functions, the selected zone is divided into a number of components. For each component, different alternatives can be assigned, namely, renovation, refurbishment, destruction, and change in function. The decision Variable in this model is to choose an alternative for each component. A set of decisions made upon all components results in a plan. A plan developed in this way can be evaluated based on the decision maker’s point of view. That is, interactions between selected alternatives can make a foundation for the assessment of urban context to design a historical-cultural landscape. A genetic algorithm (GA) approach is used to search for optimal future land use within the historical-culture landscape for a sustainable high-growth city.

Keywords: urban sustainability, green city, regeneration, genetic algorithm

Procedia PDF Downloads 69
4690 Resource-Constrained Heterogeneous Workflow Scheduling Algorithms in Heterogeneous Computing Clusters

Authors: Lei Wang, Jiahao Zhou

Abstract:

The development of heterogeneous computing clusters provides a strong computility guarantee for large-scale workflows (e.g., scientific computing, artificial intelligence (AI), etc.). However, the tasks within large-scale workflows have also gradually become heterogeneous due to different demands on computing resources, which leads to the addition of a task resource-restricted constraint to the workflow scheduling problem on heterogeneous computing platforms. In this paper, we propose a heterogeneous constrained minimum makespan scheduling algorithm based on the idea of greedy strategy, which provides an efficient solution to the heterogeneous workflow scheduling problem in a heterogeneous platform. In this paper, we test the effectiveness of our proposed scheduling algorithm by randomly generating heterogeneous workflows with heterogeneous computing platform, and the experiments show that our method improves 15.2% over the state-of-the-art methods.

Keywords: heterogeneous computing, workflow scheduling, constrained resources, minimal makespan

Procedia PDF Downloads 34
4689 Web Page Design Optimisation Based on Segment Analytics

Authors: Varsha V. Rohini, P. R. Shreya, B. Renukadevi

Abstract:

In the web analytics the information delivery and the web usage is optimized and the analysis of data is done. The analytics is the measurement, collection and analysis of webpage data. Page statistics and user metrics are the important factor in most of the web analytics tool. This is the limitation of the existing tools. It does not provide design inputs for the optimization of information. This paper aims at providing an extension for the scope of web analytics to provide analysis and statistics of each segment of a webpage. The number of click count is calculated and the concentration of links in a web page is obtained. Its user metrics are used to help in proper design of the displayed content in a webpage by Vision Based Page Segmentation (VIPS) algorithm. When the algorithm is applied on the web page it divides the entire web page into the visual block tree. The visual block tree generated will further divide the web page into visual blocks or segments which help us to understand the usage of each segment in a page and its content. The dynamic web pages and deep web pages are used to extend the scope of web page segment analytics. Space optimization concept is used with the help of the output obtained from the Vision Based Page Segmentation (VIPS) algorithm. This technique provides us the visibility of the user interaction with the WebPages and helps us to place the important links in the appropriate segments of the webpage and effectively manage space in a page and the concentration of links.

Keywords: analytics, design optimization, visual block trees, vision based technology

Procedia PDF Downloads 266
4688 Algorithms for Run-Time Task Mapping in NoC-Based Heterogeneous MPSoCs

Authors: M. K. Benhaoua, A. K. Singh, A. E. Benyamina, P. Boulet

Abstract:

Mapping parallelized tasks of applications onto these MPSoCs can be done either at design time (static) or at run-time (dynamic). Static mapping strategies find the best placement of tasks at design-time, and hence, these are not suitable for dynamic workload and seem incapable of runtime resource management. The number of tasks or applications executing in MPSoC platform can exceed the available resources, requiring efficient run-time mapping strategies to meet these constraints. This paper describes a new Spiral Dynamic Task Mapping heuristic for mapping applications onto NoC-based Heterogeneous MPSoC. This heuristic is based on packing strategy and routing Algorithm proposed also in this paper. Heuristic try to map the tasks of an application in a clustering region to reduce the communication overhead between the communicating tasks. The heuristic proposed in this paper attempts to map the tasks of an application that are most related to each other in a spiral manner and to find the best possible path load that minimizes the communication overhead. In this context, we have realized a simulation environment for experimental evaluations to map applications with varying number of tasks onto an 8x8 NoC-based Heterogeneous MPSoCs platform, we demonstrate that the new mapping heuristics with the new modified dijkstra routing algorithm proposed are capable of reducing the total execution time and energy consumption of applications when compared to state-of-the-art run-time mapping heuristics reported in the literature.

Keywords: multiprocessor system on chip, MPSoC, network on chip, NoC, heterogeneous architectures, run-time mapping heuristics, routing algorithm

Procedia PDF Downloads 489
4687 Simulation and Analysis of Inverted Pendulum Controllers

Authors: Sheren H. Salah

Abstract:

The inverted pendulum is a highly nonlinear and open-loop unstable system. An inverted pendulum (IP) is a pendulum which has its mass above its pivot point. It is often implemented with the pivot point mounted on a cart that can move horizontally and may be called a cart and pole. The characteristics of the inverted pendulum make identification and control more challenging. This paper presents the simulation study of several control strategies for an inverted pendulum system. The goal is to determine which control strategy delivers better performance with respect to pendulum’s angle. The inverted pendulum represents a challenging control problem, which continually moves toward an uncontrolled state. For controlling the inverted pendulum. The simulation study that sliding mode control (SMC) control produced better response compared to Genetic Algorithm Control (GAs) and proportional-integral-derivative(PID) control.

Keywords: Inverted Pendulum (IP) Proportional-Integral-Derivative (PID), Genetic Algorithm Control (GAs), Sliding Mode Control (SMC)

Procedia PDF Downloads 555
4686 The Effect of Internal Electrical Ion Mobility on Molten Salts through Atomistic Simulations

Authors: Carlos F. Sanz-Navarro, Sonia Fereres

Abstract:

Binary and ternary mixtures of molten salts are excellent thermal energy storage systems and have been widely used in commercial tanks both in nuclear and solar thermal applications. However, the energy density of the commercially used mixtures is still insufficient, and therefore, new systems based on latent heat storage (or phase change materials, PCM) are currently being investigated. In order to shed some light on the macroscopic physical properties of the molten salt phases, knowledge of the microscopic structure and dynamics is required. Several molecular dynamics (MD) simulations have been performed to model the thermal behavior of (Li,K)2CO3 mixtures. Up to this date, this particular molten salt mixture has not been extensively studied but it is of fundamental interest for understanding the behavior of other commercial salts. Molten salt diffusivities, the internal electrical ion mobility, and the physical properties of the solid-liquid phase transition have been calculated and compared to available data from literature. The effect of anion polarization and the application of a strong external electric field have also been investigated. The influence of electrical ion mobility on local composition is explained through the Chemla effect, well known in electrochemistry. These results open a new way to design optimal high temperature energy storage materials.

Keywords: atomistic simulations, thermal storage, latent heat, molten salt, ion mobility

Procedia PDF Downloads 326
4685 Biogeography Based CO2 and Cost Optimization of RC Cantilever Retaining Walls

Authors: Ibrahim Aydogdu, Alper Akin

Abstract:

In this study, the development of minimizing the cost and the CO2 emission of the RC retaining wall design has been performed by Biogeography Based Optimization (BBO) algorithm. This has been achieved by developing computer programs utilizing BBO algorithm which minimize the cost and the CO2 emission of the RC retaining walls. Objective functions of the optimization problem are defined as the minimized cost, the CO2 emission and weighted aggregate of the cost and the CO2 functions of the RC retaining walls. In the formulation of the optimum design problem, the height and thickness of the stem, the length of the toe projection, the thickness of the stem at base level, the length and thickness of the base, the depth and thickness of the key, the distance from the toe to the key, the number and diameter of the reinforcement bars are treated as design variables. In the formulation of the optimization problem, flexural and shear strength constraints and minimum/maximum limitations for the reinforcement bar areas are derived from American Concrete Institute (ACI 318-14) design code. Moreover, the development length conditions for suitable detailing of reinforcement are treated as a constraint. The obtained optimum designs must satisfy the factor of safety for failure modes (overturning, sliding and bearing), strength, serviceability and other required limitations to attain practically acceptable shapes. To demonstrate the efficiency and robustness of the presented BBO algorithm, the optimum design example for retaining walls is presented and the results are compared to the previously obtained results available in the literature.

Keywords: bio geography, meta-heuristic search, optimization, retaining wall

Procedia PDF Downloads 399
4684 Improvement of Ground Truth Data for Eye Location on Infrared Driver Recordings

Authors: Sorin Valcan, Mihail Gaianu

Abstract:

Labeling is a very costly and time consuming process which aims to generate datasets for training neural networks in several functionalities and projects. For driver monitoring system projects, the need for labeled images has a significant impact on the budget and distribution of effort. This paper presents the modifications done to an algorithm used for the generation of ground truth data for 2D eyes location on infrared images with drivers in order to improve the quality of the data and performance of the trained neural networks. The algorithm restrictions become tougher, which makes it more accurate but also less constant. The resulting dataset becomes smaller and shall not be altered by any kind of manual label adjustment before being used in the neural networks training process. These changes resulted in a much better performance of the trained neural networks.

Keywords: labeling automation, infrared camera, driver monitoring, eye detection, convolutional neural networks

Procedia PDF Downloads 117
4683 Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization

Authors: Lana Dalawr Jalal

Abstract:

This paper addresses the problem of offline path planning for Unmanned Aerial Vehicles (UAVs) in complex three-dimensional environment with obstacles, which is modelled by 3D Cartesian grid system. Path planning for UAVs require the computational intelligence methods to move aerial vehicles along the flight path effectively to target while avoiding obstacles. In this paper Modified Particle Swarm Optimization (MPSO) algorithm is applied to generate the optimal collision free 3D flight path for UAV. The simulations results clearly demonstrate effectiveness of the proposed algorithm in guiding UAV to the final destination by providing optimal feasible path quickly and effectively.

Keywords: obstacle avoidance, particle swarm optimization, three-dimensional path planning unmanned aerial vehicles

Procedia PDF Downloads 410
4682 A Second Order Genetic Algorithm for Traveling Salesman Problem

Authors: T. Toathom, M. Munlin, P. Sugunnasil

Abstract:

The traveling salesman problem (TSP) is one of the best-known problems in optimization problem. There are many research regarding the TSP. One of the most usage tool for this problem is the genetic algorithm (GA). The chromosome of the GA for TSP is normally encoded by the order of the visited city. However, the traditional chromosome encoding scheme has some limitations which are twofold: the large solution space and the inability to encapsulate some information. The number of solution for a certain problem is exponentially grow by the number of city. Moreover, the traditional chromosome encoding scheme fails to recognize the misplaced correct relation. It implies that the tradition method focuses only on exact solution. In this work, we relax some of the concept in the GA for TSP which is the exactness of the solution. The proposed work exploits the relation between cities in order to reduce the solution space in the chromosome encoding. In this paper, a second order GA is proposed to solve the TSP. The term second order refers to how the solution is encoded into chromosome. The chromosome is divided into 2 types: the high order chromosome and the low order chromosome. The high order chromosome is the chromosome that focus on the relation between cities such as the city A should be visited before city B. On the other hand, the low order chromosome is a type of chromosome that is derived from a high order chromosome. In other word, low order chromosome is encoded by the traditional chromosome encoding scheme. The genetic operation, mutation and crossover, will be performed on the high order chromosome. Then, the high order chromosome will be mapped to a group of low order chromosomes whose characteristics are satisfied with the high order chromosome. From the mapped set of chromosomes, the champion chromosome will be selected based on the fitness value which will be later used as a representative for the high order chromosome. The experiment is performed on the city data from TSPLIB.

Keywords: genetic algorithm, traveling salesman problem, initial population, chromosomes encoding

Procedia PDF Downloads 270
4681 Characterization and Modelling of Aerosol Droplet in Absorption Columns

Authors: Hammad Majeed, Hanna Knuutila, Magne Hillestad, Hallvard F. Svendsen

Abstract:

Formation of aerosols can cause serious complications in industrial exhaust gas CO2 capture processes. SO3 present in the flue gas can cause aerosol formation in an absorption based capture process. Small mist droplets and fog formed can normally not be removed in conventional demisting equipment because their submicron size allows the particles or droplets to follow the gas flow. As a consequence of this aerosol based emissions in the order of grams per Nm3 have been identified from PCCC plants. In absorption processes aerosols are generated by spontaneous condensation or desublimation processes in supersaturated gas phases. Undesired aerosol development may lead to amine emissions many times larger than what would be encountered in a mist free gas phase in PCCC development. It is thus of crucial importance to understand the formation and build-up of these aerosols in order to mitigate the problem. Rigorous modelling of aerosol dynamics leads to a system of partial differential equations. In order to understand mechanics of a particle entering an absorber an implementation of the model is created in Matlab. The model predicts the droplet size, the droplet internal variable profiles and the mass transfer fluxes as function of position in the absorber. The Matlab model is based on a subclass method of weighted residuals for boundary value problems named, orthogonal collocation method. The model comprises a set of mass transfer equations for transferring components and the essential diffusion reaction equations to describe the droplet internal profiles for all relevant constituents. Also included is heat transfer across the interface and inside the droplet. This paper presents results describing the basic simulation tool for the characterization of aerosols formed in CO2 absorption columns and gives examples as to how various entering droplets grow or shrink through an absorber and how their composition changes with respect to time. Below are given some preliminary simulation results for an aerosol droplet composition and temperature profiles.

Keywords: absorption columns, aerosol formation, amine emissions, internal droplet profiles, monoethanolamine (MEA), post combustion CO2 capture, simulation

Procedia PDF Downloads 246
4680 An Algorithm to Depreciate the Energy Utilization Using a Bio-Inspired Method in Wireless Sensor Network

Authors: Navdeep Singh Randhawa, Shally Sharma

Abstract:

Wireless Sensor Network is an autonomous technology emanating in the current scenario at a fast pace. This technology faces a number of defiance’s and energy management is one of them, which has a huge impact on the network lifetime. To sustain energy the different types of routing protocols have been flourished. The classical routing protocols are no more compatible to perform in complicated environments. Hence, in the field of routing the intelligent algorithms based on nature systems is a turning point in Wireless Sensor Network. These nature-based algorithms are quite efficient to handle the challenges of the WSN as they are capable of achieving local and global best optimization solutions for the complex environments. So, the main attention of this paper is to develop a routing algorithm based on some swarm intelligent technique to enhance the performance of Wireless Sensor Network.

Keywords: wireless sensor network, routing, swarm intelligence, MPRSO

Procedia PDF Downloads 352
4679 Algorithm for Recognizing Trees along Power Grid Using Multispectral Imagery

Authors: C. Hamamura, V. Gialluca

Abstract:

Much of the Eclectricity Distributors has about 70% of its electricity interruptions arising from cause "trees", alone or associated with wind and rain and with or without falling branch and / or trees. This contributes inexorably and significantly to outages, resulting in high costs as compensation in addition to the operation and maintenance costs. On the other hand, there is little data structure and solutions to better organize the trees pruning plan effectively, minimizing costs and environmentally friendly. This work describes the development of an algorithm to provide data of trees associated to power grid. The method is accomplished on several steps using satellite imagery and geographically vectorized grid. A sliding window like approach is performed to seek the area around the grid. The proposed method counted 764 trees on a patch of the grid, which was very close to the 738 trees counted manually. The trees data was used as a part of a larger project that implements a system to optimize tree pruning plan.

Keywords: image pattern recognition, trees pruning, trees recognition, neural network

Procedia PDF Downloads 499
4678 Application First and Second Digits Number in the Benford Law

Authors: Teguh Sugiarto

Abstract:

Background: This study aims to explore the fraud that occurred in the financial statements using the Benford distribution law of 1st and 2nd case study of PT AKR Corporindo Tbk. Research Methods: In this study the authors use the first digit of the analysis and the analysis of the second digit of Bedford’s law. Having obtained the results of the analysis of the first and second digits, authors will make the difference between implementations using the scale above and below 5%. The number that has the level of difference in the range of 5% above or below, then a financial report in may, to analyse in the followup to the direction of the audit investigation, and authors assume happens a confusion in the financial statements. Findings: From research done, we found that there was a difference in the results of the appearance of the first digit of the number with the proper use of Benford's law, according to PT AKR Corporindo financial reports Tbk for the fiscal year 2006-2010, above and below the level the difference in set 5%. Conclusions: From the research that has been done, it can be concluded that on PT AKR Corporindo financial report 2006, 2007, 2008, 2009 and 2010, there is a level difference of appearance of numbers according to Benford's law is significant, as presented in the table analysis.

Keywords: Benford law, first digits, second digits, Indonesian company

Procedia PDF Downloads 427
4677 Descriptive Analysis of Community-Based Needs among Asylum Seekers in New England before and after COVID-19

Authors: Viknesh Kasthuri, Victoria Angenent-Mari, Jade Wexler

Abstract:

The COVID-19 pandemic dramatically altered the landscape of asylum medicine. Brown Human Rights Asylum Clinic (BHRAC) is a medical-student-run asylum clinic that provides pro-bono medical evaluations and forensic affidavits for individuals seeking asylum in New England. After the outbreak of COVID-19 in March 2020, BHRAC experienced numerous changes both in the number of clients requesting services as well as in the resource needs of these clients. Uniquely, BHRAC assesses the needs of clients during their affidavit interview and seeks to address these needs by connecting clients to local community organizations and resources. Data regarding the specific needs of clients range from 2019-present day. Analysis of internal BHRAC’s internal data suggested a small increase in requests for assistance with light and gas (from 5% of total resource requests pre-COVID to 11%), as well as a decrease in requests for mental health services (from 20% of resources pre-COVID to 13% post-COVID). Furthermore, BHRAC witnessed a decline in clinic volume during the second half of 2020. In short, our data suggest that the pandemic affected asylum seekers' access to medico-legal services and the resources they need. Future research with larger sample sizes and in other geographic locations is required to determine the holistic impact of the COVID-19 pandemic on asylum seekers.

Keywords: asylum clinic, asylum medicine, COVID, social determinants of health

Procedia PDF Downloads 103
4676 Transfer Knowledge From Multiple Source Problems to a Target Problem in Genetic Algorithm

Authors: Terence Soule, Tami Al Ghamdi

Abstract:

To study how to transfer knowledge from multiple source problems to the target problem, we modeled the Transfer Learning (TL) process using Genetic Algorithms as the model solver. TL is the process that aims to transfer learned data from one problem to another problem. The TL process aims to help Machine Learning (ML) algorithms find a solution to the problems. The Genetic Algorithms (GA) give researchers access to information that we have about how the old problem is solved. In this paper, we have five different source problems, and we transfer the knowledge to the target problem. We studied different scenarios of the target problem. The results showed combined knowledge from multiple source problems improves the GA performance. Also, the process of combining knowledge from several problems results in promoting diversity of the transferred population.

Keywords: transfer learning, genetic algorithm, evolutionary computation, source and target

Procedia PDF Downloads 140
4675 Saving Energy at a Wastewater Treatment Plant through Electrical and Production Data Analysis

Authors: Adriano Araujo Carvalho, Arturo Alatrista Corrales

Abstract:

This paper intends to show how electrical energy consumption and production data analysis were used to find opportunities to save energy at Taboada wastewater treatment plant in Callao, Peru. In order to access the data, it was used independent data networks for both electrical and process instruments, which were taken to analyze under an ISO 50001 energy audit, which considered, thus, Energy Performance Indexes for each process and a step-by-step guide presented in this text. Due to the use of aforementioned methodology and data mining techniques applied on information gathered through electronic multimeters (conveniently placed on substation switchboards connected to a cloud network), it was possible to identify thoroughly the performance of each process and thus, evidence saving opportunities which were previously hidden before. The data analysis brought both costs and energy reduction, allowing the plant to save significant resources and to be certified under ISO 50001.

Keywords: energy and production data analysis, energy management, ISO 50001, wastewater treatment plant energy analysis

Procedia PDF Downloads 193
4674 Disease Level Assessment in Wheat Plots Using a Residual Deep Learning Algorithm

Authors: Felipe A. Guth, Shane Ward, Kevin McDonnell

Abstract:

The assessment of disease levels in crop fields is an important and time-consuming task that generally relies on expert knowledge of trained individuals. Image classification in agriculture problems historically has been based on classical machine learning strategies that make use of hand-engineered features in the top of a classification algorithm. This approach tends to not produce results with high accuracy and generalization to the classes classified by the system when the nature of the elements has a significant variability. The advent of deep convolutional neural networks has revolutionized the field of machine learning, especially in computer vision tasks. These networks have great resourcefulness of learning and have been applied successfully to image classification and object detection tasks in the last years. The objective of this work was to propose a new method based on deep learning convolutional neural networks towards the task of disease level monitoring. Common RGB images of winter wheat were obtained during a growing season. Five categories of disease levels presence were produced, in collaboration with agronomists, for the algorithm classification. Disease level tasks performed by experts provided ground truth data for the disease score of the same winter wheat plots were RGB images were acquired. The system had an overall accuracy of 84% on the discrimination of the disease level classes.

Keywords: crop disease assessment, deep learning, precision agriculture, residual neural networks

Procedia PDF Downloads 331
4673 Federal Bureau of Investigation Opposition to German Nationalist Organizations in the United States (1941-45)

Authors: Yaroslav Alexandrovich Levin

Abstract:

In modern research on the history of the United States in World War II, it is quite popular to study the opposition of the American special services and, in particular, the Federal Bureau of Investigation (FBI) to various organizations of the German diasporas in new historical conditions. The appeal to traditional methods of historical research, comparative studies, and the principles of historicism will make it possible to more accurately trace the process of tightening the counterintelligence work of the Bureau and the close connection of concerns about the involvement of public organizations in the intelligence activities of the enemy. The broadcast of nationalist ideas by various communities of Germans under the auspices of their governments quickly attracted the attention of the FBI, which is in the process of consolidating its powers as the main US counterintelligence service. At the same time, the investigations and trials conducted by the John Edgar Hoover Department following these investigations often had an openly political color and increasingly consolidated the beginning of a political investigation in this service. This practice and its implementation ran into a tough contradiction between the legal norms of America, which proclaimed "democratic values," the right to freedom of speech, and the need to strengthen the internal security of the state and society in wartime. All these processes and the associated nuances and complexities are considered in specific examples of the work of federal agents against various pro-German organizations in the period 1941-45.

Keywords: World War II, internal security, countering extremism, counterintelligence, political investigation, FBI

Procedia PDF Downloads 86
4672 Application of Smplify-X Algorithm with Enhanced Gender Classifier in 3D Human Pose Estimation

Authors: Jiahe Liu, Hongyang Yu, Miao Luo, Feng Qian

Abstract:

The widespread application of 3D human body reconstruction spans various fields. Smplify-X, an algorithm reliant on single-image input, employs three distinct body parameter templates, necessitating gender classification of individuals within the input image. Researchers employed a ResNet18 network to train a gender classifier within the Smplify-X framework, setting the threshold at 0.9, designating images falling below this threshold as having neutral gender. This model achieved 62.38% accurate predictions and 7.54% incorrect predictions. Our improvement involved refining the MobileNet network, resulting in a raised threshold of 0.97. Consequently, we attained 78.89% accurate predictions and a mere 0.2% incorrect predictions, markedly enhancing prediction precision and enabling more precise 3D human body reconstruction.

Keywords: SMPLX, mobileNet, gender classification, 3D human reconstruction

Procedia PDF Downloads 99
4671 An Ensemble-based Method for Vehicle Color Recognition

Authors: Saeedeh Barzegar Khalilsaraei, Manoocheher Kelarestaghi, Farshad Eshghi

Abstract:

The vehicle color, as a prominent and stable feature, helps to identify a vehicle more accurately. As a result, vehicle color recognition is of great importance in intelligent transportation systems. Unlike conventional methods which use only a single Convolutional Neural Network (CNN) for feature extraction or classification, in this paper, four CNNs, with different architectures well-performing in different classes, are trained to extract various features from the input image. To take advantage of the distinct capability of each network, the multiple outputs are combined using a stack generalization algorithm as an ensemble technique. As a result, the final model performs better than each CNN individually in vehicle color identification. The evaluation results in terms of overall average accuracy and accuracy variance show the proposed method’s outperformance compared to the state-of-the-art rivals.

Keywords: Vehicle Color Recognition, Ensemble Algorithm, Stack Generalization, Convolutional Neural Network

Procedia PDF Downloads 85
4670 Iterative Segmentation and Application of Hausdorff Dilation Distance in Defect Detection

Authors: S. Shankar Bharathi

Abstract:

Inspection of surface defects on metallic components has always been challenging due to its specular property. Occurrences of defects such as scratches, rust, pitting are very common in metallic surfaces during the manufacturing process. These defects if unchecked can hamper the performance and reduce the life time of such component. Many of the conventional image processing algorithms in detecting the surface defects generally involve segmentation techniques, based on thresholding, edge detection, watershed segmentation and textural segmentation. They later employ other suitable algorithms based on morphology, region growing, shape analysis, neural networks for classification purpose. In this paper the work has been focused only towards detecting scratches. Global and other thresholding techniques were used to extract the defects, but it proved to be inaccurate in extracting the defects alone. However, this paper does not focus on comparison of different segmentation techniques, but rather describes a novel approach towards segmentation combined with hausdorff dilation distance. The proposed algorithm is based on the distribution of the intensity levels, that is, whether a certain gray level is concentrated or evenly distributed. The algorithm is based on extraction of such concentrated pixels. Defective images showed higher level of concentration of some gray level, whereas in non-defective image, there seemed to be no concentration, but were evenly distributed. This formed the basis in detecting the defects in the proposed algorithm. Hausdorff dilation distance based on mathematical morphology was used to strengthen the segmentation of the defects.

Keywords: metallic surface, scratches, segmentation, hausdorff dilation distance, machine vision

Procedia PDF Downloads 427
4669 Assessment of Cattle Welfare Traveling Long Distance from Jessore (Indian Border) to Chittagong, Bangladesh

Authors: Mahabub Alam, Mohammad Mahmudul Hassan, M. Hasanuzzaman, M. Ahasanul Hoque

Abstract:

Animals are transported from one place to another for different purposes in Bangladesh. However, the potential effect of long-distance transport on cattle health has not frequently been studied. Therefore, this study was conducted to assess health conditions of cattle transported from a long distance to Chittagong in Bangladesh. A total of 100 adult cattle, regardless of breed and sex, were selected at Benapole live cattle market in Jessore between August and September 2015 for the study. Blood samples were taken from 50 randomly selected cattle at 0 hours before transportation, just after transportation, at 12-16 hours post-conclusion of transportation, and 24 hours after transportation. The external health conditions and injuries of the cattle were assessed by close inspection, and the trader was interviewed using the structured questionnaire. Images of cattle injuries were taken with a camera. The basic internal health of the cattle was evaluated using standard hemato-biochemical tests. Animals were fasted and remained standing within a small space allocation (8-10 sq feet/animal) in the vehicle during transportation. Animals were provided only with paddy straw and water prior to selling at the destination market. The overall frequency of cattle injuries varied significantly (26% before vs. 47% after transportation; p < 0.001). The frequency of different cattle injuries also significantly varied by types such as abrasion (11% vs. 21%; p < 0.05) and barbed wire injury (9% vs. 18%; p < 0.05). Single cattle injury differed significantly (21% vs. 36%; p < 0.001). Cattle health conditions varied significantly (nasal discharge: 15% vs. 28%; p < 0.05; diarrhea: 15% vs. 23%; p < 0.05 and severe dehydration: 8% vs. 20%; p < 0.001). The values of hemoglobin (Hb), total erythrocyte count (TEC), total leukocyte count (TLC), lymphocyte (L), neutrophil (N) and eosinophil (E) varied significantly (p ≤ 0.01) (Hb: 11.1mg/dl vs. 12.3mg/dl; TEC: 4.7 million/ml vs. 5.7million/ml; TLC: 6.2 thousand/ml vs. 7.3 thousand/ml; L: 61.7% vs. 58.1%; N: 29.7% vs. 32.8%; E: 3.8% vs. 4.7%). The values of serum total protein (TP), creatine kinase (CK), triglyceride (TG), calcium (Ca), phosphorus (P) and alkaline phosphatase (ALP) significantly differed (p ≤ 0.05) (TP: 6.8g/dl vs. 8.2g/dl; CK:574.9u/l vs. 1288u/l; TG: 104.7mg/dl vs. 127.7mg/dl; Ca: 11.3mg/dl vs. 13mg/dl; P: 7.3mg/dl vs. 7.6mg/dl; ALP: 303u/l vs. 363u/l). The identified status of external and internal health conditions of the cattle for trading purpose due to long-distance transportation in the present study indicates a high degree of transport stress and poor animal welfare.

Keywords: animal welfare, cattle, external and internal health conditions, transportation

Procedia PDF Downloads 181
4668 Hit-Or-Miss Transform as a Tool for Similar Shape Detection

Authors: Osama Mohamed Elrajubi, Idris El-Feghi, Mohamed Abu Baker Saghayer

Abstract:

This paper describes an identification of specific shapes within binary images using the morphological Hit-or-Miss Transform (HMT). Hit-or-Miss transform is a general binary morphological operation that can be used in searching of particular patterns of foreground and background pixels in an image. It is actually a basic operation of binary morphology since almost all other binary morphological operators are derived from it. The input of this method is a binary image and a structuring element (a template which will be searched in a binary image) while the output is another binary image. In this paper a modification of Hit-or-Miss transform has been proposed. The accuracy of algorithm is adjusted according to the similarity of the template and the sought template. The implementation of this method has been done by C language. The algorithm has been tested on several images and the results have shown that this new method can be used for similar shape detection.

Keywords: hit-or-miss operator transform, HMT, binary morphological operation, shape detection, binary images processing

Procedia PDF Downloads 332
4667 Generating 3D Anisotropic Centroidal Voronoi Tessellations

Authors: Alexandre Marin, Alexandra Bac, Laurent Astart

Abstract:

New numerical methods for PDE resolution (such as Finite Volumes (FV) or Virtual Elements Method (VEM)) open new needs in terms of meshing of domains of interest, and in particular, polyhedral meshes have many advantages. One way to build such meshes consists of constructing Restricted Voronoi Diagrams (RVDs) whose boundaries respect the domain of interest. By minimizing a function defined for RVDs, the shapes of cells can be controlled, e.g., elongated according to user-defined directions or adjusted to comply with given aspect ratios (anisotropy) and density variations. In this paper, our contribution is threefold: First, we introduce a new gradient formula for the Voronoi tessellation energy under a continuous anisotropy field. Second, we describe a meshing algorithm based on the optimisation of this function that we validate against state-of-the-art approaches. Finally, we propose a hierarchical approach to speed up our meshing algorithm.

Keywords: anisotropic Voronoi diagrams, meshes for numerical simulations, optimisation, volumic polyhedral meshing

Procedia PDF Downloads 116
4666 Human Posture Estimation Based on Multiple Viewpoints

Authors: Jiahe Liu, HongyangYu, Feng Qian, Miao Luo

Abstract:

This study aimed to address the problem of improving the confidence of key points by fusing multi-view information, thereby estimating human posture more accurately. We first obtained multi-view image information and then used the MvP algorithm to fuse this multi-view information together to obtain a set of high-confidence human key points. We used these as the input for the Spatio-Temporal Graph Convolution (ST-GCN). ST-GCN is a deep learning model used for processing spatio-temporal data, which can effectively capture spatio-temporal relationships in video sequences. By using the MvP algorithm to fuse multi-view information and inputting it into the spatio-temporal graph convolution model, this study provides an effective method to improve the accuracy of human posture estimation and provides strong support for further research and application in related fields.

Keywords: multi-view, pose estimation, ST-GCN, joint fusion

Procedia PDF Downloads 70
4665 The Use Support Vector Machine and Back Propagation Neural Network for Prediction of Daily Tidal Levels Along The Jeddah Coast, Saudi Arabia

Authors: E. A. Mlybari, M. S. Elbisy, A. H. Alshahri, O. M. Albarakati

Abstract:

Sea level rise threatens to increase the impact of future storms and hurricanes on coastal communities. Accurate sea level change prediction and supplement is an important task in determining constructions and human activities in coastal and oceanic areas. In this study, support vector machines (SVM) is proposed to predict daily tidal levels along the Jeddah Coast, Saudi Arabia. The optimal parameter values of kernel function are determined using a genetic algorithm. The SVM results are compared with the field data and with back propagation (BP). Among the models, the SVM is superior to BPNN and has better generalization performance.

Keywords: tides, prediction, support vector machines, genetic algorithm, back-propagation neural network, risk, hazards

Procedia PDF Downloads 468