Search results for: line search algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7502

Search results for: line search algorithm

5792 Optimal Relaxation Parameters for Obtaining Efficient Iterative Methods for the Solution of Electromagnetic Scattering Problems

Authors: Nadaniela Egidi, Pierluigi Maponi

Abstract:

The approximate solution of a time-harmonic electromagnetic scattering problem for inhomogeneous media is required in several application contexts, and its two-dimensional formulation is a Fredholm integral equation of the second kind. This integral equation provides a formulation for the direct scattering problem, but it has to be solved several times also in the numerical solution of the corresponding inverse scattering problem. The discretization of this Fredholm equation produces large and dense linear systems that are usually solved by iterative methods. In order to improve the efficiency of these iterative methods, we use the Symmetric SOR preconditioning, and we propose an algorithm for the evaluation of the associated relaxation parameter. We show the efficiency of the proposed algorithm by several numerical experiments, where we use two Krylov subspace methods, i.e., Bi-CGSTAB and GMRES.

Keywords: Fredholm integral equation, iterative method, preconditioning, scattering problem

Procedia PDF Downloads 87
5791 Cross Reactivity of Risperidone in Fentanyl Point of Care Devices

Authors: Barry D. Kyle, Jessica Boyd, Robin Pickersgill, Nicole Squires, Cynthia Balion

Abstract:

Background-Aim: Fentanyl is a highly-potent synthetic μ-opioid receptor agonist used for exceptional pain management. Its main metabolite, norfentanyl, is typically present in urine at significantly high concentrations (i.e. ~20%) representing an effective targeting molecule for immunoassay detection. Here, we evaluated the NCSTM One Step Fentanyl Test Device© and the BTNX Rapid ResponseTM Single Drug Test Strip© point of care (POC) test strips targeting norfentanyl (20 ng/ml) and fentanyl (100 ng/ml) molecules for potential risperidone interference. Methods: POC tests calibrated against norfentanyl (20 ng/ml) used [immunochromatographic] lateral flow devices to provide qualitative results within five minutes of urine sample contact. Results were recorded as negative if lines appeared in the test and control regions according to manufacturer’s instructions. Positive results were recorded if no line appeared in the test region (i.e., control line only visible). Pooled patient urine (n=20), that screened negative for drugs of abuse (using NCS One Step Multi-Line Screen) and fentanyl (using BTNX Rapid Response Strip) was used for spiking studies. Urine was spiked with risperidone alone and with combinations of fentanyl, norfentanyl and/or risperidone to evaluate cross-reactivity in each test device. Results: A positive screen result was obtained when 8,000 ng/mL of risperidone was spiked into drug free urine using the NCS test device. Positive screen results were also obtained in spiked urine samples containing fentanyl and norfentanyl combinations below the cut-off concentrations when 4000 ng/mL risperidone was present using the NCS testing device. There were no screen positive test results using the BTNX test strip with up to 8,000 ng/mL alone or in combination with concentrations of fentanyl and norfentanyl below the cut-off. Both devices screened positive when either fentanyl or norfentanyl exceeded the cut-off threshold in the absence and presence of risperidone. Conclusion: We report that urine samples containing risperidone may give a false positive result using the NCS One Step Fentanyl Test Device.

Keywords: fentanyl, interferences, point of care test, Risperidone

Procedia PDF Downloads 252
5790 Matching on Bipartite Graphs with Applications to School Course Registration Systems

Authors: Zhihan Li

Abstract:

Nowadays, most universities use the course enrollment system considering students’ registration orders. However, the students’ preference level to certain courses is also one important factor to consider. In this research, the possibility of applying a preference-first system has been discussed and analyzed compared to the order-first system. A bipartite graph is applied to resemble the relationship between students and courses they tend to register. With the graph set up, we apply Ford-Fulkerson (F.F.) Algorithm to maximize parings between two sets of nodes, in our case, students and courses. Two models are proposed in this paper: the one considered students’ order first, and the one considered students’ preference first. By comparing and contrasting the two models, we highlight the usability of models which potentially leads to better designs for school course registration systems.

Keywords: bipartite graph, Ford-Fulkerson (F.F.) algorithm, graph theory, maximum matching

Procedia PDF Downloads 96
5789 Coordinated Interference Canceling Algorithm for Uplink Massive Multiple Input Multiple Output Systems

Authors: Messaoud Eljamai, Sami Hidouri

Abstract:

Massive multiple-input multiple-output (MIMO) is an emerging technology for new cellular networks such as 5G systems. Its principle is to use many antennas per cell in order to maximize the network's spectral efficiency. Inter-cellular interference remains a fundamental problem. The use of massive MIMO will not derogate from the rule. It improves performances only when the number of antennas is significantly greater than the number of users. This, considerably, limits the networks spectral efficiency. In this paper, a coordinated detector for an uplink massive MIMO system is proposed in order to mitigate the inter-cellular interference. The proposed scheme combines the coordinated multipoint technique with an interference-cancelling algorithm. It requires the serving cell to send their received symbols, after processing, decision and error detection, to the interfered cells via a backhaul link. Each interfered cell is capable of eliminating intercellular interferences by generating and subtracting the user’s contribution from the received signal. The resulting signal is more reliable than the original received signal. This allows the uplink massive MIMO system to improve their performances dramatically. Simulation results show that the proposed detector improves system spectral efficiency compared to classical linear detectors.

Keywords: massive MIMO, COMP, interference canceling algorithm, spectral efficiency

Procedia PDF Downloads 132
5788 Application of Artificial Immune Systems Combined with Collaborative Filtering in Movie Recommendation System

Authors: Pei-Chann Chang, Jhen-Fu Liao, Chin-Hung Teng, Meng-Hui Chen

Abstract:

This research combines artificial immune system with user and item based collaborative filtering to create an efficient and accurate recommendation system. By applying the characteristic of antibodies and antigens in the artificial immune system and using Pearson correlation coefficient as the affinity threshold to cluster the data, our collaborative filtering can effectively find useful users and items for rating prediction. This research uses MovieLens dataset as our testing target to evaluate the effectiveness of the algorithm developed in this study. The experimental results show that the algorithm can effectively and accurately predict the movie ratings. Compared to some state of the art collaborative filtering systems, our system outperforms them in terms of the mean absolute error on the MovieLens dataset.

Keywords: artificial immune system, collaborative filtering, recommendation system, similarity

Procedia PDF Downloads 517
5787 Path Planning for Unmanned Aerial Vehicles in Constrained Environments for Locust Elimination

Authors: Aadiv Shah, Hari Nair, Vedant Mittal, Alice Cheeran

Abstract:

Present-day agricultural practices such as blanket spraying not only lead to excessive usage of pesticides but also harm the overall crop yield. This paper introduces an algorithm to optimize the traversal of an unmanned aerial vehicle (UAV) in constrained environments. The proposed system focuses on the agricultural application of targeted spraying for locust elimination. Given a satellite image of a farm, target zones that are prone to locust swarm formation are detected through the calculation of the normalized difference vegetation index (NDVI). This is followed by determining the optimal path for traversal of a UAV through these target zones using the proposed algorithm in order to perform pesticide spraying in the most efficient manner possible. Unlike the classic travelling salesman problem involving point-to-point optimization, the proposed algorithm determines an optimal path for multiple regions, independent of its geometry. Finally, the paper explores the idea of implementing reinforcement learning to model complex environmental behaviour and make the path planning mechanism for UAVs agnostic to external environment changes. This system not only presents a solution to the enormous losses incurred due to locust attacks but also an efficient way to automate agricultural practices across the globe in order to improve farmer ergonomics.

Keywords: locust, NDVI, optimization, path planning, reinforcement learning, UAV

Procedia PDF Downloads 231
5786 Highly Accurate Target Motion Compensation Using Entropy Function Minimization

Authors: Amin Aghatabar Roodbary, Mohammad Hassan Bastani

Abstract:

One of the defects of stepped frequency radar systems is their sensitivity to target motion. In such systems, target motion causes range cell shift, false peaks, Signal to Noise Ratio (SNR) reduction and range profile spreading because of power spectrum interference of each range cell in adjacent range cells which induces distortion in High Resolution Range Profile (HRRP) and disrupt target recognition process. Thus Target Motion Parameters (TMPs) effects compensation should be employed. In this paper, such a method for estimating TMPs (velocity and acceleration) and consequently eliminating or suppressing the unwanted effects on HRRP based on entropy minimization has been proposed. This method is carried out in two major steps: in the first step, a discrete search method has been utilized over the whole acceleration-velocity lattice network, in a specific interval seeking to find a less-accurate minimum point of the entropy function. Then in the second step, a 1-D search over velocity is done in locus of the minimum for several constant acceleration lines, in order to enhance the accuracy of the minimum point found in the first step. The provided simulation results demonstrate the effectiveness of the proposed method.

Keywords: automatic target recognition (ATR), high resolution range profile (HRRP), motion compensation, stepped frequency waveform technique (SFW), target motion parameters (TMPs)

Procedia PDF Downloads 138
5785 Development of Construction Cost Optimization System Using Genetic Algorithm Method

Authors: Hyeon-Seung Kim, Young-Hwan Kim, Sang-Mi Park, Min-Seo Kim, Jong-Myeung Shin, Leen-Seok Kang

Abstract:

The project budget at the planned stage might be changed by the insufficient government budget or the design change. There are many cases more especially in the case of a project performed for a long period of time. If the actual construction budget is insufficient comparing with the planned budget, the construction schedule should also be changed to match the changed budget. In that case, most project managers change the planned construction schedule by a heuristic approach without a reasonable consideration on the work priority. This study suggests an optimized methodology to modify the construction schedule according to the changed budget. The genetic algorithm was used to optimize the modified construction schedule within the changed budget. And a simulation system of construction cost histogram in accordance with the construction schedule was developed in the BIM (Building Information Modeling) environment.

Keywords: 5D, BIM, GA, cost optimization

Procedia PDF Downloads 577
5784 Optimization Based Extreme Learning Machine for Watermarking of an Image in DWT Domain

Authors: RAM PAL SINGH, VIKASH CHAUDHARY, MONIKA VERMA

Abstract:

In this paper, we proposed the implementation of optimization based Extreme Learning Machine (ELM) for watermarking of B-channel of color image in discrete wavelet transform (DWT) domain. ELM, a regularization algorithm, works based on generalized single-hidden-layer feed-forward neural networks (SLFNs). However, hidden layer parameters, generally called feature mapping in context of ELM need not to be tuned every time. This paper shows the embedding and extraction processes of watermark with the help of ELM and results are compared with already used machine learning models for watermarking.Here, a cover image is divide into suitable numbers of non-overlapping blocks of required size and DWT is applied to each block to be transformed in low frequency sub-band domain. Basically, ELM gives a unified leaning platform with a feature mapping, that is, mapping between hidden layer and output layer of SLFNs, is tried for watermark embedding and extraction purpose in a cover image. Although ELM has widespread application right from binary classification, multiclass classification to regression and function estimation etc. Unlike SVM based algorithm which achieve suboptimal solution with high computational complexity, ELM can provide better generalization performance results with very small complexity. Efficacy of optimization method based ELM algorithm is measured by using quantitative and qualitative parameters on a watermarked image even though image is subjected to different types of geometrical and conventional attacks.

Keywords: BER, DWT, extreme leaning machine (ELM), PSNR

Procedia PDF Downloads 292
5783 An Optimization Model for Maximum Clique Problem Based on Semidefinite Programming

Authors: Derkaoui Orkia, Lehireche Ahmed

Abstract:

The topic of this article is to exploring the potentialities of a powerful optimization technique, namely Semidefinite Programming, for solving NP-hard problems. This approach provides tight relaxations of combinatorial and quadratic problems. In this work, we solve the maximum clique problem using this relaxation. The clique problem is the computational problem of finding cliques in a graph. It is widely acknowledged for its many applications in real-world problems. The numerical results show that it is possible to find a maximum clique in polynomial time, using an algorithm based on semidefinite programming. We implement a primal-dual interior points algorithm to solve this problem based on semidefinite programming. The semidefinite relaxation of this problem can be solved in polynomial time.

Keywords: semidefinite programming, maximum clique problem, primal-dual interior point method, relaxation

Procedia PDF Downloads 205
5782 Optimization of Feeder Bus Routes at Urban Rail Transit Stations Based on Link Growth Probability

Authors: Yu Song, Yuefei Jin

Abstract:

Urban public transportation can be integrated when there is an efficient connection between urban rail lines, however, there are currently no effective or quick solutions being investigated for this connection. This paper analyzes the space-time distribution and travel demand of passenger connection travel based on taxi track data and data from the road network, excavates potential bus connection stations based on potential connection demand data, and introduces the link growth probability model in the complex network to solve the basic connection bus lines in order to ascertain the direction of the bus lines that are the most connected given the demand characteristics. Then, a tree view exhaustive approach based on constraints is suggested based on graph theory, which can hasten the convergence of findings while doing chain calculations. This study uses WEI QU NAN Station, the Xi'an Metro Line 2 terminal station in Shaanxi Province, as an illustration, to evaluate the model's and the solution method's efficacy. According to the findings, 153 prospective stations have been dug up in total, the feeder bus network for the entire line has been laid out, and the best route adjustment strategy has been found.

Keywords: feeder bus, route optimization, link growth probability, the graph theory

Procedia PDF Downloads 59
5781 Training a Neural Network Using Input Dropout with Aggressive Reweighting (IDAR) on Datasets with Many Useless Features

Authors: Stylianos Kampakis

Abstract:

This paper presents a new algorithm for neural networks called “Input Dropout with Aggressive Re-weighting” (IDAR) aimed specifically at datasets with many useless features. IDAR combines two techniques (dropout of input neurons and aggressive re weighting) in order to eliminate the influence of noisy features. The technique can be seen as a generalization of dropout. The algorithm is tested on two different benchmark data sets: a noisy version of the iris dataset and the MADELON data set. Its performance is compared against three other popular techniques for dealing with useless features: L2 regularization, LASSO and random forests. The results demonstrate that IDAR can be an effective technique for handling data sets with many useless features.

Keywords: neural networks, feature selection, regularization, aggressive reweighting

Procedia PDF Downloads 436
5780 Continuous Blood Pressure Measurement from Pulse Transit Time Techniques

Authors: Chien-Lin Wang, Cha-Ling Ko, Tainsong Chen

Abstract:

Pulse Blood pressure (BP) is one of the vital signs, and is an index that helps determining the stability of life. In this respect, some spinal cord injury patients need to take the tilt table test. While doing the test, the posture changes abruptly, and may cause a patient’s BP to change abnormally. This may cause patients to feel discomfort, and even feel as though their life is threatened. Therefore, if a continuous non-invasive BP assessment system were built, it could help to alert health care professionals in the process of rehabilitation when the BP value is out of range. In our research, BP assessed by the pulse transit time technique was developed. In the system, we use a self-made photoplethysmograph (PPG) sensor and filter circuit to detect two PPG signals and to calculate the time difference. The BP can immediately be assessed by the trend line. According to the results of this study, the relationship between the systolic BP and PTT has a highly negative linear correlation (R2=0.8). Further, we used the trend line to assess the value of the BP and compared it to a commercial sphygmomanometer (Omron MX3); the error rate of the system was found to be in the range of ±10%, which is within the permissible error range of a commercial sphygmomanometer. The continue blood pressure measurement from pulse transit time technique may have potential to become a convenience method for clinical rehabilitation.

Keywords: continous blood pressure measurement, PPG, time transit time, transit velocity

Procedia PDF Downloads 333
5779 EEG Signal Processing Methods to Differentiate Mental States

Authors: Sun H. Hwang, Young E. Lee, Yunhan Ga, Gilwon Yoon

Abstract:

EEG is a very complex signal with noises and other bio-potential interferences. EOG is the most distinct interfering signal when EEG signals are measured and analyzed. It is very important how to process raw EEG signals in order to obtain useful information. In this study, the EEG signal processing techniques such as EOG filtering and outlier removal were examined to minimize unwanted EOG signals and other noises. The two different mental states of resting and focusing were examined through EEG analysis. A focused state was induced by letting subjects to watch a red dot on the white screen. EEG data for 32 healthy subjects were measured. EEG data after 60-Hz notch filtering were processed by a commercially available EOG filtering and our presented algorithm based on the removal of outliers. The ratio of beta wave to theta wave was used as a parameter for determining the degree of focusing. The results show that our algorithm was more appropriate than the existing EOG filtering.

Keywords: EEG, focus, mental state, outlier, signal processing

Procedia PDF Downloads 267
5778 Offline Signature Verification in Punjabi Based On SURF Features and Critical Point Matching Using HMM

Authors: Rajpal Kaur, Pooja Choudhary

Abstract:

Biometrics, which refers to identifying an individual based on his or her physiological or behavioral characteristics, has the capabilities to the reliably distinguish between an authorized person and an imposter. The Signature recognition systems can categorized as offline (static) and online (dynamic). This paper presents Surf Feature based recognition of offline signatures system that is trained with low-resolution scanned signature images. The signature of a person is an important biometric attribute of a human being which can be used to authenticate human identity. However the signatures of human can be handled as an image and recognized using computer vision and HMM techniques. With modern computers, there is need to develop fast algorithms for signature recognition. There are multiple techniques are defined to signature recognition with a lot of scope of research. In this paper, (static signature) off-line signature recognition & verification using surf feature with HMM is proposed, where the signature is captured and presented to the user in an image format. Signatures are verified depended on parameters extracted from the signature using various image processing techniques. The Off-line Signature Verification and Recognition is implemented using Mat lab platform. This work has been analyzed or tested and found suitable for its purpose or result. The proposed method performs better than the other recently proposed methods.

Keywords: offline signature verification, offline signature recognition, signatures, SURF features, HMM

Procedia PDF Downloads 365
5777 Web Page Design Optimisation Based on Segment Analytics

Authors: Varsha V. Rohini, P. R. Shreya, B. Renukadevi

Abstract:

In the web analytics the information delivery and the web usage is optimized and the analysis of data is done. The analytics is the measurement, collection and analysis of webpage data. Page statistics and user metrics are the important factor in most of the web analytics tool. This is the limitation of the existing tools. It does not provide design inputs for the optimization of information. This paper aims at providing an extension for the scope of web analytics to provide analysis and statistics of each segment of a webpage. The number of click count is calculated and the concentration of links in a web page is obtained. Its user metrics are used to help in proper design of the displayed content in a webpage by Vision Based Page Segmentation (VIPS) algorithm. When the algorithm is applied on the web page it divides the entire web page into the visual block tree. The visual block tree generated will further divide the web page into visual blocks or segments which help us to understand the usage of each segment in a page and its content. The dynamic web pages and deep web pages are used to extend the scope of web page segment analytics. Space optimization concept is used with the help of the output obtained from the Vision Based Page Segmentation (VIPS) algorithm. This technique provides us the visibility of the user interaction with the WebPages and helps us to place the important links in the appropriate segments of the webpage and effectively manage space in a page and the concentration of links.

Keywords: analytics, design optimization, visual block trees, vision based technology

Procedia PDF Downloads 252
5776 Algorithms for Run-Time Task Mapping in NoC-Based Heterogeneous MPSoCs

Authors: M. K. Benhaoua, A. K. Singh, A. E. Benyamina, P. Boulet

Abstract:

Mapping parallelized tasks of applications onto these MPSoCs can be done either at design time (static) or at run-time (dynamic). Static mapping strategies find the best placement of tasks at design-time, and hence, these are not suitable for dynamic workload and seem incapable of runtime resource management. The number of tasks or applications executing in MPSoC platform can exceed the available resources, requiring efficient run-time mapping strategies to meet these constraints. This paper describes a new Spiral Dynamic Task Mapping heuristic for mapping applications onto NoC-based Heterogeneous MPSoC. This heuristic is based on packing strategy and routing Algorithm proposed also in this paper. Heuristic try to map the tasks of an application in a clustering region to reduce the communication overhead between the communicating tasks. The heuristic proposed in this paper attempts to map the tasks of an application that are most related to each other in a spiral manner and to find the best possible path load that minimizes the communication overhead. In this context, we have realized a simulation environment for experimental evaluations to map applications with varying number of tasks onto an 8x8 NoC-based Heterogeneous MPSoCs platform, we demonstrate that the new mapping heuristics with the new modified dijkstra routing algorithm proposed are capable of reducing the total execution time and energy consumption of applications when compared to state-of-the-art run-time mapping heuristics reported in the literature.

Keywords: multiprocessor system on chip, MPSoC, network on chip, NoC, heterogeneous architectures, run-time mapping heuristics, routing algorithm

Procedia PDF Downloads 473
5775 Mixing Behaviors of Shear-Thinning Fluids in Serpentine-Channel Micromixers

Authors: Rei-Tang Tsai, Chih-Yang Wu, Chia-Yuan Chang, Ming-Ying Kuo

Abstract:

This study aims to investigate the mixing behaviors of deionized (DI) water and carboxymethyl cellulose (CMC) solutions in C-shaped serpentine micromixers over a wide range of flow conditions. The flow of CMC solutions exhibits shear-thinning behaviors. Numerical simulations are performed to investigate the effects of the mean flow speed, fluid properties and geometry parameters on flow and mixing in the micromixers with serpentine channel of the same overall channel length. From the results, we can find the following trends. When fluid mixing is dominated by convection, the curvature-induced vortices enhance fluid mixing effectively. The mixing efficiency of a micromixer consisting of semicircular C-shaped repeating units with a smaller center-line radius is better than that of a micromixer consisting of major-segment repeating units with a larger center-line radius. The viscosity of DI water is less than the overall average apparent viscosity of CMC solutions, and so the effect of curvature-induced vortices on fluid mixing in DI water is larger than that in CMC solutions for the cases with the same mean flow speed.

Keywords: curved channel, microfluidics, mixing, non-newtonian fluids, vortex

Procedia PDF Downloads 424
5774 Simulation and Analysis of Inverted Pendulum Controllers

Authors: Sheren H. Salah

Abstract:

The inverted pendulum is a highly nonlinear and open-loop unstable system. An inverted pendulum (IP) is a pendulum which has its mass above its pivot point. It is often implemented with the pivot point mounted on a cart that can move horizontally and may be called a cart and pole. The characteristics of the inverted pendulum make identification and control more challenging. This paper presents the simulation study of several control strategies for an inverted pendulum system. The goal is to determine which control strategy delivers better performance with respect to pendulum’s angle. The inverted pendulum represents a challenging control problem, which continually moves toward an uncontrolled state. For controlling the inverted pendulum. The simulation study that sliding mode control (SMC) control produced better response compared to Genetic Algorithm Control (GAs) and proportional-integral-derivative(PID) control.

Keywords: Inverted Pendulum (IP) Proportional-Integral-Derivative (PID), Genetic Algorithm Control (GAs), Sliding Mode Control (SMC)

Procedia PDF Downloads 542
5773 Improvement of Ground Truth Data for Eye Location on Infrared Driver Recordings

Authors: Sorin Valcan, Mihail Gaianu

Abstract:

Labeling is a very costly and time consuming process which aims to generate datasets for training neural networks in several functionalities and projects. For driver monitoring system projects, the need for labeled images has a significant impact on the budget and distribution of effort. This paper presents the modifications done to an algorithm used for the generation of ground truth data for 2D eyes location on infrared images with drivers in order to improve the quality of the data and performance of the trained neural networks. The algorithm restrictions become tougher, which makes it more accurate but also less constant. The resulting dataset becomes smaller and shall not be altered by any kind of manual label adjustment before being used in the neural networks training process. These changes resulted in a much better performance of the trained neural networks.

Keywords: labeling automation, infrared camera, driver monitoring, eye detection, convolutional neural networks

Procedia PDF Downloads 93
5772 Pattern Recognition Approach Based on Metabolite Profiling Using In vitro Cancer Cell Line

Authors: Amanina Iymia Jeffree, Reena Thriumani, Mohammad Iqbal Omar, Ammar Zakaria, Yumi Zuhanis Has-Yun Hashim, Ali Yeon Md Shakaff

Abstract:

Metabolite profiling is a strategy to be approached in the pattern recognition method focused on three types of cancer cell line that driving the most to death specifically lung, breast, and colon cancer. The purpose of this study was to discriminate the VOCs pattern among cancerous and control group based on metabolite profiling. The sampling was executed utilizing the cell culture technique. All culture flasks were incubated till 72 hours and data collection started after 24 hours. Every running sample took 24 minutes to be completed accordingly. The comparative metabolite patterns were identified by the implementation of headspace-solid phase micro-extraction (HS-SPME) sampling coupled with gas chromatography-mass spectrometry (GCMS). The optimizations of the main experimental variables such as oven temperature and time were evaluated by response surface methodology (RSM) to get the optimal condition. Volatiles were acknowledged through the National Institute of Standards and Technology (NIST) mass spectral database and retention time libraries. To improve the reliability of significance, it is of crucial importance to eliminate background noise which data from 3rd minutes to 17th minutes were selected for statistical analysis. Targeted metabolites, of which were annotated as known compounds with the peak area greater than 0.5 percent were highlighted and subsequently treated statistically. Volatiles produced contain hundreds to thousands of compounds; therefore, it will be optimized by chemometric analysis, such as principal component analysis (PCA) as a preliminary analysis before subjected to a pattern classifier for identification of VOC samples. The volatile organic compound profiling has shown to be significantly distinguished among cancerous and control group based on metabolite profiling.

Keywords: in vitro cancer cell line, metabolite profiling, pattern recognition, volatile organic compounds

Procedia PDF Downloads 348
5771 Design of Single Point Mooring Buoy System by Parametric Analysis

Authors: Chul-Hee Jo, Do-Youb Kim, Seok-Jin Cho, Yu-Ho Rho

Abstract:

The Catenary Anchor Leg Mooring (CALM) Single Point Mooring (SPM) buoy system is the most popular and widely used type of offshore loading terminals. SPM buoy mooring systems have been deployed worldwide for a variety of applications, water depths and vessel sizes ranging from small production carriers to Very Large Crude Carriers (VLCCs). Because of safe and easy berthing and un-berthing operations, the SPM buoy mooring system is also preferred for offshore terminals. The SPM buoy consists of a buoy that is permanently moored to the seabed by means of multiple mooring lines. The buoy contains a bearing system that allows a part of it to rotate around the moored geostatic part. When moored to the rotating part of the buoy, a vessel is able to freely weathervane around the buoy. This study was verified the effects of design variables in order to design an SPM buoy mooring system through parametric analysis. The design variables have independent and nonlinear characteristics. Using parametric analysis, this research was found that the fairlead departure angle, wave height and period, chain diameter and line length effect to the mooring top tension, buoy excursion and line layback.

Keywords: Single Point Mooring (SPM), Catenary Anchor Leg Mooring(CALM), design variables, parametric analysis, mooring system optimization

Procedia PDF Downloads 370
5770 A Second Order Genetic Algorithm for Traveling Salesman Problem

Authors: T. Toathom, M. Munlin, P. Sugunnasil

Abstract:

The traveling salesman problem (TSP) is one of the best-known problems in optimization problem. There are many research regarding the TSP. One of the most usage tool for this problem is the genetic algorithm (GA). The chromosome of the GA for TSP is normally encoded by the order of the visited city. However, the traditional chromosome encoding scheme has some limitations which are twofold: the large solution space and the inability to encapsulate some information. The number of solution for a certain problem is exponentially grow by the number of city. Moreover, the traditional chromosome encoding scheme fails to recognize the misplaced correct relation. It implies that the tradition method focuses only on exact solution. In this work, we relax some of the concept in the GA for TSP which is the exactness of the solution. The proposed work exploits the relation between cities in order to reduce the solution space in the chromosome encoding. In this paper, a second order GA is proposed to solve the TSP. The term second order refers to how the solution is encoded into chromosome. The chromosome is divided into 2 types: the high order chromosome and the low order chromosome. The high order chromosome is the chromosome that focus on the relation between cities such as the city A should be visited before city B. On the other hand, the low order chromosome is a type of chromosome that is derived from a high order chromosome. In other word, low order chromosome is encoded by the traditional chromosome encoding scheme. The genetic operation, mutation and crossover, will be performed on the high order chromosome. Then, the high order chromosome will be mapped to a group of low order chromosomes whose characteristics are satisfied with the high order chromosome. From the mapped set of chromosomes, the champion chromosome will be selected based on the fitness value which will be later used as a representative for the high order chromosome. The experiment is performed on the city data from TSPLIB.

Keywords: genetic algorithm, traveling salesman problem, initial population, chromosomes encoding

Procedia PDF Downloads 255
5769 Delineation of Soil Physical Properties Using Electrical Conductivity, Case Study: Volcanic Soil Simulation Model

Authors: Twin Aji Kusumagiani, Eleonora Agustine, Dini Fitriani

Abstract:

The value changes of soil physical properties in the agricultural area are giving impacts on soil fertility. This can be caused by excessive usage of inorganic fertilizers and imbalances on organic fertilization. Soil physical parameters that can be measured include soil electrical conductivity, water content volume, soil porosity, dielectric permittivity, etc. This study used the electrical conductivity and volume water content as the measured physical parameters. The study was conducted on volcanic soil obtained from agricultural land conditioned with NPK fertilizer and salt in a certain amount. The dimension of the conditioned soil being used is 1 x 1 x 0.5 meters. By using this method, we can delineate the soil electrical conductivity value of land due to changes in the provision of inorganic NPK fertilizer and the salinity in the soil. Zone with the additional 1 kg of salt has the dimension of 60 cm in width, 20 cm in depth and 1 cm in thickness while zone with the additional of 10 kg NPK fertilizer has the dimensions of 70 cm in width, 20 cm in depth and 3 cm in thickness. This salt addition resulted in EC values changes from the original condition. Changes of the EC value tend to occur at a depth of 20 to 40 cm on the line 1B at 9:45 dS/cm and line 1C of 9.35 dS/cm and tend to have the direction to the Northeast.

Keywords: EC, electrical conductivity, VWC, volume water content, NPK fertilizer, salt, volcanic soil

Procedia PDF Downloads 298
5768 An Algorithm to Depreciate the Energy Utilization Using a Bio-Inspired Method in Wireless Sensor Network

Authors: Navdeep Singh Randhawa, Shally Sharma

Abstract:

Wireless Sensor Network is an autonomous technology emanating in the current scenario at a fast pace. This technology faces a number of defiance’s and energy management is one of them, which has a huge impact on the network lifetime. To sustain energy the different types of routing protocols have been flourished. The classical routing protocols are no more compatible to perform in complicated environments. Hence, in the field of routing the intelligent algorithms based on nature systems is a turning point in Wireless Sensor Network. These nature-based algorithms are quite efficient to handle the challenges of the WSN as they are capable of achieving local and global best optimization solutions for the complex environments. So, the main attention of this paper is to develop a routing algorithm based on some swarm intelligent technique to enhance the performance of Wireless Sensor Network.

Keywords: wireless sensor network, routing, swarm intelligence, MPRSO

Procedia PDF Downloads 338
5767 Risk Factors Associated with Outbreak of Cerebrospinal Meningitis in Kano State- Nigeria, March-May 2017

Authors: Visa I. Tyakaray, M. Abdulaziz, O. Badmus, N. Karaye, M. Dalhat, A. Shehu, I. Bello, T. Hussaini, S. Akar, G. Effah, P. Nguku

Abstract:

Introduction: Nigeria has recorded outbreaks of meningitis in the past, being in the meningitis belt. A multi-state outbreak of Cerebrospinal Meningitis (CSM) from Neisseria meningitides occurred in 2017 involving 24 states, and Kano State reported its first two confirmed CSM cases on 22nd March, 2017. We conducted the outbreak investigation to characterize the outbreak, determine its associated risk factors and institute appropriate control measures. Method: We conducted an unmatched Case-control study with ratio 1:2. A case was defined as any person with sudden onset of fever (>38.5˚C rectal or 38.0˚C axillary) and one of the following: neck stiffness, altered consciousness or bulging fontanelle in toddlers while a control was defined as any person who resides around the case such as family members, caregivers, neighbors, and healthcare personnel. We reviewed and validated line list and conducted active case search in health facilities and neighboring communities. Descriptive, bivariate, stratified and multivariate analysis were performed. Laboratory confirmation was by Latex agglutination and/or Culture. Results: We recruited 48 cases with median age of 11 years (1 month – 65 years), attack rate was 2.4/100,000 population with case fatality rate of 8%; 34 of 44 local government areas were affected.On stratification, age was found to be a confounder. Independent factors associated with the outbreak were age (Adjusted Odds Ratio, AOR =6.58; 95% Confidence Interval (CI) =2.85-15.180, history of Vaccination (AOR=0.37; 95% CI=0.13-0.99) and history of travel (AOR=10.16; (1.99-51.85). Laboratory results showed 22 positive cases for Neisseria meningitides types C and A/Y. Conclusion: Major risk factors associated with this outbreak were age (>14years), not being vaccinated and history of travel. We sensitized communities and strengthened case management. We recommended immediate reactive vaccination and enhanced surveillance in bordering communities.

Keywords: cerebrospinal, factors, Kano-Nigeria, meningitis, risk

Procedia PDF Downloads 196
5766 Reactive X Proactive Searches on Internet After Leprosy Institutional Campaigns in Brazil: A Google Trends Analysis

Authors: Paulo Roberto Vasconcellos-Silva

Abstract:

The "Janeiro Roxo" (Purple January) campaign in Brazil aims to promote awareness of leprosy and its early symptoms. The COVID-19 pandemic has adversely affected institutional campaigns, mostly considering leprosy a neglected disease by the media. Google Trends (GT) is a tool that tracks user searches on Google, providing insights into the popularity of specific search terms. Our prior research has categorized online searches into two types: "Reactive searches," driven by transient campaign-related stimuli, and "Proactive searches," driven by personal interest in early symptoms and self-diagnosis. Using GT we studied: (i) the impact of "Janeiro Roxo" on public interest in leprosy (assessed through reactive searches) and its early symptoms (evaluated through proactive searches) over the past five years; (ii) changes in public interest during and after the COVID-19 pandemic; (iii) patterns in the dynamics of reactive and proactive searches Methods: We used GT's "Relative Search Volume" (RSV) to gauge public interest on a scale from 0 to 100. "HANSENÍASE" (HAN) was a proxy for reactive searches, and "HANSENÍASE SINTOMAS" (leprosy symptoms) (H.SIN) for proactive searches (interest in leprosy or in self-diagnosis). We analyzed 261 weeks of data from 2018 to 2023, using polynomial trend lines to model trends over this period. Analysis of Variance (ANOVA) was used to compare weekly RSV, monthly (MM) and annual means (AM). Results: Over a span of 261 weeks, there was consistently higher Relative Search Volume (RSV) for HAN compared to H.SIN. Both search terms exhibited their highest (MM) in January months during all periods. COVID-19 pandemic: a decline was observed during the pandemic years (2020-2021). There was a 24% decrease in RSV for HAN and a 32.5% decrease for H.SIN. Both HAN and H.SIN regained their pre-pandemic search levels in January 2022-2023. Breakpoints indicated abrupt changes - in the 26th week (February 2019), 55th and 213th weeks (September 2019 and 2022) related to September regional campaigns (interrupted in 2020-2021). Trend lines for HAN exhibited an upward curve between 33rd-45th week (April to June 2019), a pandemic-related downward trend between 120th-136th week (December 2020 to March 2021), and an upward trend between 220th-240th week (November 2022 to March 2023). Conclusion: The "Janeiro Roxo" campaign, along with other media-driven activities, exerts a notable influence on both reactive and proactive searches related to leprosy topics. Reactive searches, driven by campaign stimuli, significantly outnumber proactive searches. Despite the interruption of the campaign due to the pandemic, there was a subsequent resurgence in both types of searches. The recovery observed in reactive and proactive searches post-campaign interruption underscores the effectiveness of such initiatives, particularly at the national level. This suggests that regional campaigns aimed at leprosy awareness can be considered highly successful in stimulating proactive public engagement. The evaluation of internet-based campaign programs proves valuable not only for assessing their impact but also for identifying the needs of vulnerable regions. These programs can play a crucial role in integrating regions and highlighting their needs for assistance services in the context of leprosy awareness.

Keywords: health communication, leprosy, health campaigns, information seeking behavior, Google Trends, reactive searches, proactive searches, leprosy early identification

Procedia PDF Downloads 42
5765 Generation and Diagnostics of Atmospheric Pressure Dielectric Barrier Discharge in Argon/Air

Authors: R. Shrestha, D. P. Subedi, R. B. Tyata, C. S. Wong,

Abstract:

In this paper, a technique for the determination of electron temperatures and electron densities in atmospheric pressure Argon/air discharge by the analysis of optical emission spectra (OES) is reported. The discharge was produced using a high voltage (0-20) kV power supply operating at a frequency of 27 kHz in parallel electrode system, with glass as dielectric. The dielectric layers covering the electrodes act as current limiters and prevent the transition to an arc discharge. Optical emission spectra in the range of (300nm-850nm) were recorded for the discharge with different inter electrode gap keeping electric field constant. Electron temperature (Te) and electron density (ne) are estimated from electrical and optical methods. Electron density was calculated using power balance method. The optical methods are related with line intensity ratio from the relative intensities of Ar-I and Ar-II lines in Argon plasma. The electron density calculated by using line intensity ratio method was compared with the electron density calculated by stark broadening method. The effect of dielectric thickness on plasma parameters (Te and ne) have also been studied and found that Te and ne increases as thickness of dielectric decrease for same inter electrode distance and applied voltage.

Keywords: electron density, electron temperature, optical emission spectra,

Procedia PDF Downloads 480
5764 Algorithm for Recognizing Trees along Power Grid Using Multispectral Imagery

Authors: C. Hamamura, V. Gialluca

Abstract:

Much of the Eclectricity Distributors has about 70% of its electricity interruptions arising from cause "trees", alone or associated with wind and rain and with or without falling branch and / or trees. This contributes inexorably and significantly to outages, resulting in high costs as compensation in addition to the operation and maintenance costs. On the other hand, there is little data structure and solutions to better organize the trees pruning plan effectively, minimizing costs and environmentally friendly. This work describes the development of an algorithm to provide data of trees associated to power grid. The method is accomplished on several steps using satellite imagery and geographically vectorized grid. A sliding window like approach is performed to seek the area around the grid. The proposed method counted 764 trees on a patch of the grid, which was very close to the 738 trees counted manually. The trees data was used as a part of a larger project that implements a system to optimize tree pruning plan.

Keywords: image pattern recognition, trees pruning, trees recognition, neural network

Procedia PDF Downloads 482
5763 Transfer Knowledge From Multiple Source Problems to a Target Problem in Genetic Algorithm

Authors: Terence Soule, Tami Al Ghamdi

Abstract:

To study how to transfer knowledge from multiple source problems to the target problem, we modeled the Transfer Learning (TL) process using Genetic Algorithms as the model solver. TL is the process that aims to transfer learned data from one problem to another problem. The TL process aims to help Machine Learning (ML) algorithms find a solution to the problems. The Genetic Algorithms (GA) give researchers access to information that we have about how the old problem is solved. In this paper, we have five different source problems, and we transfer the knowledge to the target problem. We studied different scenarios of the target problem. The results showed combined knowledge from multiple source problems improves the GA performance. Also, the process of combining knowledge from several problems results in promoting diversity of the transferred population.

Keywords: transfer learning, genetic algorithm, evolutionary computation, source and target

Procedia PDF Downloads 125