Search results for: edge detection algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7074

Search results for: edge detection algorithm

3924 Liquid Chromatography Microfluidics for Detection and Quantification of Urine Albumin Using Linear Regression Method

Authors: Patricia B. Cruz, Catrina Jean G. Valenzuela, Analyn N. Yumang

Abstract:

Nearly a hundred per million of the Filipino population is diagnosed with Chronic Kidney Disease (CKD). The early stage of CKD has no symptoms and can only be discovered once the patient undergoes urinalysis. Over the years, different methods were discovered and used for the quantification of the urinary albumin such as the immunochemical assays where most of these methods require large machinery that has a high cost in maintenance and resources, and a dipstick test which is yet to be proven and is still debated as a reliable method in detecting early stages of microalbuminuria. This research study involves the use of the liquid chromatography concept in microfluidic instruments with biosensor as a means of separation and detection respectively, and linear regression to quantify human urinary albumin. The researchers’ main objective was to create a miniature system that quantifies and detect patients’ urinary albumin while reducing the amount of volume used per five test samples. For this study, 30 urine samples of unknown albumin concentrations were tested using VITROS Analyzer and the microfluidic system for comparison. Based on the data shared by both methods, the actual vs. predicted regression were able to create a positive linear relationship with an R2 of 0.9995 and a linear equation of y = 1.09x + 0.07, indicating that the predicted values and actual values are approximately equal. Furthermore, the microfluidic instrument uses 75% less in total volume – sample and reagents combined, compared to the VITROS Analyzer per five test samples.

Keywords: Chronic Kidney Disease, Linear Regression, Microfluidics, Urinary Albumin

Procedia PDF Downloads 125
3923 The Fabrication of Stress Sensing Based on Artificial Antibodies to Cortisol by Molecular Imprinted Polymer

Authors: Supannika Klangphukhiew, Roongnapa Srichana, Rina Patramanon

Abstract:

Cortisol has been used as a well-known commercial stress biomarker. A homeostasis response to psychological stress is indicated by an increased level of cortisol produced in hypothalamus-pituitary-adrenal (HPA) axis. Chronic psychological stress contributing to the high level of cortisol relates to several health problems. In this study, the cortisol biosensor was fabricated that mimicked the natural receptors. The artificial antibodies were prepared using molecular imprinted polymer technique that can imitate the performance of natural anti-cortisol antibody with high stability. Cortisol-molecular imprinted polymer (cortisol-MIP) was obtained using the multi-step swelling and polymerization protocol with cortisol as a target molecule combining methacrylic acid:acrylamide (2:1) with bisacryloyl-1,2-dihydroxy-1,2-ethylenediamine and ethylenedioxy-N-methylamphetamine as cross-linkers. Cortisol-MIP was integrated to the sensor. It was coated on the disposable screen-printed carbon electrode (SPCE) for portable electrochemical analysis. The physical properties of Cortisol-MIP were characterized by means of electron microscope techniques. The binding characteristics were evaluated via covalent patterns changing in FTIR spectra which were related to voltammetry response. The performance of cortisol-MIP modified SPCE was investigated in terms of detection range, high selectivity with a detection limit of 1.28 ng/ml. The disposable cortisol biosensor represented an application of MIP technique to recognize steroids according to their structures with feasibility and cost-effectiveness that can be developed to use in point-of-care.

Keywords: stress biomarker, cortisol, molecular imprinted polymer, screen-printed carbon electrode

Procedia PDF Downloads 260
3922 Use of a Chagas Urine Nanoparticle Test (Chunap) to Correlate with Parasitemia Levels in T. cruzi/HIV Co-Infected Patients

Authors: Yagahira E. Castro-Sesquen, Robert H. Gilman, Carolina Mejia, Daniel E. Clark, Jeong Choi, Melissa J. Reimer-Mcatee, Rocio Castro, Jorge Flores, Edward Valencia-Ayala, Faustino Torrico, Ricardo Castillo-Neyra, Lance Liotta, Caryn Bern, Alessandra Luchini

Abstract:

Early diagnosis of reactivation of Chagas disease in HIV patients could be lifesaving; however, in Latin American the diagnosis is performed by detection of parasitemia by microscopy which lacks sensitivity. To evaluate if levels of T. cruzi antigens in urine determined by Chunap (Chagas urine nanoparticle test) are correlated with parasitemia levels in T. cruzi/HIV co-infected patients. T. cruzi antigens in urine of HIV patients (N=55: 31 T. cruzi infected and 24 T. cruzi serology negative) were concentrated using hydrogel particles and quantified by Western Blot and a calibration curve. The percentage of Chagas positive patients determined by Chunap compared to blood microscopy, qPCR, and ELISA was 100% (6/6), 95% (18/19) and 74% (23/31), respectively. Chunap specificity was 91.7%. Linear regression analysis demonstrated a direct relationship between parasitemia levels (determined by qPCR) and urine T. cruzi antigen concentrations (p<0.001). A cut-off of > 105 pg was chosen to determine patients with reactivation of Chagas disease (6/6). Urine antigen concentration was significantly higher among patients with CD4+ lymphocyte counts below 200/mL (p=0.045). Chunap shows potential for early detection of reactivation and with appropriate adaptation can be used for monitoring Chagas disease status in T. cruzi/HIV co-infected patients.

Keywords: antigenuria, Chagas disease, Chunap, nanoparticles, parasitemia, poly N-isopropylacrylamide (NIPAm)/trypan blue particles (polyNIPAm/TB), reactivation of Chagas disease.

Procedia PDF Downloads 365
3921 Delineation of Green Infrastructure Buffer Areas with a Simulated Annealing: Consideration of Ecosystem Services Trade-Offs in the Objective Function

Authors: Andres Manuel Garcia Lamparte, Rocio Losada Iglesias, Marcos BoullóN Magan, David Miranda Barros

Abstract:

The biodiversity strategy of the European Union for 2030, mentions climate change as one of the key factors for biodiversity loss and considers green infrastructure as one of the solutions to this problem. In this line, the European Commission has developed a green infrastructure strategy which commits members states to consider green infrastructure in their territorial planning. This green infrastructure is aimed at granting the provision of a wide number of ecosystem services to support biodiversity and human well-being by countering the effects of climate change. Yet, there are not too many tools available to delimit green infrastructure. The available ones consider the potential of the territory to provide ecosystem services. However, these methods usually aggregate several maps of ecosystem services potential without considering possible trade-offs. This can lead to excluding areas with a high potential for providing ecosystem services which have many trade-offs with other ecosystem services. In order to tackle this problem, a methodology is proposed to consider ecosystem services trade-offs in the objective function of a simulated annealing algorithm aimed at delimiting green infrastructure multifunctional buffer areas. To this end, the provision potential maps of the regulating ecosystem services considered to delimit the multifunctional buffer areas are clustered in groups, so that ecosystem services that create trade-offs are excluded in each group. The normalized provision potential maps of the ecosystem services in each group are added to obtain a potential map per group which is normalized again. Then the potential maps for each group are combined in a raster map that shows the highest provision potential value in each cell. The combined map is then used in the objective function of the simulated annealing algorithm. The algorithm is run both using the proposed methodology and considering the ecosystem services individually. The results are analyzed with spatial statistics and landscape metrics to check the number of ecosystem services that the delimited areas produce, as well as their regularity and compactness. It has been observed that the proposed methodology increases the number of ecosystem services produced by delimited areas, improving their multifunctionality and increasing their effectiveness in preventing climate change impacts.

Keywords: ecosystem services trade-offs, green infrastructure delineation, multifunctional buffer areas, climate change

Procedia PDF Downloads 163
3920 Active Flutter Suppression of Sports Aircraft Tailplane by Supplementary Control Surface

Authors: Aleš Kratochvíl, Svatomír Slavík

Abstract:

The paper presents an aircraft flutter suppression by active damping of supplementary control surface at trailing edge. The mathematical model of thin oscillation airfoil with control surface driven by pilot is developed. The supplementary control surface driven by control law is added. Active damping of flutter by several control law is present. The structural model of tailplane with an aerodynamic strip theory based on the airfoil model is developed by a finite element method. The optimization process of stiffens parameters is carried out to match the structural model with results from a ground vibration test of a small sport airplane. The implementation of supplementary control surface driven by control law is present. The active damping of tailplane model is shown.

Keywords: active damping, finite element method, flutter, tailplane model

Procedia PDF Downloads 285
3919 Plasma Actuator Application to Control Surfaces of a Model Aircraft

Authors: Yuta Moriyama, Etsuo Morishita

Abstract:

Plasma actuator is very effective to recover stall flows over an upper airfoil surface. We first manufacture the actuator, test the stability of the device by trial and error basis and find the conditions for steady operations. We visualize the flow around an airfoil in the smoke tunnel and observe the stall recovery. The plasma actuator is stationary device and has no moving parts, and it might be an ideal device to control a model aircraft. We can use the actuator not only as a stall recovery device but also as a spoiler. We put the actuator near the leading edge of an elevator of a model aircraft as a spoiler, and measure the aerodynamic forces by a three-component balance. We observe the effect of the plasma actuator on the aerodynamic forces and the device effectiveness changes depending on the angle of attack whether it is positive or negative. We also visualize the flow caused by the plasma actuator by a desk-top Schlieren photography which is otherwise very difficult in a low-speed wind tunnel experiment.

Keywords: aerodynamics, plasma actuator, model aircraft, wind tunnel

Procedia PDF Downloads 362
3918 Random Forest Classification for Population Segmentation

Authors: Regina Chua

Abstract:

To reduce the costs of re-fielding a large survey, a Random Forest classifier was applied to measure the accuracy of classifying individuals into their assigned segments with the fewest possible questions. Given a long survey, one needed to determine the most predictive ten or fewer questions that would accurately assign new individuals to custom segments. Furthermore, the solution needed to be quick in its classification and usable in non-Python environments. In this paper, a supervised Random Forest classifier was modeled on a dataset with 7,000 individuals, 60 questions, and 254 features. The Random Forest consisted of an iterative collection of individual decision trees that result in a predicted segment with robust precision and recall scores compared to a single tree. A random 70-30 stratified sampling for training the algorithm was used, and accuracy trade-offs at different depths for each segment were identified. Ultimately, the Random Forest classifier performed at 87% accuracy at a depth of 10 with 20 instead of 254 features and 10 instead of 60 questions. With an acceptable accuracy in prioritizing feature selection, new tools were developed for non-Python environments: a worksheet with a formulaic version of the algorithm and an embedded function to predict the segment of an individual in real-time. Random Forest was determined to be an optimal classification model by its feature selection, performance, processing speed, and flexible application in other environments.

Keywords: machine learning, supervised learning, data science, random forest, classification, prediction, predictive modeling

Procedia PDF Downloads 85
3917 Development of a Direct Immunoassay for Human Ferritin Using Diffraction-Based Sensing Method

Authors: Joel Ballesteros, Harriet Jane Caleja, Florian Del Mundo, Cherrie Pascual

Abstract:

Diffraction-based sensing was utilized in the quantification of human ferritin in blood serum to provide an alternative to label-based immunoassays currently used in clinical diagnostics and researches. The diffraction intensity was measured by the diffractive optics technology or dotLab™ system. Two methods were evaluated in this study: direct immunoassay and direct sandwich immunoassay. In the direct immunoassay, human ferritin was captured by human ferritin antibodies immobilized on an avidin-coated sensor while the direct sandwich immunoassay had an additional step for the binding of a detector human ferritin antibody on the analyte complex. Both methods were repeatable with coefficient of variation values below 15%. The direct sandwich immunoassay had a linear response from 10 to 500 ng/mL which is wider than the 100-500 ng/mL of the direct immunoassay. The direct sandwich immunoassay also has a higher calibration sensitivity with value 0.002 Diffractive Intensity (ng mL-1)-1) compared to the 0.004 Diffractive Intensity (ng mL-1)-1 of the direct immunoassay. The limit of detection and limit of quantification values of the direct immunoassay were found to be 29 ng/mL and 98 ng/mL, respectively, while the direct sandwich immunoassay has a limit of detection (LOD) of 2.5 ng/mL and a limit of quantification (LOQ) of 8.2 ng/mL. In terms of accuracy, the direct immunoassay had a percent recovery of 88.8-93.0% in PBS while the direct sandwich immunoassay had 94.1 to 97.2%. Based on the results, the direct sandwich immunoassay is a better diffraction-based immunoassay in terms of accuracy, LOD, LOQ, linear range, and sensitivity. The direct sandwich immunoassay was utilized in the determination of human ferritin in blood serum and the results are validated by Chemiluminescent Magnetic Immunoassay (CMIA). The calculated Pearson correlation coefficient was 0.995 and the p-values of the paired-sample t-test were less than 0.5 which show that the results of the direct sandwich immunoassay was comparable to that of CMIA and could be utilized as an alternative analytical method.

Keywords: biosensor, diffraction, ferritin, immunoassay

Procedia PDF Downloads 344
3916 Geochemical Investigation of Weathering and Sorting for Tepeköy Sandstones

Authors: M. Yavuz Hüseyinca, Şuayip Küpeli

Abstract:

The Chemical Index of Alteration (CIA) values of Late Eocene-Oligocene aged sandstones that exposed on the eastern edge of Tuz Lake (Central Anatolia, Turkey) range from 49 to 59 with an average of 51. The A-CN-K diagram indicates that sandstones underwent post-depositional K-metasomatism. The original average CIA value before the K-metasomatism is calculated as 55. This value is lower than that of Post Archean Australian Shale (PAAS) and defines a low intense chemical weathering in the source-area. Extrapolation of sandstones back to the plagioclase-alkali feldspar line in the A-CN-K diagram suggests a high average plagioclase to alkali feldspar ratio in the provenance and a composition close to granodiorite. The Zr/Sc and Th/Sc ratios with the Al₂O₃-Zr-TiO₂ space do not show zircon addition that refuse both recycling of sediments and sorting effect. All these data suggest direct and rapid transportation from the source due to topographic uplift and probably arid to semi-arid climate conditions for the sandstones.

Keywords: central Anatolia, sandstone, sorting, weathering

Procedia PDF Downloads 367
3915 Modern Information Security Management and Digital Technologies: A Comprehensive Approach to Data Protection

Authors: Mahshid Arabi

Abstract:

With the rapid expansion of digital technologies and the internet, information security has become a critical priority for organizations and individuals. The widespread use of digital tools such as smartphones and internet networks facilitates the storage of vast amounts of data, but simultaneously, vulnerabilities and security threats have significantly increased. The aim of this study is to examine and analyze modern methods of information security management and to develop a comprehensive model to counteract threats and information misuse. This study employs a mixed-methods approach, including both qualitative and quantitative analyses. Initially, a systematic review of previous articles and research in the field of information security was conducted. Then, using the Delphi method, interviews with 30 information security experts were conducted to gather their insights on security challenges and solutions. Based on the results of these interviews, a comprehensive model for information security management was developed. The proposed model includes advanced encryption techniques, machine learning-based intrusion detection systems, and network security protocols. AES and RSA encryption algorithms were used for data protection, and machine learning models such as Random Forest and Neural Networks were utilized for intrusion detection. Statistical analyses were performed using SPSS software. To evaluate the effectiveness of the proposed model, T-Test and ANOVA statistical tests were employed, and results were measured using accuracy, sensitivity, and specificity indicators of the models. Additionally, multiple regression analysis was conducted to examine the impact of various variables on information security. The findings of this study indicate that the comprehensive proposed model reduced cyber-attacks by an average of 85%. Statistical analysis showed that the combined use of encryption techniques and intrusion detection systems significantly improves information security. Based on the obtained results, it is recommended that organizations continuously update their information security systems and use a combination of multiple security methods to protect their data. Additionally, educating employees and raising public awareness about information security can serve as an effective tool in reducing security risks. This research demonstrates that effective and up-to-date information security management requires a comprehensive and coordinated approach, including the development and implementation of advanced techniques and continuous training of human resources.

Keywords: data protection, digital technologies, information security, modern management

Procedia PDF Downloads 18
3914 Design of 3-Step Skew BLAC Motor for Better Performance in Electric Power Steering System

Authors: Subrato Saha, Yun-Hyun Cho

Abstract:

In electric power steering (EPS), spoke type brushless ac (BLAC) motors offer distinct advantages over other electric motor types in terms torque smoothness, reliability and efficiency. This paper deals with the shape optimization of spoke type BLAC motor, in order to reduce cogging torque. This paper examines 3 steps skewing rotor angle, optimizing rotor core edge and rotor overlap length for reducing cogging torque in spoke type BLAC motor. The methods were applied to existing machine designs and their performance was calculated using finite- element analysis (FEA). Prototypes of the machine designs were constructed and experimental results obtained. It is shown that the FEA predicted the cogging torque to be nearly reduce using those methods.

Keywords: EPS, 3-Step skewing, spoke type BLAC, cogging torque, FEA, optimization

Procedia PDF Downloads 482
3913 Contactless Electromagnetic Detection of Stress Fluctuations in Steel Elements

Authors: M. A. García, J. Vinolas, A. Hernando

Abstract:

Steel is nowadays one of the most important structural materials because of its outstanding mechanical properties. Therefore, in order to look for a sustainable economic model and to optimize the use of extensive resources, new methods to monitor and prevent failure of steel-based facilities are required. The classical mechanical tests, as for instance building tasting, are invasive and destructive. Moreover, for facilities where the steel element is embedded, (as reinforced concrete) these techniques are directly non applicable. Hence, non-invasive monitoring techniques to prevent failure, without altering the structural properties of the elements are required. Among them, electromagnetic methods are particularly suitable for non-invasive inspection of the mechanical state of steel-based elements. The magnetoelastic coupling effects induce a modification of the electromagnetic properties of an element upon applied stress. Since most steels are ferromagnetic because of their large Fe content, it is possible to inspect their structure and state in a non-invasive way. We present here a distinct electromagnetic method for contactless evaluation of internal stress in steel-based elements. In particular, this method relies on measuring the magnetic induction between two coils with the steel specimen in between them. We found that the alteration of electromagnetic properties of the steel specimen induced by applied stress-induced changes in the induction allowed us to detect stress well below half of the elastic limit of the material. Hence, it represents an outstanding non-invasive method to prevent failure in steel-based facilities. We here describe the theoretical model, present experimental results to validate it and finally we show a practical application for detection of stress and inhomogeneities in train railways.

Keywords: magnetoelastic, magnetic induction, mechanical stress, steel

Procedia PDF Downloads 29
3912 Open-Loop Vector Control of Induction Motor with Space Vector Pulse Width Modulation Technique

Authors: Karchung, S. Ruangsinchaiwanich

Abstract:

This paper presents open-loop vector control method of induction motor with space vector pulse width modulation (SVPWM) technique. Normally, the closed loop speed control is preferred and is believed to be more accurate. However, it requires a position sensor to track the rotor position which is not desirable to use it for certain workspace applications. This paper exhibits the performance of three-phase induction motor with the simplest control algorithm without the use of a position sensor nor an estimation block to estimate rotor position for sensorless control. The motor stator currents are measured and are transformed to synchronously rotating (d-q-axis) frame by use of Clarke and Park transformation. The actual control happens in this frame where the measured currents are compared with the reference currents. The error signal is fed to a conventional PI controller, and the corrected d-q voltage is generated. The controller outputs are transformed back to three phase voltages and are fed to SVPWM block which generates PWM signal for the voltage source inverter. The open loop vector control model along with SVPWM algorithm is modeled in MATLAB/Simulink software and is experimented and validated in TMS320F28335 DSP board.

Keywords: electric drive, induction motor, open-loop vector control, space vector pulse width modulation technique

Procedia PDF Downloads 141
3911 Flexible Ethylene-Propylene Copolymer Nanofibers Decorated with Ag Nanoparticles as Effective 3D Surface-Enhanced Raman Scattering Substrates

Authors: Yi Li, Rui Lu, Lianjun Wang

Abstract:

With the rapid development of chemical industry, the consumption of volatile organic compounds (VOCs) has increased extensively. In the process of VOCs production and application, plenty of them have been transferred to environment. As a result, it has led to pollution problems not only in soil and ground water but also to human beings. Thus, it is important to develop a sensitive and cost-effective analytical method for trace VOCs detection in environment. Surface-enhanced Raman Spectroscopy (SERS), as one of the most sensitive optical analytical technique with rapid response, pinpoint accuracy and noninvasive detection, has been widely used for ultratrace analysis. Based on the plasmon resonance on the nanoscale metallic surface, SERS technology can even detect single molecule due to abundant nanogaps (i.e. 'hot spots') on the nanosubstrate. In this work, a self-supported flexible silver nitrate (AgNO3)/ethylene-propylene copolymer (EPM) hybrid nanofibers was fabricated by electrospinning. After an in-situ chemical reduction using ice-cold sodium borohydride as reduction agent, numerous silver nanoparticles were formed on the nanofiber surface. By adjusting the reduction time and AgNO3 content, the morphology and dimension of silver nanoparticles could be controlled. According to the principles of solid-phase extraction, the hydrophobic substance is more likely to partition into the hydrophobic EPM membrane in an aqueous environment while water and other polar components are excluded from the analytes. By the enrichment of EPM fibers, the number of hydrophobic molecules located on the 'hot spots' generated from criss-crossed nanofibers is greatly increased, which further enhances SERS signal intensity. The as-prepared Ag/EPM hybrid nanofibers were first employed to detect common SERS probe molecule (p-aminothiophenol) with the detection limit down to 10-12 M, which demonstrated an excellent SERS performance. To further study the application of the fabricated substrate for monitoring hydrophobic substance in water, several typical VOCs, such as benzene, toluene and p-xylene, were selected as model compounds. The results showed that the characteristic peaks of these target analytes in the mixed aqueous solution could be distinguished even at a concentration of 10-6 M after multi-peaks gaussian fitting process, including C-H bending (850 cm-1), C-C ring stretching (1581 cm-1, 1600 cm-1) of benzene, C-H bending (844 cm-1 ,1151 cm-1), C-C ring stretching (1001 cm-1), CH3 bending vibration (1377 cm-1) of toluene, C-H bending (829 cm-1), C-C stretching (1614 cm-1) of p-xylene. The SERS substrate has remarkable advantages which combine the enrichment capacity from EPM and the Raman enhancement of Ag nanoparticles. Meanwhile, the huge specific surface area resulted from electrospinning is benificial to increase the number of adsoption sites and promotes 'hot spots' formation. In summary, this work provides powerful potential in rapid, on-site and accurate detection of trace VOCs using a portable Raman.

Keywords: electrospinning, ethylene-propylene copolymer, silver nanoparticles, SERS, VOCs

Procedia PDF Downloads 155
3910 Hybrid Hunger Games Search Optimization Based on the Neural Networks Approach Applied to UAVs

Authors: Nadia Samantha Zuñiga-Peña, Norberto Hernández-Romero, Omar Aguilar-Mejia, Salatiel García-Nava

Abstract:

Using unmanned aerial vehicles (UAVs) for load transport has gained significant importance in various sectors due to their ability to improve efficiency, reduce costs, and access hard-to-reach areas. Although UAVs offer numerous advantages for load transport, several complications and challenges must be addressed to exploit their potential fully. Complexity relays on UAVs are underactuated, non-linear systems with a high degree of coupling between their variables and are subject to forces with uncertainty. One of the biggest challenges is modeling and controlling the system formed by UAVs carrying a load. In order to solve the controller problem, in this work, a hybridization of Neural Network and Hunger Games Search (HGS) metaheuristic algorithm is developed and implemented to find the parameters of the Super Twisting Sliding Mode Controller for the 8 degrees of freedom model of UAV with payload. The optimized controller successfully tracks the UAV through the three-dimensional desired path, demonstrating the effectiveness of the proposed solution. A comparison of performance shows the superiority of the neural network HGS (NNHGS) over the HGS algorithm, minimizing the tracking error by 57.5 %.

Keywords: neural networks, hunger games search, super twisting sliding mode controller, UAVs.

Procedia PDF Downloads 15
3909 Modeling of Crack Propagation Path in Concrete with Coarse Trapezoidal Aggregates by Boundary Element Method

Authors: Chong Wang, Alexandre Urbano Hoffmann

Abstract:

Interaction between a crack and a trapezoidal aggregate in a single edge notched concrete beam is simulated using boundary element method with an automatic crack extension program. The stress intensity factors of the growing crack are obtained from the J-integral. Three crack extension paths: deflecting around the particulate, growing along the interface and penetrating into the particulate are achieved in terms of the mismatch state of mechanical characteristics of matrix and the particulate. The toughening is also given by the ratio of stress intensity factors. The results reveal that as stress shielding occurs, toughening is obtained when the crack is approaching to a stiff and strong aggregate weakly bonded to a relatively soft matrix. The present work intends to help for the design of aggregate reinforced concretes.

Keywords: aggregate concrete, boundary element method, two-phase composite, crack extension path, crack/particulate interaction

Procedia PDF Downloads 420
3908 Obtaining High-Dimensional Configuration Space for Robotic Systems Operating in a Common Environment

Authors: U. Yerlikaya, R. T. Balkan

Abstract:

In this research, a method is developed to obtain high-dimensional configuration space for path planning problems. In typical cases, the path planning problems are solved directly in the 3-dimensional (D) workspace. However, this method is inefficient in handling the robots with various geometrical and mechanical restrictions. To overcome these difficulties, path planning may be formalized and solved in a new space which is called configuration space. The number of dimensions of the configuration space comes from the degree of freedoms of the system of interest. The method can be applied in two ways. In the first way, the point clouds of all the bodies of the system and interaction of them are used. The second way is performed via using the clearance function of simulation software where the minimum distances between surfaces of bodies are simultaneously measured. A double-turret system is held in the scope of this study. The 4-D configuration space of a double-turret system is obtained in these two ways. As a result, the difference between these two methods is around 1%, depending on the density of the point cloud. The disparity between the two forms steadily decreases as the point cloud density increases. At the end of the study, in order to verify 4-D configuration space obtained, 4-D path planning problem was realized as 2-D + 2-D and a sample path planning is carried out with using A* algorithm. Then, the accuracy of the configuration space is proved using the obtained paths on the simulation model of the double-turret system.

Keywords: A* algorithm, autonomous turrets, high-dimensional C-space, manifold C-space, point clouds

Procedia PDF Downloads 133
3907 A Variable Neighborhood Search with Tabu Conditions for the Roaming Salesman Problem

Authors: Masoud Shahmanzari

Abstract:

The aim of this paper is to present a Variable Neighborhood Search (VNS) with Tabu Search (TS) conditions for the Roaming Salesman Problem (RSP). The RSP is a special case of the well-known traveling salesman problem (TSP) where a set of cities with time-dependent rewards and a set of campaign days are given. Each city can be visited on any day and a subset of cities can be visited multiple times. The goal is to determine an optimal campaign schedule consist of daily open/closed tours that visit some cities and maximizes the total net benefit while respecting daily maximum tour duration constraints and the necessity to return campaign base frequently. This problem arises in several real-life applications and particularly in election logistics where depots are not fixed. We formulate the problem as a mixed integer linear programming (MILP), in which we capture as many real-world aspects of the RSP as possible. We also present a hybrid metaheuristic algorithm based on a VNS with TS conditions. The initial feasible solution is constructed via a new matheuristc approach based on the decomposition of the original problem. Next, this solution is improved in terms of the collected rewards using the proposed local search procedure. We consider a set of 81 cities in Turkey and a campaign of 30 days as our largest instance. Computational results on real-world instances show that the developed algorithm could find near-optimal solutions effectively.

Keywords: optimization, routing, election logistics, heuristics

Procedia PDF Downloads 82
3906 Quality of Service of Transportation Networks: A Hybrid Measurement of Travel Time and Reliability

Authors: Chin-Chia Jane

Abstract:

In a transportation network, travel time refers to the transmission time from source node to destination node, whereas reliability refers to the probability of a successful connection from source node to destination node. With an increasing emphasis on quality of service (QoS), both performance indexes are significant in the design and analysis of transportation systems. In this work, we extend the well-known flow network model for transportation networks so that travel time and reliability are integrated into the QoS measurement simultaneously. In the extended model, in addition to the general arc capacities, each intermediate node has a time weight which is the travel time for per unit of commodity going through the node. Meanwhile, arcs and nodes are treated as binary random variables that switch between operation and failure with associated probabilities. For pre-specified travel time limitation and demand requirement, the QoS of a transportation network is the probability that source can successfully transport the demand requirement to destination while the total transmission time is under the travel time limitation. This work is pioneering, since existing literatures that evaluate travel time reliability via a single optimization path, the proposed QoS focuses the performance of the whole network system. To compute the QoS of transportation networks, we first transfer the extended network model into an equivalent min-cost max-flow network model. In the transferred network, each arc has a new travel time weight which takes value 0. Each intermediate node is replaced by two nodes u and v, and an arc directed from u to v. The newly generated nodes u and v are perfect nodes. The new direct arc has three weights: travel time, capacity, and operation probability. Then the universal set of state vectors is recursively decomposed into disjoint subsets of reliable, unreliable, and stochastic vectors until no stochastic vector is left. The decomposition is made possible by applying existing efficient min-cost max-flow algorithm. Because the reliable subsets are disjoint, QoS can be obtained directly by summing the probabilities of these reliable subsets. Computational experiments are conducted on a benchmark network which has 11 nodes and 21 arcs. Five travel time limitations and five demand requirements are set to compute the QoS value. To make a comparison, we test the exhaustive complete enumeration method. Computational results reveal the proposed algorithm is much more efficient than the complete enumeration method. In this work, a transportation network is analyzed by an extended flow network model where each arc has a fixed capacity, each intermediate node has a time weight, and both arcs and nodes are independent binary random variables. The quality of service of the transportation network is an integration of customer demands, travel time, and the probability of connection. We present a decomposition algorithm to compute the QoS efficiently. Computational experiments conducted on a prototype network show that the proposed algorithm is superior to existing complete enumeration methods.

Keywords: quality of service, reliability, transportation network, travel time

Procedia PDF Downloads 214
3905 Data Compression in Ultrasonic Network Communication via Sparse Signal Processing

Authors: Beata Zima, Octavio A. Márquez Reyes, Masoud Mohammadgholiha, Jochen Moll, Luca de Marchi

Abstract:

This document presents the approach of using compressed sensing in signal encoding and information transferring within a guided wave sensor network, comprised of specially designed frequency steerable acoustic transducers (FSATs). Wave propagation in a damaged plate was simulated using commercial FEM-based software COMSOL. Guided waves were excited by means of FSATs, characterized by the special shape of its electrodes, and modeled using PIC255 piezoelectric material. The special shape of the FSAT, allows for focusing wave energy in a certain direction, accordingly to the frequency components of its actuation signal, which makes available a larger monitored area. The process begins when a FSAT detects and records reflection from damage in the structure, this signal is then encoded and prepared for transmission, using a combined approach, based on Compressed Sensing Matching Pursuit and Quadrature Amplitude Modulation (QAM). After codification of the signal is in binary chars the information is transmitted between the nodes in the network. The message reaches the last node, where it is finally decoded and processed, to be used for damage detection and localization purposes. The main aim of the investigation is to determine the location of detected damage using reconstructed signals. The study demonstrates that the special steerable capabilities of FSATs, not only facilitate the detection of damage but also permit transmitting the damage information to a chosen area in a specific direction of the investigated structure.

Keywords: data compression, ultrasonic communication, guided waves, FEM analysis

Procedia PDF Downloads 117
3904 Wind Fragility for Honeycomb Roof Cladding Panels Using Screw Pull-Out Capacity

Authors: Viriyavudh Sim, Woo Young Jung

Abstract:

The failure of roof cladding mostly occurs due to the failing of the connection between claddings and purlins, which is the pull-out of the screw connecting the two parts when the pull-out load, i.e. typhoon, is higher than the resistance of the connection screw. As typhoon disasters in Korea are constantly on the rise, probability risk assessment (PRA) has become a vital tool to evaluate the performance of civil structures. In this study, we attempted to determine the fragility of roof cladding with the screw connection. Experimental study was performed to evaluate the pull-out resistance of screw joints between honeycomb panels and back frames. Subsequently, by means of Monte Carlo Simulation method, probability of failure for these types of roof cladding was determined. The results that the failure of roof cladding was depends on their location on the roof, for example, the edge most panel has the highest probability of failure.

Keywords: Monte Carlo Simulation, roof cladding, screw pull-out strength, wind fragility

Procedia PDF Downloads 245
3903 Supercomputer Simulation of Magnetic Multilayers Films

Authors: Vitalii Yu. Kapitan, Aleksandr V. Perzhu, Konstantin V. Nefedev

Abstract:

The necessity of studying magnetic multilayer structures is explained by the prospects of their practical application as a technological base for creating new storages medium. Magnetic multilayer films have many unique features that contribute to increasing the density of information recording and the speed of storage devices. Multilayer structures are structures of alternating magnetic and nonmagnetic layers. In frame of the classical Heisenberg model, lattice spin systems with direct short- and long-range exchange interactions were investigated by Monte Carlo methods. The thermodynamic characteristics of multilayer structures, such as the temperature behavior of magnetization, energy, and heat capacity, were investigated. The processes of magnetization reversal of multilayer structures in external magnetic fields were investigated. The developed software is based on the new, promising programming language Rust. Rust is a new experimental programming language developed by Mozilla. The language is positioned as an alternative to C and C++. For the Monte Carlo simulation, the Metropolis algorithm and its parallel implementation using MPI and the Wang-Landau algorithm were used. We are planning to study of magnetic multilayer films with asymmetric Dzyaloshinskii–Moriya (DM) interaction, interfacing effects and skyrmions textures. This work was supported by the state task of the Ministry of Education and Science of the Russia # 3.7383.2017/8.9

Keywords: The Monte Carlo methods, Heisenberg model, multilayer structures, magnetic skyrmion

Procedia PDF Downloads 161
3902 Array Type Miniaturized Ultrasonic Sensors for Detecting Sinkhole in the City

Authors: Won Young Choi, Kwan Kyu Park

Abstract:

Recently, the road depression happening in the urban area is different from the cause of the sink hole and the generation mechanism occurring in the limestone area. The main cause of sinkholes occurring in the city center is the loss of soil due to the damage of old underground buried materials and groundwater discharge due to large underground excavation works. The method of detecting the sinkhole in the urban area is mostly using the Ground Penetration Radar (GPR). However, it is challenging to implement compact system and detecting watery state since it is based on electromagnetic waves. Although many ultrasonic underground detection studies have been conducted, near-ground detection (several tens of cm to several meters) has been developed for bulk systems using geophones as a receiver. The goal of this work is to fabricate a miniaturized sinkhole detecting system based on low-cost ultrasonic transducers of 40 kHz resonant frequency with high transmission pressure and receiving sensitivity. Motived by biomedical ultrasonic imaging methods, we detect air layers below the ground such as asphalt through the pulse-echo method. To improve image quality using multi-channel, linear array system is implemented, and image is acquired by classical synthetic aperture imaging method. We present the successful feasibility test of multi-channel sinkhole detector based on ultrasonic transducer. In this work, we presented and analyzed image results which are imaged by single channel pulse-echo imaging, synthetic aperture imaging.

Keywords: road depression, sinkhole, synthetic aperture imaging, ultrasonic transducer

Procedia PDF Downloads 135
3901 Nine-Level Shunt Active Power Filter Associated with a Photovoltaic Array Coupled to the Electrical Distribution Network

Authors: Zahzouh Zoubir, Bouzaouit Azzeddine, Gahgah Mounir

Abstract:

The use of more and more electronic power switches with a nonlinear behavior generates non-sinusoidal currents in distribution networks, which causes damage to domestic and industrial equipment. The multi-level shunt power active filter is subsequently shown to be an adequate solution to the problem raised. Nevertheless, the difficulty of adjusting the active filter DC supply voltage requires another technology to ensure it. In this article, a photovoltaic generator is associated with the DC bus power terminals of the active filter. The proposed system consists of a field of solar panels, three multi-level voltage inverters connected to the power grid and a non-linear load consisting of a six-diode rectifier bridge supplying a resistive-inductive load. Current control techniques of active and reactive power are used to compensate for both harmonic currents and reactive power as well as to inject active solar power into the distribution network. An algorithm of the search method of the maximum power point of type Perturb and observe is applied. Simulation results of the system proposed under the MATLAB/Simulink environment shows that the performance of control commands that reassure the solar power injection in the network, harmonic current compensation and power factor correction.

Keywords: Actif power filter, MPPT, pertub&observe algorithm, PV array, PWM-control

Procedia PDF Downloads 328
3900 Intrusion Detection in Cloud Computing Using Machine Learning

Authors: Faiza Babur Khan, Sohail Asghar

Abstract:

With an emergence of distributed environment, cloud computing is proving to be the most stimulating computing paradigm shift in computer technology, resulting in spectacular expansion in IT industry. Many companies have augmented their technical infrastructure by adopting cloud resource sharing architecture. Cloud computing has opened doors to unlimited opportunities from application to platform availability, expandable storage and provision of computing environment. However, from a security viewpoint, an added risk level is introduced from clouds, weakening the protection mechanisms, and hardening the availability of privacy, data security and on demand service. Issues of trust, confidentiality, and integrity are elevated due to multitenant resource sharing architecture of cloud. Trust or reliability of cloud refers to its capability of providing the needed services precisely and unfailingly. Confidentiality is the ability of the architecture to ensure authorization of the relevant party to access its private data. It also guarantees integrity to protect the data from being fabricated by an unauthorized user. So in order to assure provision of secured cloud, a roadmap or model is obligatory to analyze a security problem, design mitigation strategies, and evaluate solutions. The aim of the paper is twofold; first to enlighten the factors which make cloud security critical along with alleviation strategies and secondly to propose an intrusion detection model that identifies the attackers in a preventive way using machine learning Random Forest classifier with an accuracy of 99.8%. This model uses less number of features. A comparison with other classifiers is also presented.

Keywords: cloud security, threats, machine learning, random forest, classification

Procedia PDF Downloads 313
3899 Photoelectrochemical Study of Nanostructured Acropora-Like Lead Sulfide Thin Films

Authors: S. Kaci, A. Keffous, O. Fellahi, I. Bozetine, H. Menari

Abstract:

In this paper, we report the fabrication and characterization of Acropora-like lead sulfide nanostructured thin films using chemical bath deposition. The method has the strong points of low temperature and no surfactant, comparing with the other method. The preferential growth directions of the broad branches were indexed as along (200) directions. The photoelectrochemical property of the as-deposited thin films was also investigated. Photoelectrochemical characterization was performed in the aim to determine the flat band potential (Vfb) and to confirm the n-type character of PbS, elucidated from the J(V) curves both in the dark and under illumination. The apparition of the photocurrent Jph started at a potential VON of −0.41 V/ECS and increased towards the anodic direction, which is typical of n-type behavior. The near infrared absorbance spectrum displayed an absorbance edge at 1959 nm, showing blue shift comparing to bulk PbS (3020 nm). These nanostructured lead sulfide thin films may have potential application as dispersed photoelectrode capable of generating H2 under visible light.

Keywords: lead sulfide, nanostructures, photo-conversion, thin films

Procedia PDF Downloads 350
3898 Hard Disk Failure Predictions in Supercomputing System Based on CNN-LSTM and Oversampling Technique

Authors: Yingkun Huang, Li Guo, Zekang Lan, Kai Tian

Abstract:

Hard disk drives (HDD) failure of the exascale supercomputing system may lead to service interruption and invalidate previous calculations, and it will cause permanent data loss. Therefore, initiating corrective actions before hard drive failures materialize is critical to the continued operation of jobs. In this paper, a highly accurate analysis model based on CNN-LSTM and oversampling technique was proposed, which can correctly predict the necessity of a disk replacement even ten days in advance. Generally, the learning-based method performs poorly on a training dataset with long-tail distribution, especially fault prediction is a very classic situation as the scarcity of failure data. To overcome the puzzle, a new oversampling was employed to augment the data, and then, an improved CNN-LSTM with the shortcut was built to learn more effective features. The shortcut transmits the results of the previous layer of CNN and is used as the input of the LSTM model after weighted fusion with the output of the next layer. Finally, a detailed, empirical comparison of 6 prediction methods is presented and discussed on a public dataset for evaluation. The experiments indicate that the proposed method predicts disk failure with 0.91 Precision, 0.91 Recall, 0.91 F-measure, and 0.90 MCC for 10 days prediction horizon. Thus, the proposed algorithm is an efficient algorithm for predicting HDD failure in supercomputing.

Keywords: HDD replacement, failure, CNN-LSTM, oversampling, prediction

Procedia PDF Downloads 69
3897 Regularizing Software for Aerosol Particles

Authors: Christine Böckmann, Julia Rosemann

Abstract:

We present an inversion algorithm that is used in the European Aerosol Lidar Network for the inversion of data collected with multi-wavelength Raman lidar. These instruments measure backscatter coefficients at 355, 532, and 1064 nm, and extinction coefficients at 355 and 532 nm. The algorithm is based on manually controlled inversion of optical data which allows for detailed sensitivity studies and thus provides us with comparably high quality of the derived data products. The algorithm allows us to derive particle effective radius, volume, surface-area concentration with comparably high confidence. The retrieval of the real and imaginary parts of the complex refractive index still is a challenge in view of the accuracy required for these parameters in climate change studies in which light-absorption needs to be known with high accuracy. Single-scattering albedo (SSA) can be computed from the retrieve microphysical parameters and allows us to categorize aerosols into high and low absorbing aerosols. From mathematical point of view the algorithm is based on the concept of using truncated singular value decomposition as regularization method. This method was adapted to work for the retrieval of the particle size distribution function (PSD) and is called hybrid regularization technique since it is using a triple of regularization parameters. The inversion of an ill-posed problem, such as the retrieval of the PSD, is always a challenging task because very small measurement errors will be amplified most often hugely during the solution process unless an appropriate regularization method is used. Even using a regularization method is difficult since appropriate regularization parameters have to be determined. Therefore, in a next stage of our work we decided to use two regularization techniques in parallel for comparison purpose. The second method is an iterative regularization method based on Pade iteration. Here, the number of iteration steps serves as the regularization parameter. We successfully developed a semi-automated software for spherical particles which is able to run even on a parallel processor machine. From a mathematical point of view, it is also very important (as selection criteria for an appropriate regularization method) to investigate the degree of ill-posedness of the problem which we found is a moderate ill-posedness. We computed the optical data from mono-modal logarithmic PSD and investigated particles of spherical shape in our simulations. We considered particle radii as large as 6 nm which does not only cover the size range of particles in the fine-mode fraction of naturally occurring PSD but also covers a part of the coarse-mode fraction of PSD. We considered errors of 15% in the simulation studies. For the SSA, 100% of all cases achieve relative errors below 12%. In more detail, 87% of all cases for 355 nm and 88% of all cases for 532 nm are well below 6%. With respect to the absolute error for non- and weak-absorbing particles with real parts 1.5 and 1.6 in all modes the accuracy limit +/- 0.03 is achieved. In sum, 70% of all cases stay below +/-0.03 which is sufficient for climate change studies.

Keywords: aerosol particles, inverse problem, microphysical particle properties, regularization

Procedia PDF Downloads 334
3896 Method Validation for Heavy Metal Determination in Spring Water and Sediments

Authors: Habtamu Abdisa

Abstract:

Spring water is particularly valuable due to its high mineral content, which is beneficial for human health. However, anthropogenic activities usually imbalance the natural levels of its composition, which can cause adverse health effects. Regular monitoring of a naturally given environmental resource is of great concern in the world today. The spectrophotometric application is one of the best methods for qualifying and quantifying the mineral contents of environmental water samples. This research was conducted to evaluate the quality of spring water concerning its heavy metal composition. A grab sampling technique was employed to collect representative samples, including duplicates. The samples were then treated with concentrated HNO3 to a pH level below 2 and stored at 4oC. The samples were digested and analyzed for cadmium (Cd), chromium (Cr), manganese (Mn), copper (Cu), iron (Fe), and zinc (Zn) following method validation. Atomic Absorption Spectrometry (AAS) was utilized for the sample analysis. Quality control measures, including blanks, duplicates, and certified reference materials (CRMs), were implemented to ensure the accuracy and precision of the analytical results. Of the metals analyzed in the water samples, Cd and Cr were found to be below the detection limit. However, the concentrations of Mn, Cu, Fe, and Zn ranged from mean values of 0.119-0.227 mg/L, 0.142-0.166 mg/L, 0.183-0.267 mg/L, and 0.074-0.181 mg/L, respectively. Sediment analysis revealed mean concentration ranges of 348.31-429.21 mg/kg, 0.23-0.28 mg/kg, 18.73-22.84 mg/kg, 2.76-3.15 mg/kg, 941.84-1128.56 mg/kg, and 42.39-66.53 mg/kg for Mn, Cd, Cu, Cr, Fe, and Zn, respectively. The study results established that the evaluated spring water and its associated sediment met the regulatory standards and guidelines for heavy metal concentrations. Furthermore, this research can enhance the quality assurance and control processes for environmental sample analysis, ensuring the generation of reliable data.

Keywords: method validation, heavy metal, spring water, sediment, method detection limit

Procedia PDF Downloads 60
3895 Increasing the Resilience of Cyber Physical Systems in Smart Grid Environments using Dynamic Cells

Authors: Andrea Tundis, Carlos García Cordero, Rolf Egert, Alfredo Garro, Max Mühlhäuser

Abstract:

Resilience is an important system property that relies on the ability of a system to automatically recover from a degraded state so as to continue providing its services. Resilient systems have the means of detecting faults and failures with the added capability of automatically restoring their normal operations. Mastering resilience in the domain of Cyber-Physical Systems is challenging due to the interdependence of hybrid hardware and software components, along with physical limitations, laws, regulations and standards, among others. In order to overcome these challenges, this paper presents a modeling approach, based on the concept of Dynamic Cells, tailored to the management of Smart Grids. Additionally, a heuristic algorithm that works on top of the proposed modeling approach, to find resilient configurations, has been defined and implemented. More specifically, the model supports a flexible representation of Smart Grids and the algorithm is able to manage, at different abstraction levels, the resource consumption of individual grid elements on the presence of failures and faults. Finally, the proposal is evaluated in a test scenario where the effectiveness of such approach, when dealing with complex scenarios where adequate solutions are difficult to find, is shown.

Keywords: cyber-physical systems, energy management, optimization, smart grids, self-healing, resilience, security

Procedia PDF Downloads 321