Search results for: mathematical algorithms of targeting and persecution
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4251

Search results for: mathematical algorithms of targeting and persecution

2331 The Delaying Influence of Degradation on the Divestment of Gas Turbines for Associated Gas Utilisation: Part 1

Authors: Mafel Obhuo, Dodeye I. Igbong, Duabari S. Aziaka, Pericles Pilidis

Abstract:

An important feature of the exploitation of associated gas as fuel for gas turbine engines is a declining supply. So when exploiting this resource, the divestment of prime movers is very important as the fuel supply diminishes with time. This paper explores the influence of engine degradation on the timing of divestments. Hypothetical but realistic gas turbine engines were modelled with Turbomatch, the Cranfield University gas turbine performance simulation tool. The results were deployed in three degradation scenarios within the TERA (Techno-economic and environmental risk analysis) framework to develop economic models. An optimisation with Genetic Algorithms was carried out to maximize the economic benefit. The results show that degradation will have a significant impact. It will delay the divestment of power plants, while they are running less efficiently. Over a 20 year investment, a decrease of $0.11bn, $0.26bn and $0.45bn (billion US dollars) were observed for the three degradation scenarios as against the clean case.

Keywords: economic return, flared associated gas, net present value, optimization

Procedia PDF Downloads 119
2330 A Parallel Algorithm for Solving the PFSP on the Grid

Authors: Samia Kouki

Abstract:

Solving NP-hard combinatorial optimization problems by exact search methods, such as Branch-and-Bound, may degenerate to complete enumeration. For that reason, exact approaches limit us to solve only small or moderate size problem instances, due to the exponential increase in CPU time when problem size increases. One of the most promising ways to reduce significantly the computational burden of sequential versions of Branch-and-Bound is to design parallel versions of these algorithms which employ several processors. This paper describes a parallel Branch-and-Bound algorithm called GALB for solving the classical permutation flowshop scheduling problem as well as its implementation on a Grid computing infrastructure. The experimental study of our distributed parallel algorithm gives promising results and shows clearly the benefit of the parallel paradigm to solve large-scale instances in moderate CPU time.

Keywords: grid computing, permutation flow shop problem, branch and bound, load balancing

Procedia PDF Downloads 269
2329 Formulation of the N-Acylethanolamine, Linoleoylethanolamide into Cubosomes for Delivery across the Blood-Brain Barrier

Authors: Younus Mohammad, Anita B. Fallah, Ben J. Boyd, Shakila B. Rizwan

Abstract:

N-acylethanolamines (NAEs) are endogenous lipids, which have neuromodulatory properties. NAEs have shown neuroprotective properties in various neurodegenerative diseases including Alzheimer's disease, Parkinson's disease and ischemic stroke. However, NAEs are eliminated rapidly in vivo by enzymatic hydrolysis. We propose to encapsulate NAEs in liquid crystalline nanoparticles (cubosomes) to increase their biological half-life and explore their therapeutic potential. Recently, we have reported the co-formulation and nanostructural characterization of cubosomes containing the NAE, oleoylethanolamide and a synthetic cubosome forming lipid phytantriol. Here, we report on the formulation of cubosomes with the NAE, linoleoylethanolamide (LEA) as the core cubosome forming lipid. LEA-cubosomes were formulated in the presence of three different steric stabilisers: two brain targeting ligands, Tween 80 and Pluronic P188 and a control, Pluronic F127. Size, morphology and internal structure of formulations were characterized by dynamic light scattering (DLS), cryogenic transmission electron microscopy (Cryo–TEM) and small angle X–ray scattering (SAXS), respectively. Chemical stability of LEA in formulations was investigated using high-performance liquid chromatography (HPLC). Cytotoxicity of formulations towards human cerebral microvascular endothelial cell line (hCMEC/D3) was also investigated using an MTT (3-[4, 5- dimethylthiazol-2-yl]-2, 5-diphenyl tetrazolium bromide) assay. All cubosome formulations had mean particle size of less than 250 nm and were uniformly distributed with polydispersity indices less than 0.2. Cubosomes produced had a bicontinuous cubic internal structure with an Im3m space group but different lattice parameters, indicating the different modes of interaction between the stabilisers and LEA. LEA in formulations was found to be chemically stable. At concentrations of up to 20 µg/mL LEA in the presence of all the stabilisers, greater than 80% cell viability was observed.

Keywords: blood-brain barrier, cubosomes, linoleoyl ethanolamide, N-acylethanolamines (NAEs)

Procedia PDF Downloads 193
2328 Quantification of Global Cerebrovascular Reactivity in the Principal Feeding Arteries of the Human Brain

Authors: Ravinder Kaur

Abstract:

Introduction Global cerebrovascular reactivity (CVR) mapping is a promising clinical assessment for stress-testing the brain using physiological challenges, such as CO₂, to elicit changes in perfusion. It enables real-time assessment of cerebrovascular integrity and health. Conventional imaging approaches solely use steady-state parameters, like cerebral blood flow (CBF), to evaluate the integrity of the resting parenchyma and can erroneously show a healthy brain at rest, despite the underlying pathogenesis in the presence of cerebrovascular disease. Conversely, coupling CO₂ inhalation with phase-contrast MRI neuroimaging interrogates the capacity of the vasculature to respond to changes under stress. It shows promise in providing prognostic value as a novel health marker to measure neurovascular function in disease and to detect early brain vasculature dysfunction. Objective This exploratory study was established to:(a) quantify the CBF response to CO₂ in hypocapnia and hypercapnia,(b) evaluate disparities in CVR between internal carotid (ICA) and vertebral artery (VA), and (c) assess sex-specific variation in CVR. Methodology Phase-contrast MRI was employed to measure the cerebrovascular reactivity to CO₂ (±10 mmHg). The respiratory interventions were presented using the prospectively end-tidal targeting RespirActTM Gen3 system. Post-processing and statistical analysis were conducted. Results In 9 young, healthy subjects, the CBF increased from hypocapnia to hypercapnia in all vessels (4.21±0.76 to 7.20±1.83 mL/sec in ICA, 1.36±0.55 to 2.33±1.31 mL/sec in VA, p < 0.05). The CVR was quantitatively higher in ICA than VA (slope of linear regression: 0.23 vs. 0.07 mL/sec/mmHg, p < 0.05). No statistically significant effect was observed in CVR between male and female (0.25 vs 0.20 mL/sec/mmHg in ICA, 0.09 vs 0.11 mL/sec/mmHg in VA, p > 0.05). Conclusions The principal finding in this investigation validated the modulation of CBF by CO₂. Moreover, it has indicated that regional heterogeneity in hemodynamic response exists in the brain. This study provides scope to standardize the quantification of CVR prior to its clinical translation.

Keywords: cerebrovascular disease, neuroimaging, phase contrast MRI, cerebrovascular reactivity, carbon dioxide

Procedia PDF Downloads 131
2327 Hardware Implementation for the Contact Force Reconstruction in Tactile Sensor Arrays

Authors: María-Luisa Pinto-Salamanca, Wilson-Javier Pérez-Holguín

Abstract:

Reconstruction of contact forces is a fundamental technique for analyzing the properties of a touched object and is essential for regulating the grip force in slip control loops. This is based on the processing of the distribution, intensity, and direction of the forces during the capture of the sensors. Currently, efficient hardware alternatives have been used more frequently in different fields of application, allowing the implementation of computationally complex algorithms, as is the case with tactile signal processing. The use of hardware for smart tactile sensing systems is a research area that promises to improve the processing time and portability requirements of applications such as artificial skin and robotics, among others. The literature review shows that hardware implementations are present today in almost all stages of smart tactile detection systems except in the force reconstruction process, a stage in which they have been less applied. This work presents a hardware implementation of a model-driven reported in the literature for the contact force reconstruction of flat and rigid tactile sensor arrays from normal stress data. From the analysis of a software implementation of such a model, this implementation proposes the parallelization of tasks that facilitate the execution of matrix operations and a two-dimensional optimization function to obtain a vector force by each taxel in the array. This work seeks to take advantage of the parallel hardware characteristics of Field Programmable Gate Arrays, FPGAs, and the possibility of applying appropriate techniques for algorithms parallelization using as a guide the rules of generalization, efficiency, and scalability in the tactile decoding process and considering the low latency, low power consumption, and real-time execution as the main parameters of design. The results show a maximum estimation error of 32% in the tangential forces and 22% in the normal forces with respect to the simulation by the Finite Element Modeling (FEM) technique of Hertzian and non-Hertzian contact events, over sensor arrays of 10×10 taxels of different sizes. The hardware implementation was carried out on an MPSoC XCZU9EG-2FFVB1156 platform of Xilinx® that allows the reconstruction of force vectors following a scalable approach, from the information captured by means of tactile sensor arrays composed of up to 48 × 48 taxels that use various transduction technologies. The proposed implementation demonstrates a reduction in estimation time of x / 180 compared to software implementations. Despite the relatively high values of the estimation errors, the information provided by this implementation on the tangential and normal tractions and the triaxial reconstruction of forces allows to adequately reconstruct the tactile properties of the touched object, which are similar to those obtained in the software implementation and in the two FEM simulations taken as reference. Although errors could be reduced, the proposed implementation is useful for decoding contact forces for portable tactile sensing systems, thus helping to expand electronic skin applications in robotic and biomedical contexts.

Keywords: contact forces reconstruction, forces estimation, tactile sensor array, hardware implementation

Procedia PDF Downloads 181
2326 Mechanisms of Action in Mindfulness-Based Cognitive Therapy (MBCT) and Mindfulness-Based Stress Reduction (MBSR) in People with Physical and/or Psychological Conditions: A Systematic Review

Authors: Modi Alsubaie, Willem Kuyken, Rebecca Abbott, Barnaby Dunn, Chris Dickens, Tina Keil, William Henley

Abstract:

Background: Recently, there has been an increased interest in studying the effects of mindfulness-based interventions for people with psychological and physical problems. However, the mechanisms of action in these interventions that lead to beneficial physical and psychological outcomes have yet to be clearly identified. Purpose: The aim of this paper is to review, systematically, the evidence to date on the mechanisms of action in mindfulness interventions in populations with physical and/or psychological conditions. Method: Searches of seven databases (PsycINFO, Medline (Ovid), Cochrane Central Register of Controlled Trials, EMBASE, CINAHL, AMED, ClinicalTrials.gov) were undertaken in June 2014 and July 2015. We evaluated to what extent the studies we identified met the criteria suggested by Kazdin for establishing mechanisms of action within a psychological treatment (2007, 2009). Results: We identified four trials examining mechanisms of mindfulness interventions in those with comorbid psychological and physical health problems and 14 in those with psychological conditions. These studies examined a diverse range of potential mechanisms, including mindfulness and rumination. Of these candidate mechanisms, the most consistent finding was that greater self-reported change in mindfulness mediated superior clinical outcomes. However, very few studies fully met the Kazdin criteria for examining treatment mechanisms. Conclusion: There was evidence that global changes in mindfulness are linked to better outcomes. This evidence pertained more to interventions targeting psychological rather than physical health conditions. While there is promising evidence that MBCT/MBSR intervention effects are mediated by hypothesised mechanisms, there is a lack of methodological rigour in the field of testing mechanisms of action for both MBCT and MBSR, which precludes definitive conclusions.

Keywords: MBCT, MBSR, mechanisms, physical conditions, psychological conditions, systematic review

Procedia PDF Downloads 315
2325 Applying Spanning Tree Graph Theory for Automatic Database Normalization

Authors: Chetneti Srisa-an

Abstract:

In Knowledge and Data Engineering field, relational database is the best repository to store data in a real world. It has been using around the world more than eight decades. Normalization is the most important process for the analysis and design of relational databases. It aims at creating a set of relational tables with minimum data redundancy that preserve consistency and facilitate correct insertion, deletion, and modification. Normalization is a major task in the design of relational databases. Despite its importance, very few algorithms have been developed to be used in the design of commercial automatic normalization tools. It is also rare technique to do it automatically rather manually. Moreover, for a large and complex database as of now, it make even harder to do it manually. This paper presents a new complete automated relational database normalization method. It produces the directed graph and spanning tree, first. It then proceeds with generating the 2NF, 3NF and also BCNF normal forms. The benefit of this new algorithm is that it can cope with a large set of complex function dependencies.

Keywords: relational database, functional dependency, automatic normalization, primary key, spanning tree

Procedia PDF Downloads 337
2324 Electromagnetic Tuned Mass Damper Approach for Regenerative Suspension

Authors: S. Kopylov, C. Z. Bo

Abstract:

This study is aimed at exploring the possibility of energy recovery through the suppression of vibrations. The article describes design of electromagnetic dynamic damper. The magnetic part of the device performs the function of a tuned mass damper, thereby providing both energy regeneration and damping properties to the protected mass. According to the theory of tuned mass damper, equations of mathematical models were obtained. Then, under given properties of current system, amplitude frequency response was investigated. Therefore, main ideas and methods for further research were defined.

Keywords: electromagnetic damper, oscillations with two degrees of freedom, regeneration systems, tuned mass damper

Procedia PDF Downloads 196
2323 A Context-Sensitive Algorithm for Media Similarity Search

Authors: Guang-Ho Cha

Abstract:

This paper presents a context-sensitive media similarity search algorithm. One of the central problems regarding media search is the semantic gap between the low-level features computed automatically from media data and the human interpretation of them. This is because the notion of similarity is usually based on high-level abstraction but the low-level features do not sometimes reflect the human perception. Many media search algorithms have used the Minkowski metric to measure similarity between image pairs. However those functions cannot adequately capture the aspects of the characteristics of the human visual system as well as the nonlinear relationships in contextual information given by images in a collection. Our search algorithm tackles this problem by employing a similarity measure and a ranking strategy that reflect the nonlinearity of human perception and contextual information in a dataset. Similarity search in an image database based on this contextual information shows encouraging experimental results.

Keywords: context-sensitive search, image search, similarity ranking, similarity search

Procedia PDF Downloads 347
2322 Segmentation of Gray Scale Images of Dropwise Condensation on Textured Surfaces

Authors: Helene Martin, Solmaz Boroomandi Barati, Jean-Charles Pinoli, Stephane Valette, Yann Gavet

Abstract:

In the present work we developed an image processing algorithm to measure water droplets characteristics during dropwise condensation on pillared surfaces. The main problem in this process is the similarity between shape and size of water droplets and the pillars. The developed method divides droplets into four main groups based on their size and applies the corresponding algorithm to segment each group. These algorithms generate binary images of droplets based on both their geometrical and intensity properties. The information related to droplets evolution during time including mean radius and drops number per unit area are then extracted from the binary images. The developed image processing algorithm is verified using manual detection and applied to two different sets of images corresponding to two kinds of pillared surfaces.

Keywords: dropwise condensation, textured surface, image processing, watershed

Procedia PDF Downloads 210
2321 Drug-Based Nanoparticles: Comparative Study of the Effect Drug Type on Release Kinetics and Cell Viability

Authors: Chukwudalu C. Nwazojie, Wole W. Soboyejo, John Obayemi, Ali Salifu Azeko, Sandra M. Jusu, Chinyerem M. Onyekanne

Abstract:

The conventional methods for the diagnosis and treatment of breast cancer include bulk systematic mammography, ultrasound, dynamic contrast-enhanced fast 3D gradient-echo (GRE) magnetic resonance imaging (MRI), surgery, chemotherapy, and radiotherapy. However, nanoparticles and drug-loaded polymer microspheres for disease (cancer) targeting and treatment have enormous potential to enhance the approaches that are used today. The goal is to produce an implantable biomedical device for localized breast cancer drug delivery within Africa and the world. The main advantage of localized delivery is that it reduces the amount of drug that is needed to have a therapeutic effect. Polymer blends of poly (D,L-lactide-co-glycolide) (PLGA) and polycaprolactone (PCL), which are biodegradable, is used as a drug excipient. This work focuses on the development of PLGA-PCL (poly (D,L-lactide-co-glycolide) (PLGA) blended with based injectable drug microspheres and are loaded with anticancer drugs (prodigiosin (PG), and paclitaxel (PTX) control) and also the conjugated forms of the drug functionalized with LHRH (luteinizing hormone-releasing hormone) (PG-LHRH, and PTX- LHRH control), using a single-emulsion solvent evaporation technique. The encapsulation was done in the presence of PLGA-PCL (as a polymer matrix) and poly-(vinyl alcohol) (PVA) (as an emulsifier). Comparative study of the various drugs release kinetics and degradation mechanisms of the PLGA-PCL with an encapsulated drug is achieved, and the implication of this study is for the potential application of prodigiosin PLGA-PCL loaded microparticles for controlled delivery of cancer drug and treatment to prevent the regrowth or locoregional recurrence, following surgical resection of triple-negative breast tumor.

Keywords: cancer, polymers, drug kinetics, nanoparticles

Procedia PDF Downloads 85
2320 Automated Prediction of HIV-associated Cervical Cancer Patients Using Data Mining Techniques for Survival Analysis

Authors: O. J. Akinsola, Yinan Zheng, Rose Anorlu, F. T. Ogunsola, Lifang Hou, Robert Leo-Murphy

Abstract:

Cervical Cancer (CC) is the 2nd most common cancer among women living in low and middle-income countries, with no associated symptoms during formative periods. With the advancement and innovative medical research, there are numerous preventive measures being utilized, but the incidence of cervical cancer cannot be truncated with the application of only screening tests. The mortality associated with this invasive cervical cancer can be nipped in the bud through the important role of early-stage detection. This study research selected an array of different top features selection techniques which was aimed at developing a model that could validly diagnose the risk factors of cervical cancer. A retrospective clinic-based cohort study was conducted on 178 HIV-associated cervical cancer patients in Lagos University teaching Hospital, Nigeria (U54 data repository) in April 2022. The outcome measure was the automated prediction of the HIV-associated cervical cancer cases, while the predictor variables include: demographic information, reproductive history, birth control, sexual history, cervical cancer screening history for invasive cervical cancer. The proposed technique was assessed with R and Python programming software to produce the model by utilizing the classification algorithms for the detection and diagnosis of cervical cancer disease. Four machine learning classification algorithms used are: the machine learning model was split into training and testing dataset into ratio 80:20. The numerical features were also standardized while hyperparameter tuning was carried out on the machine learning to train and test the data. Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), and K-Nearest Neighbor (KNN). Some fitting features were selected for the detection and diagnosis of cervical cancer diseases from selected characteristics in the dataset using the contribution of various selection methods for the classification cervical cancer into healthy or diseased status. The mean age of patients was 49.7±12.1 years, mean age at pregnancy was 23.3±5.5 years, mean age at first sexual experience was 19.4±3.2 years, while the mean BMI was 27.1±5.6 kg/m2. A larger percentage of the patients are Married (62.9%), while most of them have at least two sexual partners (72.5%). Age of patients (OR=1.065, p<0.001**), marital status (OR=0.375, p=0.011**), number of pregnancy live-births (OR=1.317, p=0.007**), and use of birth control pills (OR=0.291, p=0.015**) were found to be significantly associated with HIV-associated cervical cancer. On top ten 10 features (variables) considered in the analysis, RF claims the overall model performance, which include: accuracy of (72.0%), the precision of (84.6%), a recall of (84.6%) and F1-score of (74.0%) while LR has: an accuracy of (74.0%), precision of (70.0%), recall of (70.0%) and F1-score of (70.0%). The RF model identified 10 features predictive of developing cervical cancer. The age of patients was considered as the most important risk factor, followed by the number of pregnancy livebirths, marital status, and use of birth control pills, The study shows that data mining techniques could be used to identify women living with HIV at high risk of developing cervical cancer in Nigeria and other sub-Saharan African countries.

Keywords: associated cervical cancer, data mining, random forest, logistic regression

Procedia PDF Downloads 68
2319 Assessment of Mortgage Applications Using Fuzzy Logic

Authors: Swathi Sampath, V. Kalaichelvi

Abstract:

The assessment of the risk posed by a borrower to a lender is one of the common problems that financial institutions have to deal with. Consumers vying for a mortgage are generally compared to each other by the use of a number called the Credit Score, which is generated by applying a mathematical algorithm to information in the applicant’s credit report. The higher the credit score, the lower the risk posed by the candidate, and the better he is to be taken on by the lender. The objective of the present work is to use fuzzy logic and linguistic rules to create a model that generates Credit Scores.

Keywords: credit scoring, fuzzy logic, mortgage, risk assessment

Procedia PDF Downloads 390
2318 Design of a 4-DOF Robot Manipulator with Optimized Algorithm for Inverse Kinematics

Authors: S. Gómez, G. Sánchez, J. Zarama, M. Castañeda Ramos, J. Escoto Alcántar, J. Torres, A. Núñez, S. Santana, F. Nájera, J. A. Lopez

Abstract:

This paper shows in detail the mathematical model of direct and inverse kinematics for a robot manipulator (welding type) with four degrees of freedom. Using the D-H parameters, screw theory, numerical, geometric and interpolation methods, the theoretical and practical values of the position of robot were determined using an optimized algorithm for inverse kinematics obtaining the values of the particular joints in order to determine the virtual paths in a relatively short time.

Keywords: kinematics, degree of freedom, optimization, robot manipulator

Procedia PDF Downloads 446
2317 Switching to the Latin Alphabet in Kazakhstan: A Brief Overview of Character Recognition Methods

Authors: Ainagul Yermekova, Liudmila Goncharenko, Ali Baghirzade, Sergey Sybachin

Abstract:

In this article, we address the problem of Kazakhstan's transition to the Latin alphabet. The transition process started in 2017 and is scheduled to be completed in 2025. In connection with these events, the problem of recognizing the characters of the new alphabet is raised. Well-known character recognition programs such as ABBYY FineReader, FormReader, MyScript Stylus did not recognize specific Kazakh letters that were used in Cyrillic. The author tries to give an assessment of the well-known method of character recognition that could be in demand as part of the country's transition to the Latin alphabet. Three methods of character recognition: template, structured, and feature-based, are considered through the algorithms of operation. At the end of the article, a general conclusion is made about the possibility of applying a certain method to a particular recognition process: for example, in the process of population census, recognition of typographic text in Latin, or recognition of photos of car numbers, store signs, etc.

Keywords: text detection, template method, recognition algorithm, structured method, feature method

Procedia PDF Downloads 169
2316 Drone Classification Using Classification Methods Using Conventional Model With Embedded Audio-Visual Features

Authors: Hrishi Rakshit, Pooneh Bagheri Zadeh

Abstract:

This paper investigates the performance of drone classification methods using conventional DCNN with different hyperparameters, when additional drone audio data is embedded in the dataset for training and further classification. In this paper, first a custom dataset is created using different images of drones from University of South California (USC) datasets and Leeds Beckett university datasets with embedded drone audio signal. The three well-known DCNN architectures namely, Resnet50, Darknet53 and Shufflenet are employed over the created dataset tuning their hyperparameters such as, learning rates, maximum epochs, Mini Batch size with different optimizers. Precision-Recall curves and F1 Scores-Threshold curves are used to evaluate the performance of the named classification algorithms. Experimental results show that Resnet50 has the highest efficiency compared to other DCNN methods.

Keywords: drone classifications, deep convolutional neural network, hyperparameters, drone audio signal

Procedia PDF Downloads 84
2315 Elucidating the Genetic Determinism of Seed Protein Plasticity in Response to the Environment Using Medicago truncatula

Authors: K. Cartelier, D. Aime, V. Vernoud, J. Buitink, J. M. Prosperi, K. Gallardo, C. Le Signor

Abstract:

Legumes can produce protein-rich seeds without nitrogen fertilizer through root symbiosis with nitrogen-fixing rhizobia. Rich in lysine, these proteins are used for human nutrition and animal feed. However, the instability of seed protein yield and quality due to environmental fluctuations limits the wider use of legumes such as pea. Breeding efforts are needed to optimize and stabilize seed nutritional value, which requires to identify the genetic determinism of seed protein plasticity in response to the environment. Towards this goal, we have studied the plasticity of protein content and composition of seeds from a collection of 200 Medicago truncatula ecotypes grown under four controlled conditions (optimal, drought, and winter/spring sowing). A quantitative analysis of one-dimensional protein profiles of these mature seeds was performed and plasticity indices were calculated from each abundant protein band. Genome-Wide Association Studies (GWAS) from these data identified major GWAS hotspots, from which a list of candidate genes was obtained. A Gene Ontology Enrichment Analysis revealed an over-representation of genes involved in several amino acid metabolic pathways. This led us to propose that environmental variations are likely to modulate amino acid balance, thus impacting seed protein composition. The selection of candidate genes for controlling the plasticity of seed protein composition was refined using transcriptomics data from developing Medicago truncatula seeds. The pea orthologs of key genes were identified for functional studies by mean of TILLING (Targeting Induced Local Lesions in Genomes) lines in this crop. We will present how this study highlighted mechanisms that could govern seed protein plasticity, providing new cues towards the stabilization of legume seed quality.

Keywords: GWAS, Medicago truncatula, plasticity, seed, storage proteins

Procedia PDF Downloads 127
2314 Cognitive Radio in Aeronautic: Comparison of Some Spectrum Sensing Technics

Authors: Abdelkhalek Bouchikhi, Elyes Benmokhtar, Sebastien Saletzki

Abstract:

The aeronautical field is experiencing issues with RF spectrum congestion due to the constant increase in the number of flights, aircrafts and telecom systems on board. In addition, these systems are bulky in size, weight and energy consumption. The cognitive radio helps particularly solving the spectrum congestion issue by its capacity to detect idle frequency channels then, allowing an opportunistic exploitation of the RF spectrum. The present work aims to propose a new use case for aeronautical spectrum sharing and to study the performances of three different detection techniques: energy detector, matched filter and cyclostationary detector within the aeronautical use case. The spectrum in the proposed cognitive radio is allocated dynamically where each cognitive radio follows a cognitive cycle. The spectrum sensing is a crucial step. The goal of the sensing is gathering data about the surrounding environment. Cognitive radio can use different sensors: antennas, cameras, accelerometer, thermometer, etc. In IEEE 802.22 standard, for example, a primary user (PU) has always the priority to communicate. When a frequency channel witch used by the primary user is idle, the secondary user (SU) is allowed to transmit in this channel. The Distance Measuring Equipment (DME) is composed of a UHF transmitter/receiver (interrogator) in the aircraft and a UHF receiver/transmitter on the ground. While the future cognitive radio will be used jointly to alleviate the spectrum congestion issue in the aeronautical field. LDACS, for example, is a good candidate; it provides two isolated data-links: ground-to-air and air-to-ground data-links. The first contribution of the present work is a strategy allowing sharing the L-band. The adopted spectrum sharing strategy is as follow: the DME will play the role of PU which is the licensed user and the LDACS1 systems will be the SUs. The SUs could use the L-band channels opportunely as long as they do not causing harmful interference signals which affect the QoS of the DME system. Although the spectrum sensing is a key step, it helps detecting holes by determining whether the primary signal is present or not in a given frequency channel. A missing detection on primary user presence creates interference between PU and SU and will affect seriously the QoS of the legacy radio. In this study, first brief definitions, concepts and the state of the art of cognitive radio will be presented. Then, a study of three communication channel detection algorithms in a cognitive radio context is carried out. The study is made from the point of view of functions, material requirements and signal detection capability in the aeronautical field. Then, we presented a modeling of the detection problem by three different methods (energy, adapted filter, and cyclostationary) as well as an algorithmic description of these detectors is done. Then, we study and compare the performance of the algorithms. Simulations were carried out using MATLAB software. We analyzed the results based on ROCs curves for SNR between -10dB and 20dB. The three detectors have been tested with a synthetics and real world signals.

Keywords: aeronautic, communication, navigation, surveillance systems, cognitive radio, spectrum sensing, software defined radio

Procedia PDF Downloads 158
2313 Automatic Facial Skin Segmentation Using Possibilistic C-Means Algorithm for Evaluation of Facial Surgeries

Authors: Elham Alaee, Mousa Shamsi, Hossein Ahmadi, Soroosh Nazem, Mohammad Hossein Sedaaghi

Abstract:

Human face has a fundamental role in the appearance of individuals. So the importance of facial surgeries is undeniable. Thus, there is a need for the appropriate and accurate facial skin segmentation in order to extract different features. Since Fuzzy C-Means (FCM) clustering algorithm doesn’t work appropriately for noisy images and outliers, in this paper we exploit Possibilistic C-Means (PCM) algorithm in order to segment the facial skin. For this purpose, first, we convert facial images from RGB to YCbCr color space. To evaluate performance of the proposed algorithm, the database of Sahand University of Technology, Tabriz, Iran was used. In order to have a better understanding from the proposed algorithm; FCM and Expectation-Maximization (EM) algorithms are also used for facial skin segmentation. The proposed method shows better results than the other segmentation methods. Results include misclassification error (0.032) and the region’s area error (0.045) for the proposed algorithm.

Keywords: facial image, segmentation, PCM, FCM, skin error, facial surgery

Procedia PDF Downloads 572
2312 Optimal Design of Redundant Hybrid Manipulator for Minimum Singularity

Authors: Arash Rahmani, Ahmad Ghanbari, Abbas Baghernezhad, Babak Safaei

Abstract:

In the design of parallel manipulators, usually mean value of a dexterity measure over the workspace volume is considered as the objective function to be used in optimization algorithms. The mentioned indexes in a hybrid parallel manipulator (HPM) are quite complicated to solve thanks to infinite solutions for every point within the workspace of the redundant manipulators. In this paper, spatial isotropic design axioms are extended as a well-known method for optimum design of manipulators. An upper limit for the isotropy measure of HPM is calculated and instead of computing and minimizing isotropy measure, minimizing the obtained limit is considered. To this end, two different objective functions are suggested which are obtained from objective functions of comprising modules. Finally, by using genetic algorithm (GA), the best geometric parameters for a specific hybrid parallel robot which is composed of two modified Gough-Stewart platforms (MGSP) are achieved.

Keywords: hybrid manipulator, spatial isotropy, genetic algorithm, optimum design

Procedia PDF Downloads 325
2311 Study on Eco-Feedback of Thermal Comfort and Cost Efficiency for Low Energy Residence

Authors: Y. Jin, N. Zhang, X. Luo, W. Zhang

Abstract:

China with annual increasing 0.5-0.6 billion squares city residence has brought in enormous energy consumption by HVAC facilities and other appliances. In this regard, governments and researchers are encouraging renewable energy like solar energy, geothermal energy using in houses. However, high cost of equipment and low energy conversion result in a very low acceptable to residents. So what’s the equilibrium point of eco-feedback to reach economic benefit and thermal comfort? That is the main question should be answered. In this paper, the objective is an on-site solar PV and heater house, which has been evaluated as a low energy building. Since HVAC system is considered as main energy consumption equipment, the residence with 24-hour monitoring system set to measure temperature, wind velocity and energy in-out value with no HVAC system for one month of summer and winter. Thermal comfort time period will be analyzed and confirmed; then the air-conditioner will be started within thermal discomfort time for the following one summer and winter month. The same data will be recorded to calculate the average energy consumption monthly for a purpose of whole day thermal comfort. Finally, two analysis work will be done: 1) Original building thermal simulation by computer at design stage with actual measured temperature after construction will be contrastive analyzed; 2) The cost of renewable energy facilities and power consumption converted to cost efficient rate to assess the feasibility of renewable energy input for residence. The results of the experiment showed that a certain deviation exists between actual measured data and simulated one for human thermal comfort, especially in summer period. Moreover, the cost-effectiveness is high for a house in targeting city Guilin now with at least 11 years of cost-covering. The conclusion proves that an eco-feedback of a low energy residence is never only consideration of its energy net value, but also the cost efficiency that is the critical factor to push renewable energy acceptable by the public.

Keywords: cost efficiency, eco-feedback, low energy residence, thermal comfort

Procedia PDF Downloads 242
2310 Chaos Cryptography in Cloud Architectures with Lower Latency

Authors: Mohammad A. Alia

Abstract:

With the rapid evolution of the internet applications, cloud computing becomes one of today’s hottest research areas due to its ability to reduce costs associated with computing. Cloud is, therefore, increasing flexibility and scalability for computing services in the internet. Cloud computing is Internet based computing due to shared resources and information which are dynamically delivered to consumers. As cloud computing share resources via the open network, hence cloud outsourcing is vulnerable to attack. Therefore, this paper will explore data security of cloud computing by implementing chaotic cryptography. The proposal scenario develops a problem transformation technique that enables customers to secretly transform their information. This work proposes the chaotic cryptographic algorithms have been applied to enhance the security of the cloud computing accessibility. However, the proposed scenario is secure, easy and straightforward process. The chaotic encryption and digital signature systems ensure the security of the proposed scenario. Though, the choice of the key size becomes crucial to prevent a brute force attack.

Keywords: chaos, cloud computing, security, cryptography

Procedia PDF Downloads 331
2309 Image Reconstruction Method Based on L0 Norm

Authors: Jianhong Xiang, Hao Xiang, Linyu Wang

Abstract:

Compressed sensing (CS) has a wide range of applications in sparse signal reconstruction. Aiming at the problems of low recovery accuracy and long reconstruction time of existing reconstruction algorithms in medical imaging, this paper proposes a corrected smoothing L0 algorithm based on compressed sensing (CSL0). First, an approximate hyperbolic tangent function (AHTF) that is more similar to the L0 norm is proposed to approximate the L0 norm. Secondly, in view of the "sawtooth phenomenon" in the steepest descent method and the problem of sensitivity to the initial value selection in the modified Newton method, the use of the steepest descent method and the modified Newton method are jointly optimized to improve the reconstruction accuracy. Finally, the CSL0 algorithm is simulated on various images. The results show that the algorithm proposed in this paper improves the reconstruction accuracy of the test image by 0-0. 98dB.

Keywords: smoothed L0, compressed sensing, image processing, sparse reconstruction

Procedia PDF Downloads 104
2308 Electrical Decomposition of Time Series of Power Consumption

Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats

Abstract:

Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).

Keywords: electrical disaggregation, DTW, general appliance modeling, event detection

Procedia PDF Downloads 63
2307 Data Poisoning Attacks on Federated Learning and Preventive Measures

Authors: Beulah Rani Inbanathan

Abstract:

In the present era, it is vivid from the numerous outcomes that data privacy is being compromised in various ways. Machine learning is one technology that uses the centralized server, and then data is given as input which is being analyzed by the algorithms present on this mentioned server, and hence outputs are predicted. However, each time the data must be sent by the user as the algorithm will analyze the input data in order to predict the output, which is prone to threats. The solution to overcome this issue is federated learning, where the models alone get updated while the data resides on the local machine and does not get exchanged with the other local models. Nevertheless, even on these local models, there are chances of data poisoning, and it is crystal clear from various experiments done by many people. This paper delves into many ways where data poisoning occurs and the many methods through which it is prevalent that data poisoning still exists. It includes the poisoning attacks on IoT devices, Edge devices, Autoregressive model, and also, on Industrial IoT systems and also, few points on how these could be evadible in order to protect our data which is personal, or sensitive, or harmful when exposed.

Keywords: data poisoning, federated learning, Internet of Things, edge computing

Procedia PDF Downloads 73
2306 A Blockchain-Based Privacy-Preserving Physical Delivery System

Authors: Shahin Zanbaghi, Saeed Samet

Abstract:

The internet has transformed the way we shop. Previously, most of our purchases came in the form of shopping trips to a nearby store. Now, it’s as easy as clicking a mouse. But with great convenience comes great responsibility. We have to be constantly vigilant about our personal information. In this work, our proposed approach is to encrypt the information printed on the physical packages, which include personal information in plain text, using a symmetric encryption algorithm; then, we store that encrypted information into a Blockchain network rather than storing them in companies or corporations centralized databases. We present, implement and assess a blockchain-based system using Ethereum smart contracts. We present detailed algorithms that explain the details of our smart contract. We present the security, cost, and performance analysis of the proposed method. Our work indicates that the proposed solution is economically attainable and provides data integrity, security, transparency, and data traceability.

Keywords: blockchain, Ethereum, smart contract, commit-reveal scheme

Procedia PDF Downloads 135
2305 Detecting Local Clusters of Childhood Malnutrition in the Island Province of Marinduque, Philippines Using Spatial Scan Statistic

Authors: Novee Lor C. Leyso, Maylin C. Palatino

Abstract:

Under-five malnutrition continues to persist in the Philippines, particularly in the island Province of Marinduque, with prevalence of some forms of malnutrition even worsening in recent years. Local spatial cluster detection provides a spatial perspective in understanding this phenomenon as key in analyzing patterns of geographic variation, identification of community-appropriate programs and interventions, and focused targeting on high-risk areas. Using data from a province-wide household-based census conducted in 2014–2016, this study aimed to determine and evaluate spatial clusters of under-five malnutrition, across the province and within each municipality at the individual level using household location. Malnutrition was defined as weight-for-age z-score that fall outside the 2 standard deviations from the median of the WHO reference population. The Kulldorff’s elliptical spatial scan statistic in binomial model was used to locate clusters with high-risk of malnutrition, while adjusting for age and membership to government conditional cash transfer program as proxy for socio-economic status. One large significant cluster of under-five malnutrition was found southwest of the province, in which living in these areas at least doubles the risk of malnutrition. Additionally, at least one significant cluster were identified within each municipality—mostly located along the coastal areas. All these indicate apparent geographical variations across and within municipalities in the province. There were also similarities and disparities in the patterns of risk of malnutrition in each cluster across municipalities, and even within municipality, suggesting underlying causes at work that warrants further investigation. Therefore, community-appropriate programs and interventions should be identified and should be focused on high-risk areas to maximize limited government resources. Further studies are also recommended to determine factors affecting variations in childhood malnutrition considering the evidence of spatial clustering found in this study.

Keywords: Binomial model, Kulldorff’s elliptical spatial scan statistic, Philippines, under-five malnutrition

Procedia PDF Downloads 124
2304 Searching the Efficient Frontier for the Coherent Covering Location Problem

Authors: Felipe Azocar Simonet, Luis Acosta Espejo

Abstract:

In this article, we will try to find an efficient boundary approximation for the bi-objective location problem with coherent coverage for two levels of hierarchy (CCLP). We present the mathematical formulation of the model used. Supported efficient solutions and unsupported efficient solutions are obtained by solving the bi-objective combinatorial problem through the weights method using a Lagrangean heuristic. Subsequently, the results are validated through the DEA analysis with the GEM index (Global efficiency measurement).

Keywords: coherent covering location problem, efficient frontier, lagragian relaxation, data envelopment analysis

Procedia PDF Downloads 316
2303 Computational Fluid Dynamics-Coupled Optimisation Strategy for Aerodynamic Design

Authors: Anvar Atayev, Karl Steinborn, Aleksander Lovric, Saif Al-Ibadi, Jorg Fliege

Abstract:

In this paper, we present results obtained from optimising the aerodynamic performance of aerostructures in external ow. The optimisation method used was developed to efficiently handle multi-variable problems with numerous black-box objective functions and constraints. To demonstrate these capabilities, a series of CFD problems were considered; (1) a two-dimensional NACA aerofoil with three variables, (2) a two-dimensional morphing aerofoil with 17 variables, and (3) a three-dimensional morphing aeroplane tail with 33 variables. The objective functions considered were related to combinations of the mean aerodynamic coefficients, as well as their relative variations/oscillations. It was observed that for each CFD problem, an improved objective value was found. Notably, the scale-up in variables for the latter problems did not greatly hinder optimisation performance. This makes the method promising for scaled-up CFD problems, which require considerable computational resources.

Keywords: computational fluid dynamics, optimisation algorithms, aerodynamic design, engineering design

Procedia PDF Downloads 105
2302 Disease Trajectories in Relation to Poor Sleep Health in the UK Biobank

Authors: Jiajia Peng, Jianqing Qiu, Jianjun Ren, Yu Zhao

Abstract:

Background: Insufficient sleep has been focused on as a public health epidemic. However, a comprehensive analysis of disease trajectory associated with unhealthy sleep habits is still unclear currently. Objective: This study sought to comprehensively clarify the disease's trajectory in relation to the overall poor sleep pattern and unhealthy sleep behaviors separately. Methods: 410,682 participants with available information on sleep behaviors were collected from the UK Biobank at the baseline visit (2006-2010). These participants were classified as having high- and low risk of each sleep behavior and were followed from 2006 to 2020 to identify the increased risks of diseases. We used Cox regression to estimate the associations of high-risk sleep behaviors with the elevated risks of diseases, and further established diseases trajectory using significant diseases. The low-risk unhealthy sleep behaviors were defined as the reference. Thereafter, we also examined the trajectory of diseases linked with the overall poor sleep pattern by combining all of these unhealthy sleep behaviors. To visualize the disease's trajectory, network analysis was used for presenting these trajectories. Results: During a median follow-up of 12.2 years, we noted 12 medical conditions in relation to unhealthy sleep behaviors and the overall poor sleep pattern among 410,682 participants with a median age of 58.0 years. The majority of participants had unhealthy sleep behaviors; in particular, 75.62% with frequent sleeplessness, and 72.12% had abnormal sleep durations. Besides, a total of 16,032 individuals with an overall poor sleep pattern were identified. In general, three major disease clusters were associated with overall poor sleep status and unhealthy sleep behaviors according to the disease trajectory and network analysis, mainly in the digestive, musculoskeletal and connective tissue, and cardiometabolic systems. Of note, two circularity disease pairs (I25→I20 and I48→I50) showed the highest risks following these unhealthy sleep habits. Additionally, significant differences in disease trajectories were observed in relation to sex and sleep medication among individuals with poor sleep status. Conclusions: We identified the major disease clusters and high-risk diseases following participants with overall poor sleep health and unhealthy sleep behaviors, respectively. It may suggest the need to investigate the potential interventions targeting these key pathways.

Keywords: sleep, poor sleep, unhealthy sleep behaviors, disease trajectory, UK Biobank

Procedia PDF Downloads 69