Search results for: gaussian selection operator
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3017

Search results for: gaussian selection operator

2207 Method to Find a ε-Optimal Control of Stochastic Differential Equation Driven by a Brownian Motion

Authors: Francys Souza, Alberto Ohashi, Dorival Leao

Abstract:

We present a general solution for finding the ε-optimal controls for non-Markovian stochastic systems as stochastic differential equations driven by Brownian motion, which is a problem recognized as a difficult solution. The contribution appears in the development of mathematical tools to deal with modeling and control of non-Markovian systems, whose applicability in different areas is well known. The methodology used consists to discretize the problem through a random discretization. In this way, we transform an infinite dimensional problem in a finite dimensional, thereafter we use measurable selection arguments, to find a control on an explicit form for the discretized problem. Then, we prove the control found for the discretized problem is a ε-optimal control for the original problem. Our theory provides a concrete description of a rather general class, among the principals, we can highlight financial problems such as portfolio control, hedging, super-hedging, pairs-trading and others. Therefore, our main contribution is the development of a tool to explicitly the ε-optimal control for non-Markovian stochastic systems. The pathwise analysis was made through a random discretization jointly with measurable selection arguments, has provided us with a structure to transform an infinite dimensional problem into a finite dimensional. The theory is applied to stochastic control problems based on path-dependent stochastic differential equations, where both drift and diffusion components are controlled. We are able to explicitly show optimal control with our method.

Keywords: dynamic programming equation, optimal control, stochastic control, stochastic differential equation

Procedia PDF Downloads 183
2206 The Spectroscopic, Molecular Structure and Electrostatic Potential, Polarizability Hyperpolarizability, and Homo–Lumo Analysis of Monomeric and Dimeric Structures of 2-Chloro-N-(2 Methylphenyl) Benzamide

Authors: N. Khelloul, N. Benhalima, A. Chouaih, F. Hamzaoui

Abstract:

The monomer and dimer structures of the title molecule have been obtained from density functional theory (DFT) B3LYP method with 6-31G (d,p) as basis set calculations. The optimized geometrical parameters obtained by B3LYP/6-31G (d,p) method shows good agreement with experimental X-ray data. The polarizability and first order hyperpolarizabilty of the title molecule were calculated and interpreted. The intermolecular N–H•••O hydrogen bonds are discussed in dimer structure of the molecule. The vibrational wave numbers and their assignments were examined theoretically using the Gaussian 09 set of quantum chemistry codes. The predicted frontier molecular orbital energies at B3LYP/6-31G(d,p) method set show that charge transfer occurs within the molecule. The frontier molecular orbital calculations clearly show the inverse relationship of HOMO–LUMO gap with the total static hyperpolarizability. The results also show that 2-Chloro-N-(2-methylphenyl) benzamide 2 molecule may have nonlinear optical (NLO) comportment with non-zero values.

Keywords: DFT, HOMO, LUMO, NLO

Procedia PDF Downloads 336
2205 Impact Evaluation and Technical Efficiency in Ethiopia: Correcting for Selectivity Bias in Stochastic Frontier Analysis

Authors: Tefera Kebede Leyu

Abstract:

The purpose of this study was to estimate the impact of LIVES project participation on the level of technical efficiency of farm households in three regions of Ethiopia. We used household-level data gathered by IRLI between February and April 2014 for the year 2013(retroactive). Data on 1,905 (754 intervention and 1, 151 control groups) sample households were analyzed using STATA software package version 14. Efforts were made to combine stochastic frontier modeling with impact evaluation methodology using the Heckman (1979) two-stage model to deal with possible selectivity bias arising from unobservable characteristics in the stochastic frontier model. Results indicate that farmers in the two groups are not efficient and operate below their potential frontiers i.e., there is a potential to increase crop productivity through efficiency improvements in both groups. In addition, the empirical results revealed selection bias in both groups of farmers confirming the justification for the use of selection bias corrected stochastic frontier model. It was also found that intervention farmers achieved higher technical efficiency scores than the control group of farmers. Furthermore, the selectivity bias-corrected model showed a different technical efficiency score for the intervention farmers while it more or less remained the same for that of control group farmers. However, the control group of farmers shows a higher dispersion as measured by the coefficient of variation compared to the intervention counterparts. Among the explanatory variables, the study found that farmer’s age (proxy to farm experience), land certification, frequency of visit to improved seed center, farmer’s education and row planting are important contributing factors for participation decisions and hence technical efficiency of farmers in the study areas. We recommend that policies targeting the design of development intervention programs in the agricultural sector focus more on providing farmers with on-farm visits by extension workers, provision of credit services, establishment of farmers’ training centers and adoption of modern farm technologies. Finally, we recommend further research to deal with this kind of methodological framework using a panel data set to test whether technical efficiency starts to increase or decrease with the length of time that farmers participate in development programs.

Keywords: impact evaluation, efficiency analysis and selection bias, stochastic frontier model, Heckman-two step

Procedia PDF Downloads 70
2204 Machine Learning Driven Analysis of Kepler Objects of Interest to Identify Exoplanets

Authors: Akshat Kumar, Vidushi

Abstract:

This paper identifies 27 KOIs, 26 of which are currently classified as candidates and one as false positives that have a high probability of being confirmed. For this purpose, 11 machine learning algorithms were implemented on the cumulative kepler dataset sourced from the NASA exoplanet archive; it was observed that the best-performing model was HistGradientBoosting and XGBoost with a test accuracy of 93.5%, and the lowest-performing model was Gaussian NB with a test accuracy of 54%, to test model performance F1, cross-validation score and RUC curve was calculated. Based on the learned models, the significant characteristics for confirm exoplanets were identified, putting emphasis on the object’s transit and stellar properties; these characteristics were namely koi_count, koi_prad, koi_period, koi_dor, koi_ror, and koi_smass, which were later considered to filter out the potential KOIs. The paper also calculates the Earth similarity index based on the planetary radius and equilibrium temperature for each KOI identified to aid in their classification.

Keywords: Kepler objects of interest, exoplanets, space exploration, machine learning, earth similarity index, transit photometry

Procedia PDF Downloads 69
2203 Ship Detection Requirements Analysis for Different Sea States: Validation on Real SAR Data

Authors: Jaime Martín-de-Nicolás, David Mata-Moya, Nerea del-Rey-Maestre, Pedro Gómez-del-Hoyo, María-Pilar Jarabo-Amores

Abstract:

Ship detection is nowadays quite an important issue in tasks related to sea traffic control, fishery management and ship search and rescue. Although it has traditionally been carried out by patrol ships or aircrafts, coverage and weather conditions and sea state can become a problem. Synthetic aperture radars can surpass these coverage limitations and work under any climatological condition. A fast CFAR ship detector based on a robust statistical modeling of sea clutter with respect to sea states in SAR images is used. In this paper, the minimum SNR required to obtain a given detection probability with a given false alarm rate for any sea state is determined. A Gaussian target model using real SAR data is considered. Results show that SNR does not depend heavily on the class considered. Provided there is some variation in the backscattering of targets in SAR imagery, the detection probability is limited and a post-processing stage based on morphology would be suitable.

Keywords: SAR, generalized gamma distribution, detection curves, radar detection

Procedia PDF Downloads 449
2202 Small Entrepreneurs as Creators of Chaos: Increasing Returns Requires Scaling

Authors: M. B. Neace, Xin GAo

Abstract:

Small entrepreneurs are ubiquitous. Regardless of location their success depends on several behavioral characteristics and several market conditions. In this concept paper, we extend this paradigm to include elements from the science of chaos. Our observations, research findings, literature search and intuition lead us to the proposition that all entrepreneurs seek increasing returns, as did the many small entrepreneurs we have interviewed over the years. There will be a few whose initial perturbations may create tsunami-like waves of increasing returns over time resulting in very large market consequences–the butterfly impact. When small entrepreneurs perturb the market-place and their initial efforts take root a series of phase-space transitions begin to occur. They sustain the stream of increasing returns by scaling up. Chaos theory contributes to our understanding of this phenomenon. Sustaining and nourishing increasing returns of small entrepreneurs as complex adaptive systems requires scaling. In this paper we focus on the most critical element of the small entrepreneur scaling process–the mindset of the owner-operator.

Keywords: entrepreneur, increasing returns, scaling, chaos

Procedia PDF Downloads 454
2201 Coordination Behavior, Theoretical Studies, and Biological Activity of Some Transition Metal Complexes with Oxime Ligands

Authors: Noura Kichou, Manel Tafergguenit, Nabila Ghechtouli, Zakia Hank

Abstract:

The aim of this work is to synthesize, characterize and evaluate the biological activity of two Ligands : glyoxime and dimethylglyoxime, and their metal Ni(II) chelates. The newly chelates were characterized by elemental analysis, IR, EPR, nuclear magnetic resonances (1H and 13C), and biological activity. The antibacterial and antifungal activities of the ligands and its metal complexes were screened against bacterial species (Staphylococcus aureus, Bacillus subtilis, and Escherichia coli) and fungi (Candida albicans). Ampicillin and amphotericin were used as references for antibacterial and antifungal studies. The activity data show that the metal complexes have a promising biological activity comparable with parent free ligand against bacterial and fungal species. A structural, energetic, and electronic theoretical study was carried out using the DFT method, with the functional B3LYP and the gaussian program 09. A complete optimization of geometries was made, followed by a calculation of the frequencies of the normal modes of vibration. The UV spectrum was also interpreted. The theoretical results were compared with the experimental data.

Keywords: glyoxime, dimetylglyoxime, nickel, antibacterial activity

Procedia PDF Downloads 96
2200 Coordination Behavior, Theoretical studies and Biological Activity of Some Transition Metal Complexes with Oxime Ligands

Authors: Noura Kichou, Manel Tafergguenit, Nabila Ghechtouli, Zakia Hank

Abstract:

The aim of this work is to synthesize, characterize and evaluate the biological activity of two Ligands: glyoxime and dimethylglyoxime, and their metal Ni(II) chelates. The newly chelates were characterized by elemental analysis, IR, EPR, nuclear magnetic resonances (1H and 13C), and biological activity. The antibacterial and antifungal activities of the ligands and its metal complexes were screened against bacterial species (Staphylococcus aureus, Bacillus subtilis, and Escherichia coli) and fungi (Candida albicans). Ampicillin and amphotericin were used as references for antibacterial and antifungal studies. The activity data show that the metal complexes have a promising biological activity comparable with parent free ligand against bacterial and fungal species. A structural, energetic, and electronic theoretical study was carried out using the DFT method, with the functional B3LYP and the gaussian program 09. A complete optimization of geometries was made, followed by a calculation of the frequencies of the normal modes of vibration. The UV spectrum was also interpreted. The theoretical results were compared with the experimental data.

Keywords: glyoxime, dimetylglyoxime, nickel, antibacterial activity

Procedia PDF Downloads 106
2199 Clustering Based Level Set Evaluation for Low Contrast Images

Authors: Bikshalu Kalagadda, Srikanth Rangu

Abstract:

The important object of images segmentation is to extract objects with respect to some input features. One of the important methods for image segmentation is Level set method. Generally medical images and synthetic images with low contrast of pixel profile, for such images difficult to locate interested features in images. In conventional level set function, develops irregularity during its process of evaluation of contour of objects, this destroy the stability of evolution process. For this problem a remedy is proposed, a new hybrid algorithm is Clustering Level Set Evolution. Kernel fuzzy particles swarm optimization clustering with the Distance Regularized Level Set (DRLS) and Selective Binary, and Gaussian Filtering Regularized Level Set (SBGFRLS) methods are used. The ability of identifying different regions becomes easy with improved speed. Efficiency of the modified method can be evaluated by comparing with the previous method for similar specifications. Comparison can be carried out by considering medical and synthetic images.

Keywords: segmentation, clustering, level set function, re-initialization, Kernel fuzzy, swarm optimization

Procedia PDF Downloads 347
2198 Development of Computational Approach for Calculation of Hydrogen Solubility in Hydrocarbons for Treatment of Petroleum

Authors: Abdulrahman Sumayli, Saad M. AlShahrani

Abstract:

For the hydrogenation process, knowing the solubility of hydrogen (H2) in hydrocarbons is critical to improve the efficiency of the process. We investigated the H2 solubility computation in four heavy crude oil feedstocks using machine learning techniques. Temperature, pressure, and feedstock type were considered as the inputs to the models, while the hydrogen solubility was the sole response. Specifically, we employed three different models: Support Vector Regression (SVR), Gaussian process regression (GPR), and Bayesian ridge regression (BRR). To achieve the best performance, the hyper-parameters of these models are optimized using the whale optimization algorithm (WOA). We evaluated the models using a dataset of solubility measurements in various feedstocks, and we compared their performance based on several metrics. Our results show that the WOA-SVR model tuned with WOA achieves the best performance overall, with an RMSE of 1.38 × 10− 2 and an R-squared of 0.991. These findings suggest that machine learning techniques can provide accurate predictions of hydrogen solubility in different feedstocks, which could be useful in the development of hydrogen-related technologies. Besides, the solubility of hydrogen in the four heavy oil fractions is estimated in different ranges of temperatures and pressures of 150 ◦C–350 ◦C and 1.2 MPa–10.8 MPa, respectively

Keywords: temperature, pressure variations, machine learning, oil treatment

Procedia PDF Downloads 65
2197 Proteome-Wide Convergent Evolution on Vocal Learning Birds Reveals Insight into cAMP-Based Learning Pathway

Authors: Chul Lee, Seoae Cho, Erich D. Jarvis, Heebal Kim

Abstract:

Vocal learning, the ability to imitate vocalizations based on auditory experience, is a homoplastic character state observed in different independent lineages of animals such as songbirds, parrots, hummingbirds and human. It has now become possible to perform genome-wide molecular analyses across vocal learners and vocal non-learners with the recent expansion of avian genome data. It was analyzed the whole genomes of human and 48 avian species including those belonging to the three avian vocal learning lineages, to determine if behavior and neural convergence are associated with molecular convergence in divergent species of vocal learners. Analyses of 8295 orthologous genes across bird species revealed 141 genes with amino acid substitutions specific to vocal learners. Out of these, 25 genes have vocal learner specific genetic homoplasies, and their functions were enriched for learning. Several sites in these genes are estimated under convergent evolution and positive selection. A potential role for a subset of these genes in vocal learning was supported by associations with gene expression profiles in vocal learning brain regions of songbirds and human disease that cause language dysfunctions. The key candidate gene with multiple independent lines of the evidences specific to vocal learners was DRD5. Our findings suggest cAMP-based learning pathway in avian vocal learners, indicating molecular homoplastic changes associated with a complex behavioral trait, vocal learning.

Keywords: amino acid substitutions, convergent evolution, positive selection, vocal learning

Procedia PDF Downloads 336
2196 Artificial Intelligent Methodology for Liquid Propellant Engine Design Optimization

Authors: Hassan Naseh, Javad Roozgard

Abstract:

This paper represents the methodology based on Artificial Intelligent (AI) applied to Liquid Propellant Engine (LPE) optimization. The AI methodology utilized from Adaptive neural Fuzzy Inference System (ANFIS). In this methodology, the optimum objective function means to achieve maximum performance (specific impulse). The independent design variables in ANFIS modeling are combustion chamber pressure and temperature and oxidizer to fuel ratio and output of this modeling are specific impulse that can be applied with other objective functions in LPE design optimization. To this end, the LPE’s parameter has been modeled in ANFIS methodology based on generating fuzzy inference system structure by using grid partitioning, subtractive clustering and Fuzzy C-Means (FCM) clustering for both inferences (Mamdani and Sugeno) and various types of membership functions. The final comparing optimization results shown accuracy and processing run time of the Gaussian ANFIS Methodology between all methods.

Keywords: ANFIS methodology, artificial intelligent, liquid propellant engine, optimization

Procedia PDF Downloads 579
2195 Automatic Staging and Subtype Determination for Non-Small Cell Lung Carcinoma Using PET Image Texture Analysis

Authors: Seyhan Karaçavuş, Bülent Yılmaz, Ömer Kayaaltı, Semra İçer, Arzu Taşdemir, Oğuzhan Ayyıldız, Kübra Eset, Eser Kaya

Abstract:

In this study, our goal was to perform tumor staging and subtype determination automatically using different texture analysis approaches for a very common cancer type, i.e., non-small cell lung carcinoma (NSCLC). Especially, we introduced a texture analysis approach, called Law’s texture filter, to be used in this context for the first time. The 18F-FDG PET images of 42 patients with NSCLC were evaluated. The number of patients for each tumor stage, i.e., I-II, III or IV, was 14. The patients had ~45% adenocarcinoma (ADC) and ~55% squamous cell carcinoma (SqCCs). MATLAB technical computing language was employed in the extraction of 51 features by using first order statistics (FOS), gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and Laws’ texture filters. The feature selection method employed was the sequential forward selection (SFS). Selected textural features were used in the automatic classification by k-nearest neighbors (k-NN) and support vector machines (SVM). In the automatic classification of tumor stage, the accuracy was approximately 59.5% with k-NN classifier (k=3) and 69% with SVM (with one versus one paradigm), using 5 features. In the automatic classification of tumor subtype, the accuracy was around 92.7% with SVM one vs. one. Texture analysis of FDG-PET images might be used, in addition to metabolic parameters as an objective tool to assess tumor histopathological characteristics and in automatic classification of tumor stage and subtype.

Keywords: cancer stage, cancer cell type, non-small cell lung carcinoma, PET, texture analysis

Procedia PDF Downloads 322
2194 Semi-Supervised Learning for Spanish Speech Recognition Using Deep Neural Networks

Authors: B. R. Campomanes-Alvarez, P. Quiros, B. Fernandez

Abstract:

Automatic Speech Recognition (ASR) is a machine-based process of decoding and transcribing oral speech. A typical ASR system receives acoustic input from a speaker or an audio file, analyzes it using algorithms, and produces an output in the form of a text. Some speech recognition systems use Hidden Markov Models (HMMs) to deal with the temporal variability of speech and Gaussian Mixture Models (GMMs) to determine how well each state of each HMM fits a short window of frames of coefficients that represents the acoustic input. Another way to evaluate the fit is to use a feed-forward neural network that takes several frames of coefficients as input and produces posterior probabilities over HMM states as output. Deep neural networks (DNNs) that have many hidden layers and are trained using new methods have been shown to outperform GMMs on a variety of speech recognition systems. Acoustic models for state-of-the-art ASR systems are usually training on massive amounts of data. However, audio files with their corresponding transcriptions can be difficult to obtain, especially in the Spanish language. Hence, in the case of these low-resource scenarios, building an ASR model is considered as a complex task due to the lack of labeled data, resulting in an under-trained system. Semi-supervised learning approaches arise as necessary tasks given the high cost of transcribing audio data. The main goal of this proposal is to develop a procedure based on acoustic semi-supervised learning for Spanish ASR systems by using DNNs. This semi-supervised learning approach consists of: (a) Training a seed ASR model with a DNN using a set of audios and their respective transcriptions. A DNN with a one-hidden-layer network was initialized; increasing the number of hidden layers in training, to a five. A refinement, which consisted of the weight matrix plus bias term and a Stochastic Gradient Descent (SGD) training were also performed. The objective function was the cross-entropy criterion. (b) Decoding/testing a set of unlabeled data with the obtained seed model. (c) Selecting a suitable subset of the validated data to retrain the seed model, thereby improving its performance on the target test set. To choose the most precise transcriptions, three confidence scores or metrics, regarding the lattice concept (based on the graph cost, the acoustic cost and a combination of both), was performed as selection technique. The performance of the ASR system will be calculated by means of the Word Error Rate (WER). The test dataset was renewed in order to extract the new transcriptions added to the training dataset. Some experiments were carried out in order to select the best ASR results. A comparison between a GMM-based model without retraining and the DNN proposed system was also made under the same conditions. Results showed that the semi-supervised ASR-model based on DNNs outperformed the GMM-model, in terms of WER, in all tested cases. The best result obtained an improvement of 6% relative WER. Hence, these promising results suggest that the proposed technique could be suitable for building ASR models in low-resource environments.

Keywords: automatic speech recognition, deep neural networks, machine learning, semi-supervised learning

Procedia PDF Downloads 337
2193 A Remedy for the Confusing Occlusal Principles - An Approach to a Passionate, In-Depth Understanding of Tooth Surfaces Dynamics

Authors: Kariem Elhelow

Abstract:

The task of optimizing teeth surface relations remains perplexing for many dental practitioners. The well-being of teeth, periodontium, and the musculoskeletal system is closely associated with occlusal stability. Dental occlusion is rather far beyond the simple contact of the occlusal surfaces of the opposite jaws, a fact that turned the word “Occlusion” into one of the most complicated puzzles in dentistry. The literature describing the pathological approaches made the practice of occlusion even more intimidating. Understanding the biomechanics of teeth and jaw movements makes the goals of occlusal rehabilitation very lively and simple to practice. The purpose of this article is to establish a path for understanding and practicing the fundamental occlusal principles in a simple yet in depth way. Relying of the evidence based core would deliver a context for showing that occlusion is not as complicated as literatures might reflect. Conclusion: Maintaining a well-defined picture of what a healthy occlusion should be like is very gratifying to both the operator and the patient, with added worth of predictability, esthetics, and function to the whole treatment.

Keywords: occlusal, temporomandibular joint, prosthetic, dentition

Procedia PDF Downloads 127
2192 Analyzing the Effect of Materials’ Selection on Energy Saving and Carbon Footprint: A Case Study Simulation of Concrete Structure Building

Authors: M. Kouhirostamkolaei, M. Kouhirostami, M. Sam, J. Woo, A. T. Asutosh, J. Li, C. Kibert

Abstract:

Construction is one of the most energy consumed activities in the urban environment that results in a significant amount of greenhouse gas emissions around the world. Thus, the impact of the construction industry on global warming is undeniable. Thus, reducing building energy consumption and mitigating carbon production can slow the rate of global warming. The purpose of this study is to determine the amount of energy consumption and carbon dioxide production during the operation phase and the impact of using new shells on energy saving and carbon footprint. Therefore, a residential building with a re-enforced concrete structure is selected in Babolsar, Iran. DesignBuilder software has been used for one year of building operation to calculate the amount of carbon dioxide production and energy consumption in the operation phase of the building. The primary results show the building use 61750 kWh of energy each year. Computer simulation analyzes the effect of changing building shells -using XPS polystyrene and new electrochromic windows- as well as changing the type of lighting on energy consumption reduction and subsequent carbon dioxide production. The results show that the amount of energy and carbon production during building operation has been reduced by approximately 70% by applying the proposed changes. The changes reduce CO2e to 11345 kg CO2/yr. The result of this study helps designers and engineers to consider material selection’s process as one of the most important stages of design for improving energy performance of buildings.

Keywords: construction materials, green construction, energy simulation, carbon footprint, energy saving, concrete structure, designbuilder

Procedia PDF Downloads 194
2191 Towards Automatic Calibration of In-Line Machine Processes

Authors: David F. Nettleton, Elodie Bugnicourt, Christian Wasiak, Alejandro Rosales

Abstract:

In this presentation, preliminary results are given for the modeling and calibration of two different industrial winding MIMO (Multiple Input Multiple Output) processes using machine learning techniques. In contrast to previous approaches which have typically used ‘black-box’ linear statistical methods together with a definition of the mechanical behavior of the process, we use non-linear machine learning algorithms together with a ‘white-box’ rule induction technique to create a supervised model of the fitting error between the expected and real force measures. The final objective is to build a precise model of the winding process in order to control de-tension of the material being wound in the first case, and the friction of the material passing through the die, in the second case. Case 1, Tension Control of a Winding Process. A plastic web is unwound from a first reel, goes over a traction reel and is rewound on a third reel. The objectives are: (i) to train a model to predict the web tension and (ii) calibration to find the input values which result in a given tension. Case 2, Friction Force Control of a Micro-Pullwinding Process. A core+resin passes through a first die, then two winding units wind an outer layer around the core, and a final pass through a second die. The objectives are: (i) to train a model to predict the friction on die2; (ii) calibration to find the input values which result in a given friction on die2. Different machine learning approaches are tested to build models, Kernel Ridge Regression, Support Vector Regression (with a Radial Basis Function Kernel) and MPART (Rule Induction with continuous value as output). As a previous step, the MPART rule induction algorithm was used to build an explicative model of the error (the difference between expected and real friction on die2). The modeling of the error behavior using explicative rules is used to help improve the overall process model. Once the models are built, the inputs are calibrated by generating Gaussian random numbers for each input (taking into account its mean and standard deviation) and comparing the output to a target (desired) output until a closest fit is found. The results of empirical testing show that a high precision is obtained for the trained models and for the calibration process. The learning step is the slowest part of the process (max. 5 minutes for this data), but this can be done offline just once. The calibration step is much faster and in under one minute obtained a precision error of less than 1x10-3 for both outputs. To summarize, in the present work two processes have been modeled and calibrated. A fast processing time and high precision has been achieved, which can be further improved by using heuristics to guide the Gaussian calibration. Error behavior has been modeled to help improve the overall process understanding. This has relevance for the quick optimal set up of many different industrial processes which use a pull-winding type process to manufacture fibre reinforced plastic parts. Acknowledgements to the Openmind project which is funded by Horizon 2020 European Union funding for Research & Innovation, Grant Agreement number 680820

Keywords: data model, machine learning, industrial winding, calibration

Procedia PDF Downloads 238
2190 Appliance of the Analytic Hierarchy Process Methodology for the Selection of a Small Modular Reactors to Enhance Maritime Traffic Decarbonisation

Authors: Sara Martín, Ying Jie Zheng, César Hueso

Abstract:

International shipping is considered one of the largest sources of pollution in the world, accounting for 812 million tons of CO2 emissions in the year 2018. Current maritime decarbonisation is based on the implementation of new fuel alternatives, such as LNG, biofuels, and methanol, among others, which are less polluting as well as less efficient. Despite being a carbon-free and highly-developed technology, nuclear propulsion is hardly discussed as an alternative. Scientifically, it is believed that Small Modular Reactors (SMR) could be a promising solution to decarbonized maritime traffic due to their small dimensions and safety capabilities. However, as of today, there are no merchant ships powered by nuclear systems. Therefore, this project aims to understand the challenges of the development of nuclear-fuelled vessels by analysing all SMR designs to choose the most suitable one. In order not to fall into subjectivities, the Analytic Hierarchy Process (AHP) will be used to make the selection. This multiple-criteria evaluation technique analyses complex decisions by pairwise comparison of a number of evaluation criteria that can be applied to each SMR. The state-of-the-art 72 SMRs presented by the International Atomic Energy Agency (IAEA) will be analysed and ranked by a global parameter, calculated by applying the AHP methodology. The main target of the work is to find an adequate SMR system to power a ship. Top designs will be described in detail, and conclusions will be drawn from the results. This project has been conceived as an effort to foster the near-term development of zero-emission maritime traffic.

Keywords: international shipping, decarbonization, SMR, AHP, nuclear-fuelled vessels

Procedia PDF Downloads 122
2189 Opportunities and Challenges in Midwifery Education: A Literature Review

Authors: Abeer M. Orabi

Abstract:

Midwives are being seen as a key factor in returning birth care to a normal physiologic process that is woman-centered. On the other hand, more needs to be done to increase access for every woman to professional midwifery care. Because of the nature of the midwifery specialty, the magnitude of the effect that can result from a lack of knowledge if midwives make a mistake in their care has the potential to affect a large number of the birthing population. So, the development, running, and management of midwifery educational programs should follow international standards and come after a thorough community needs assessment. At the same time, the number of accredited midwifery educational programs needs to be increased so that larger numbers of midwives will be educated and qualified, as well as access to skilled midwifery care will be increased. Indeed, the selection of promising midwives is important for the successful completion of an educational program, achievement of the program goals, and retention of graduates in the field. Further, the number of schooled midwives in midwifery education programs, their background, and their experience constitute some concerns in the higher education industry. Basically, preceptors and clinical sites are major contributors to the midwifery education process, as educational programs rely on them to provide clinical practice opportunities. In this regard, the selection of clinical training sites should be based on certain criteria to ensure their readiness for the intended training experiences. After that, communication, collaboration, and liaison between teaching faculty and field staff should be maintained. However, the shortage of clinical preceptors and the massive reduction in the number of practicing midwives, in addition to unmanageable workloads, act as significant barriers to midwifery education. Moreover, the medicalized approach inherent in the hospital setting makes it difficult to practice the midwifery model of care, such as watchful waiting, non-interference in normal processes, and judicious use of interventions. Furthermore, creating a motivating study environment is crucial for avoiding unnecessary withdrawal and retention in any educational program. It is well understood that research is an essential component of any profession for achieving its optimal goal and providing a foundation and evidence for its practices, and midwifery is no exception. Midwives have been playing an important role in generating their own research. However, the selection of novel, researchable, and sustainable topics considering community health needs is also a challenge. In conclusion, ongoing education and research are the lifeblood of the midwifery profession to offer a highly competent and qualified workforce. However, many challenges are being faced, and barriers are hindering their improvement.

Keywords: barriers, challenges, midwifery education, educational programs

Procedia PDF Downloads 111
2188 Code Embedding for Software Vulnerability Discovery Based on Semantic Information

Authors: Joseph Gear, Yue Xu, Ernest Foo, Praveen Gauravaran, Zahra Jadidi, Leonie Simpson

Abstract:

Deep learning methods have been seeing an increasing application to the long-standing security research goal of automatic vulnerability detection for source code. Attention, however, must still be paid to the task of producing vector representations for source code (code embeddings) as input for these deep learning models. Graphical representations of code, most predominantly Abstract Syntax Trees and Code Property Graphs, have received some use in this task of late; however, for very large graphs representing very large code snip- pets, learning becomes prohibitively computationally expensive. This expense may be reduced by intelligently pruning this input to only vulnerability-relevant information; however, little research in this area has been performed. Additionally, most existing work comprehends code based solely on the structure of the graph at the expense of the information contained by the node in the graph. This paper proposes Semantic-enhanced Code Embedding for Vulnerability Discovery (SCEVD), a deep learning model which uses semantic-based feature selection for its vulnerability classification model. It uses information from the nodes as well as the structure of the code graph in order to select features which are most indicative of the presence or absence of vulnerabilities. This model is implemented and experimentally tested using the SARD Juliet vulnerability test suite to determine its efficacy. It is able to improve on existing code graph feature selection methods, as demonstrated by its improved ability to discover vulnerabilities.

Keywords: code representation, deep learning, source code semantics, vulnerability discovery

Procedia PDF Downloads 153
2187 The Use of Image Processing Responses Tools Applied to Analysing Bouguer Gravity Anomaly Map (Tangier-Tetuan's Area-Morocco)

Authors: Saad Bakkali

Abstract:

Image processing is a powerful tool for the enhancement of edges in images used in the interpretation of geophysical potential field data. Arial and terrestrial gravimetric surveys were carried out in the region of Tangier-Tetuan. From the observed and measured data of gravity Bouguer gravity anomalies map was prepared. This paper reports the results and interpretations of the transformed maps of Bouguer gravity anomaly of the Tangier-Tetuan area using image processing. Filtering analysis based on classical image process was applied. Operator image process like logarithmic and gamma correction are used. This paper also present the results obtained from this image processing analysis of the enhancement edges of the Bouguer gravity anomaly map of the Tangier-Tetuan zone.

Keywords: bouguer, tangier, filtering, gamma correction, logarithmic enhancement edges

Procedia PDF Downloads 419
2186 Hybrid Model of Strategic and Contextual Leadership in Pluralistic Organizations- A Qualitative Multiple Case Study

Authors: Ergham Al Bachir

Abstract:

This study adopts strategic leadership (Upper Echelons) as the core theory and contextual leadership theory as the research lens. This research asks how the external context impacts strategic leadership effectiveness to achieve the outcomes in pluralistic organizations (PO). The study explores how the context influences the selection of CEOs, top management teams (TMT), and their leadership effectiveness. POs are characterized by the multiple objectives of their top management teams, divergent objectives, multiple strategies, and multiple governing authorities. The research question is explored by means of a qualitative multiple-case study focusing on healthcare, real estate, and financial services organizations. The data sources are semi-structured interviews, documents, and direct observations. The data analysis strategy is inductive and deploys thematic analysis and cross-case synthesis. The findings differentiate between national and international CEOs' delegation of authority and relationship with the Board of Directors. The findings identify the elements of the dynamic context that influence TMT and PO outcomes. The emergent hybrid strategic and contextual leadership framework shows how the different contextual factors influence strategic direction, PO context, selection of CEOs and TMT, and the outcomes in four pluralistic organizations. The study offers seven theoretical contributions to Upper Echelons, strategic leadership, and contextual leadership research. (1) The integration of two theories revealed how CEO’s impact on the organization is complementary to the contextual impact. (2) Conducting this study in the Middle East contributes to strategic leadership and contextual leadership research. (3) The demonstration of the significant contextual effects on the selection of CEOs. (4 and 5) Two contributions revealed new links between the context, the Board role, internal versus external CEOs, and national versus international CEOs. (6 and 7) This study offered two definitions: what accounts for CEO leadership effectiveness and organizational outcomes. Two methodological contributions were also identified: (1) Previous strategic leadership and Upper Echelons research are mainly quantitative, while this study adopts qualitative multiple-case research with face-to-face interviews. (2) The extrication of the CEO from the TMT advanced the data analysis in strategic leadership research. Four contributions are offered to practice: (1) The CEO's leadership effectiveness inside and outside the organization. (2) Rapid turnover of predecessor CEOs signifies the need for a strategic and contextual approach to CEOs' succession. (3) TMT composition and education impact on TMT-CEO and TMT-TMT interface. (4) Multilevel strategic contextual leadership development framework.

Keywords: strategic leadership, contextual leadership, upper echelons, pluralistic organizations, cross-cultural leadership

Procedia PDF Downloads 85
2185 Probability Fuzzy Aggregation Operators in Vehicle Routing Problem

Authors: Anna Sikharulidze, Gia Sirbiladze

Abstract:

For the evaluation of unreliability levels of movement on the closed routes in the vehicle routing problem, the fuzzy operators family is constructed. The interactions between routing factors in extreme conditions on the roads are considered. A multi-criteria decision-making model (MCDM) is constructed. Constructed aggregations are based on the Choquet integral and the associated probability class of a fuzzy measure. Propositions on the correctness of the extension are proved. Connections between the operators and the compositions of dual triangular norms are described. The conjugate connections between the constructed operators are shown. Operators reflect interactions among all the combinations of the factors in the fuzzy MCDM process. Several variants of constructed operators are used in the decision-making problem regarding the assessment of unreliability and possibility levels of movement on closed routes.

Keywords: vehicle routing problem, associated probabilities of a fuzzy measure, choquet integral, fuzzy aggregation operator

Procedia PDF Downloads 324
2184 Hybridization and Evaluation of Jatropha to Improve High Yield Varieties in Indonesia

Authors: Rully D. Purwati, Tantri D.A. Anggraeni, Bambang Heliyanto, M. Machfud, Joko Hartono

Abstract:

The availability of fuel in the world will be reduced in next few years, it is necessary to find alternative energy sources. Jatropha curcas L. is one of oil crops producing non-edible oil which is potential for bio-diesel. Jatropha cultivation and development program in Indonesia is facing several problems especially low seed yield resulting in inefficient crop cultivation cost. To cope with the problem, development of high yielding varieties is necessary. Development of new varieties to improve seed yield was conducted by hybridization and selection and resulted in fourteen potential genotypes. The yield potential of the fourteen genotypes were evaluated and compared with two check varieties. The objective of the evaluation was to find Jatropha hybrids with some characters i.e. their productivity was higher than check varieties, oil content > 40% and harvesting age ≤ 110 days. Hybridization and individual plant selection were carried out from 2010 to 2014. Evaluation of high yield was conducted in Asembagus experimental station, Situbondo, East Java in three years (2015-2017). The experimental designed was Randomized Complete Block Design with three replication, and plot size 10 m x 8 m. The characters observed were number of capsules per plant, dry seed yield (kg/ha) and seed oil content (%). The results of this experiment indicated that all the hybrids evaluated have higher productivity than check variety IP-3A. There were two superior hybrids i.e. HS-49xSP-65/32 and HS-49xSP-19/28 with highest seed yield per hectare and number of capsules per plant for three years.

Keywords: Jatropha, bio energy, hybrid, high seed yield

Procedia PDF Downloads 136
2183 Estimation of the Mean of the Selected Population

Authors: Kalu Ram Meena, Aditi Kar Gangopadhyay, Satrajit Mandal

Abstract:

Two normal populations with different means and same variance are considered, where the variances are known. The population with the smaller sample mean is selected. Various estimators are constructed for the mean of the selected normal population. Finally, they are compared with respect to the bias and MSE risks by the method of Monte-Carlo simulation and their performances are analysed with the help of graphs.

Keywords: estimation after selection, Brewster-Zidek technique, estimators, selected populations

Procedia PDF Downloads 508
2182 Optimization Technique for the Contractor’s Portfolio in the Bidding Process

Authors: Taha Anjamrooz, Sareh Rajabi, Salwa Bheiry

Abstract:

Selection between the available projects in bidding processes for the contractor is one of the essential areas to concentrate on. It is important for the contractor to choose the right projects within its portfolio during the tendering stage based on certain criteria. It should align the bidding process with its origination strategies and goals as a screening process to have the right portfolio pool to start with. Secondly, it should set the proper framework and use a suitable technique in order to optimize its selection process for concertation purpose and higher efforts during the tender stage with goals of success and winning. In this research paper, a two steps framework proposed to increase the efficiency of the contractor’s bidding process and the winning chance of getting the new projects awarded. In this framework, initially, all the projects pass through the first stage screening process, in which the portfolio basket will be evaluated and adjusted in accordance with the organization strategies to the reduced version of the portfolio pool, which is in line with organization activities. In the second stage, the contractor uses linear programming to optimize the portfolio pool based on available resources such as manpower, light equipment, heavy equipment, financial capability, return on investment, and success rate of winning the bid. Therefore, this optimization model will assist the contractor in utilizing its internal resource to its maximum and increase its winning chance for the new project considering past experience with clients, built-relation between two parties, and complexity in the exertion of the projects. The objective of this research will be to increase the contractor's winning chance in the bidding process based on the success rate and expected return on investment.

Keywords: bidding process, internal resources, optimization, contracting portfolio management

Procedia PDF Downloads 140
2181 Evaluating the Effects of an Educational Video on Running Shoe Selection and Subjective Perceptions

Authors: Andrew Fife, Jean-Francois Esculier, Codi Ramsey, Kim Hebert-Losier

Abstract:

Objectives: We aimed to identify how an evidence-based educational video influences how runners select shoes, and perceive shoe comfort, satisfaction, and performance over three months in comparison with a control video. Design: Two groups participated in a double-blind randomised controlled trial. Method: Fifty-six runners were randomly assigned to view one of two video presentations prior to purchasing new shoes for road running in speciality stores. Runners completed a survey with regards to their own shoes and one in reference to the new shoes purchased at three timepoints: before first use, onemonth post-purchase, and three-months post-purchase. Perceived shoe comfort, satisfaction, and performance were assessed using 100 mm visual analogue scales. Factors that influenced their shoe purchase were ranked in order of importance. Results: Comfort and satisfaction were not significantly different between groups and timepoints. The perceived performance of new shoes (75.6 mm) was significantly greater than own shoes (mean: 67.6 mm) before first use, but ratings returned to own-shoe levels one month later in both groups. The group receiving the evidence-based presentation reported their purchased shoes as being influenced more by the video (55.4 mm) than the control group (21.8 mm), although both chose the same brand and model as previously worn over half of the time. Runners in both groups prioritised fit, comfort, and choosing similar shoes to the ones they previously used. Conclusions: In contrast to expectations, the evidence-based educational video did not appear to influence running shoe selection, or overall perceived shoe comfort, satisfaction, or performance.

Keywords: comfort, consumer behaviour, consciousness, education, running, shoes

Procedia PDF Downloads 25
2180 Optimal Feature Extraction Dimension in Finger Vein Recognition Using Kernel Principal Component Analysis

Authors: Amir Hajian, Sepehr Damavandinejadmonfared

Abstract:

In this paper the issue of dimensionality reduction is investigated in finger vein recognition systems using kernel Principal Component Analysis (KPCA). One aspect of KPCA is to find the most appropriate kernel function on finger vein recognition as there are several kernel functions which can be used within PCA-based algorithms. In this paper, however, another side of PCA-based algorithms -particularly KPCA- is investigated. The aspect of dimension of feature vector in PCA-based algorithms is of importance especially when it comes to the real-world applications and usage of such algorithms. It means that a fixed dimension of feature vector has to be set to reduce the dimension of the input and output data and extract the features from them. Then a classifier is performed to classify the data and make the final decision. We analyze KPCA (Polynomial, Gaussian, and Laplacian) in details in this paper and investigate the optimal feature extraction dimension in finger vein recognition using KPCA.

Keywords: biometrics, finger vein recognition, principal component analysis (PCA), kernel principal component analysis (KPCA)

Procedia PDF Downloads 362
2179 Application of Sensory Thermography on Workers of a Wireless Industry in Mexico

Authors: Claudia Camargo Wilson, Enrique Javier de la Vega Bustillos, Jesús Everardo Olguín Tiznado, Juan Andrés López Barreras, Sandra K. Enriquez

Abstract:

This study focuses on the application of sensory thermography, as a non-invasive method to evaluate the musculoskeletal injuries that industry workers performing Highly Repetitive Movements (HRM) may acquire. It was made at a wireless company having the target of analyze temperatures in worker’s wrists, elbows and shoulders in workstations during their activities, this thru sensorial thermography with the goal of detecting maximum temperatures (Tmax) that could indicate possible injuries. The tests were applied during 3 hours for only 2 workers that work in workstations where there’s been the highest index of injuries and accidents. We were made comparisons for each part of the body that were study for both because of the similitude between the activities of the workstations; they were requiring both an immediate evaluation. The Tmax was recorder during the test of the worker 2, in the left wrist, reaching a temperature of 35.088ºC and with a maximum increase of 1.856°C.

Keywords: thermography, maximum temperaturas (Tmax), highly repetitive movements (HRM), operator

Procedia PDF Downloads 398
2178 Pre-Transformation Phase Reconstruction for Deformation-Induced Transformation in AISI 304 Austenitic Stainless Steel

Authors: Manendra Singh Parihar, Sandip Ghosh Chowdhury

Abstract:

Austenitic stainless steels are widely used and give a good combination of properties. When this steel is plastically deformed, a phase transformation of the metastable Face Centred Cubic Austenite to the stable Body Centred Cubic (α’) or to the Hexagonal close packed (ԑ) martensite may occur, leading to the enhancement in the mechanical properties like strength. The work was based on variant selection and corresponding texture analysis for the strain induced martensitic transformation during deformation of the parent austenite FCC phase to form the product HCP and the BCC martensite phases separately, obeying their respective orientation relationships. The automated method for reconstruction of the parent phase orientation using the EBSD data of the product phase orientation is done using the MATLAB and TSL-OIM software. The method of triplets was used which involves the formation of a triplet of neighboring product grains having a common variant and linking them using a misorientation-based criterion. This led to the proper reconstruction of the pre-transformation phase orientation data and thus to its microstructure and texture. The computational speed of current method is better compared to the previously used methods of reconstruction. The reconstruction of austenite from ԑ and α’ martensite was carried out for multiple samples and their IPF images, pole figures, inverse pole figures and ODFs were compared. Similar type of results was observed for all samples. The comparison gives the idea for estimating the correct sequence of the transformation i.e. γ → ε → α’ or γ → α’, during deformation of AISI 304 austenitic stainless steel.

Keywords: variant selection, reconstruction, EBSD, austenitic stainless steel, martensitic transformation

Procedia PDF Downloads 496