Search results for: risk priority number.
3327 Reducing SAGE Data Using Genetic Algorithms
Authors: Cheng-Hong Yang, Tsung-Mu Shih, Li-Yeh Chuang
Abstract:
Serial Analysis of Gene Expression is a powerful quantification technique for generating cell or tissue gene expression data. The profile of the gene expression of cell or tissue in several different states is difficult for biologists to analyze because of the large number of genes typically involved. However, feature selection in machine learning can successfully reduce this problem. The method allows reducing the features (genes) in specific SAGE data, and determines only relevant genes. In this study, we used a genetic algorithm to implement feature selection, and evaluate the classification accuracy of the selected features with the K-nearest neighbor method. In order to validate the proposed method, we used two SAGE data sets for testing. The results of this study conclusively prove that the number of features of the original SAGE data set can be significantly reduced and higher classification accuracy can be achieved.Keywords: Serial Analysis of Gene Expression, Feature selection, Genetic Algorithm, K-nearest neighbor method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16103326 Bed Site Selection by Wild Boar (Sus scrofa) in Baghshadi Protected Area, Yazd Province, Iran
Authors: S. Aghainajafizadeh, F. Heydari, H. Abbasian
Abstract:
Populations of wild boar present in semi-arid of central Iran. We studied features influencing bed site selection by this species in semi-arid central steppe of Iran. Habitat features of the detected bed site were compared with randomly selected by quantifying number of habitat variables in semi- arid area in Iran. The results revealed that the most important influencing factors in bed site selection were vegetation cover, number of Artemisia sieberi, percentage cover and height of Acer cinerascens, percentage cover and height of Amygdalus scoparia. This is the first ecological study of the wild boar in a protected area of the semi desert biome of Iran. Sustainability of wild boar populations in this area dependent to shrubs of Amygdalus scoparia and Acer cinerascens for thermal and camouflage cover.
Keywords: Wild boar, Bed site selection, Yazd, Iran
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13003325 Machine Learning Techniques in Bank Credit Analysis
Authors: Fernanda M. Assef, Maria Teresinha A. Steiner
Abstract:
The aim of this paper is to compare and discuss better classifier algorithm options for credit risk assessment by applying different Machine Learning techniques. Using records from a Brazilian financial institution, this study uses a database of 5,432 companies that are clients of the bank, where 2,600 clients are classified as non-defaulters, 1,551 are classified as defaulters and 1,281 are temporarily defaulters, meaning that the clients are overdue on their payments for up 180 days. For each case, a total of 15 attributes was considered for a one-against-all assessment using four different techniques: Artificial Neural Networks Multilayer Perceptron (ANN-MLP), Artificial Neural Networks Radial Basis Functions (ANN-RBF), Logistic Regression (LR) and finally Support Vector Machines (SVM). For each method, different parameters were analyzed in order to obtain different results when the best of each technique was compared. Initially the data were coded in thermometer code (numerical attributes) or dummy coding (for nominal attributes). The methods were then evaluated for each parameter and the best result of each technique was compared in terms of accuracy, false positives, false negatives, true positives and true negatives. This comparison showed that the best method, in terms of accuracy, was ANN-RBF (79.20% for non-defaulter classification, 97.74% for defaulters and 75.37% for the temporarily defaulter classification). However, the best accuracy does not always represent the best technique. For instance, on the classification of temporarily defaulters, this technique, in terms of false positives, was surpassed by SVM, which had the lowest rate (0.07%) of false positive classifications. All these intrinsic details are discussed considering the results found, and an overview of what was presented is shown in the conclusion of this study.
Keywords: Artificial Neural Networks, ANNs, classifier algorithms, credit risk assessment, logistic regression, machine learning, support vector machines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12813324 Price Prediction Line, Investment Signals and Limit Conditions Applied for the German Financial Market
Authors: Cristian Păuna
Abstract:
In the first decades of the 21st century, in the electronic trading environment, algorithmic capital investments became the primary tool to make a profit by speculations in financial markets. A significant number of traders, private or institutional investors are participating in the capital markets every day using automated algorithms. The autonomous trading software is today a considerable part in the business intelligence system of any modern financial activity. The trading decisions and orders are made automatically by computers using different mathematical models. This paper will present one of these models called Price Prediction Line. A mathematical algorithm will be revealed to build a reliable trend line, which is the base for limit conditions and automated investment signals, the core for a computerized investment system. The paper will guide how to apply these tools to generate entry and exit investment signals, limit conditions to build a mathematical filter for the investment opportunities, and the methodology to integrate all of these in automated investment software. The paper will also present trading results obtained for the leading German financial market index with the presented methods to analyze and to compare different automated investment algorithms. It was found that a specific mathematical algorithm can be optimized and integrated into an automated trading system with good and sustained results for the leading German Market. Investment results will be compared in order to qualify the presented model. In conclusion, a 1:6.12 risk was obtained to reward ratio applying the trigonometric method to the DAX Deutscher Aktienindex on 24 months investment. These results are superior to those obtained with other similar models as this paper reveal. The general idea sustained by this paper is that the Price Prediction Line model presented is a reliable capital investment methodology that can be successfully applied to build an automated investment system with excellent results.
Keywords: Algorithmic trading, automated investment system, DAX Deutscher Aktienindex.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6963323 Analysis of Education Faculty Students’ Attitudes towards E-Learning According to Different Variables
Authors: Eyup Yurt, Ahmet Kurnaz, Ismail Sahin
Abstract:
The purpose of the study is to investigate the education faculty students’ attitudes towards e-learning according to different variables. In current study, the data were collected from 393 students of an education faculty in Turkey. In this study, theattitude towards e‐learning scale and the demographic information form were used to collect data. The collected data were analyzed by t-test, ANOVA and Pearson correlation coefficient. It was found that there is a significant difference in students’ tendency towards e-learning and avoidance from e-learning based on gender. Male students have more positive attitudes towards e-learning than female students. Also, the students who used the internet lesshave higher levels of avoidance from e-learning. Additionally, it is found that there is a positive and significant relationship between the number of personal mobile learning devices and tendency towards e-learning. On the other hand, there is a negative and significant relationship between the number of personal mobile learning devices and avoidance from e-learning. Also, suggestions were presented according to findings.
Keywords: Education faculty students, attitude towards e-learning, gender, daily Internet usage time, m-learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21643322 Solver for a Magnetic Equivalent Circuit and Modeling the Inrush Current of a 3-Phase Transformer
Authors: Markus G. Ortner, Christian Magele, Klaus Krischan
Abstract:
Knowledge about the magnetic quantities in a magnetic circuit is always of great interest. On the one hand, this information is needed for the simulation of a transformer. On the other hand, parameter studies are more reliable, if the magnetic quantities are derived from a well established model. One possibility to model the 3-phase transformer is by using a magnetic equivalent circuit (MEC). Though this is a well known system, it is often not an easy task to set up such a model for a large number of lumped elements which additionally includes the nonlinear characteristic of the magnetic material. Here we show the setup of a solver for a MEC and the results of the calculation in comparison to measurements taken. The equations of the MEC are based on a rearranged system of the nodal analysis. Thus it is possible to achieve a minimum number of equations, and a clear and simple structure. Hence, it is uncomplicated in its handling and it supports the iteration process. Additional helpful tasks are implemented within the solver to enhance the performance. The electric circuit is described by an electric equivalent circuit (EEC). Our results for the 3-phase transformer demonstrate the computational efficiency of the solver, and show the benefit of the application of a MEC.
Keywords: Inrush current, magnetic equivalent circuit, nonlinear behavior, transformer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24663321 BEM Formulations Based on Kirchhoffs Hypoyhesis to Perform Linear Bending Analysis of Plates Reinforced by Beams
Authors: Gabriela R. Fernandes, Renato F. Denadai, Guido J. Denipotti
Abstract:
In this work, are discussed two formulations of the boundary element method - BEM to perform linear bending analysis of plates reinforced by beams. Both formulations are based on the Kirchhoff's hypothesis and they are obtained from the reciprocity theorem applied to zoned plates, where each sub-region defines a beam or a slab. In the first model the problem values are defined along the interfaces and the external boundary. Then, in order to reduce the number of degrees of freedom kinematics hypothesis are assumed along the beam cross section, leading to a second formulation where the collocation points are defined along the beam skeleton, instead of being placed on interfaces. On these formulations no approximation of the generalized forces along the interface is required. Moreover, compatibility and equilibrium conditions along the interface are automatically imposed by the integral equation. Thus, these formulations require less approximation and the total number of the degree s of freedom is reduced. In the numerical examples are discussed the differences between these two BEM formulations, comparing as well the results to a well-known finite element code.
Keywords: Boundary elements, Building floor structures, Platebending.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16663320 A Medical Resource Forecasting Model for Emergency Room Patients with Acute Hepatitis
Authors: R. J. Kuo, W. C. Cheng, W. C. Lien, T. J. Yang
Abstract:
Taiwan is a hyper endemic area for the Hepatitis B virus (HBV). The estimated total number of HBsAg carriers in the general population who are more than 20 years old is more than 3 million. Therefore, a case record review is conducted from January 2003 to June 2007 for all patients with a diagnosis of acute hepatitis who were admitted to the Emergency Department (ED) of a well-known teaching hospital. The cost for the use of medical resources is defined as the total medical fee. In this study, principal component analysis (PCA) is firstly employed to reduce the number of dimensions. Support vector regression (SVR) and artificial neural network (ANN) are then used to develop the forecasting model. A total of 117 patients meet the inclusion criteria. 61% patients involved in this study are hepatitis B related. The computational result shows that the proposed PCA-SVR model has superior performance than other compared algorithms. In conclusion, the Child-Pugh score and echogram can both be used to predict the cost of medical resources for patients with acute hepatitis in the ED.
Keywords: Acute hepatitis, Medical resource cost, Artificial neural network, Support vector regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19173319 Mechanism of Damping in Welded Structures using Finite Element Approach
Authors: B. Singh, B. K. Nanda
Abstract:
The characterization and modeling of the dynamic behavior of many built-up structures under vibration conditions is still a subject of current research. The present study emphasizes the theoretical investigation of slip damping in layered and jointed welded cantilever structures using finite element approach. Application of finite element method in damping analysis is relatively recent, as such, some problems particularly slip damping analysis has not received enough attention. To validate the finite element model developed, experiments have been conducted on a number of mild steel specimens under different initial conditions of vibration. Finite element model developed affirms that the damping capacity of such structures is influenced by a number of vital parameters such as; pressure distribution, kinematic coefficient of friction and micro-slip at the interfaces, amplitude, frequency of vibration, length and thickness of the specimen. Finite element model developed can be utilized effectively in the design of machine tools, automobiles, aerodynamic and space structures, frames and machine members for enhancing their damping capacity.Keywords: Amplitude, finite element method, slip damping, tack welding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19223318 Project Complexity Indices based on Topology Features
Authors: Amer A. Boushaala
Abstract:
The heuristic decision rules used for project scheduling will vary depending upon the project-s size, complexity, duration, personnel, and owner requirements. The concept of project complexity has received little detailed attention. The need to differentiate between easy and hard problem instances and the interest in isolating the fundamental factors that determine the computing effort required by these procedures inspired a number of researchers to develop various complexity measures. In this study, the most common measures of project complexity are presented. A new measure of project complexity is developed. The main privilege of the proposed measure is that, it considers size, shape and logic characteristics, time characteristics, resource demands and availability characteristics as well as number of critical activities and critical paths. The degree of sensitivity of the proposed measure for complexity of project networks has been tested and evaluated against the other measures of complexity of the considered fifty project networks under consideration in the current study. The developed measure showed more sensitivity to the changes in the network data and gives accurate quantified results when comparing the complexities of networks.Keywords: Activity networks, Complexity index, Networkcomplexity measure, Network topology, Project Network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16813317 Performance Degradation for the GLR Test-Statistics for Spatial Signal Detection
Authors: Olesya Bolkhovskaya, Alexander Maltsev
Abstract:
Antenna arrays are widely used in modern radio systems in sonar and communications. The solving of the detection problems of a useful signal on the background of noise is based on the GLRT method. There is a large number of problem which depends on the known a priori information. In this work, in contrast to the majority of already solved problems, it is used only difference spatial properties of the signal and noise for detection. We are analyzing the influence of the degree of non-coherence of signal and noise unhomogeneity on the performance characteristics of different GLRT statistics. The description of the signal and noise is carried out by means of the spatial covariance matrices C in the cases of different number of known information. The partially coherent signalis is simulated as a plane wave with a random angle of incidence of the wave concerning a normal. Background noise is simulated as random process with uniform distribution function in each element. The results of investigation of degradation of performance characteristics for different cases are represented in this work.
Keywords: GLRT, Neumann-Pearson’s criterion, test-statistics, degradation, spatial processing, multielement antenna array
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18063316 Identification of Non-Lexicon Non-Slang Unigrams in Body-enhancement Medicinal UBE
Authors: Jatinderkumar R. Saini, Apurva A. Desai
Abstract:
Email has become a fast and cheap means of online communication. The main threat to email is Unsolicited Bulk Email (UBE), commonly called spam email. The current work aims at identification of unigrams in more than 2700 UBE that advertise body-enhancement drugs. The identification is based on the requirement that the unigram is neither present in dictionary, nor is a slang term. The motives of the paper are many fold. This is an attempt to analyze spamming behaviour and employment of wordmutation technique. On the side-lines of the paper, we have attempted to better understand the spam, the slang and their interplay. The problem has been addressed by employing Tokenization technique and Unigram BOW model. We found that the non-lexicon words constitute nearly 66% of total number of lexis of corpus whereas non-slang words constitute nearly 2.4% of non-lexicon words. Further, non-lexicon non-slang unigrams composed of 2 lexicon words, form more than 71% of the total number of such unigrams. To the best of our knowledge, this is the first attempt to analyze usage of non-lexicon non-slang unigrams in any kind of UBE.Keywords: Body Enhancement, Lexicon, Medicinal, Slang, Unigram, Unsolicited Bulk e-mail (UBE)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18203315 Simulation of Lean Principles Impact in a Multi-Product Supply Chain
Authors: M. Rossini, A. Portioli Studacher
Abstract:
The market competition is moving from the single firm to the whole supply chain because of increasing competition and growing need for operational efficiencies and customer orientation. Supply chain management allows companies to look beyond their organizational boundaries to develop and leverage resources and capabilities of their supply chain partners. This creates competitive advantages in the marketplace and because of this SCM has acquired strategic importance. Lean Approach is a management strategy that focuses on reducing every type of waste present in an organization. This approach is becoming more and more popular among supply chain managers. The supply chain application of lean approach is not frequent. In particular, it is not well studied which are the impacts of lean approach principles in a supply chain context. In literature there are only few studies aimed at understanding the qualitative impact of the lean approach in supply chains. Therefore, the goal of this research work is to study the impacts of lean principles implementation along a supply chain. To achieve this, a simulation model of a threeechelon multi-product supply chain has been built. Kanban system (and several priority policies) and setup time reduction degrees are implemented in the lean-configured supply chain to apply pull and lot-sizing decrease principles respectively. To evaluate the benefits of lean approach, lean supply chain is compared with an EOQ-configured supply chain. The simulation results show that Kanban system and setup-time reduction improve inventory stock level. They also show that logistics efforts are affected to lean implementation degree. The paper concludes describing performances of lean supply chain in different contexts.Keywords: Inventory policy, Kanban, lean supply chain, simulation study, supply chain management, planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21303314 The Use of Knowledge Management Systems and ICT Service Desk Management to Minimize the Digital Divide Experienced in the Museum Sector
Authors: Ruel A. Welch
Abstract:
Since the introduction of ServiceNow, the UK’s Science Museum Group’s (SMG) ICT service desk portal, there has not been an analysis of the tools available to SMG staff for Just-in-time knowledge acquisition (Knowledge Management Systems) and reporting ICT incidents with a focus on an aspect of professional identity namely, gender. Therefore, it is important for SMG to investigate the apparent disparities so that solutions can be derived to minimize this digital divide if one exists. This study is conducted in the milieu of UK museums, galleries, arts, academic, charitable, and cultural heritage sector. It is acknowledged at SMG that there are challenges with keeping up with an ever-changing digital landscape. Subsequently, this entails the rapid upskilling of staff and developing an infrastructure that supports just-in-time technological knowledge acquisition and reporting technology related issues. This problem was addressed by analysing ServiceNow ICT incident reports and reports from knowledge articles from a six-month period from February to July. This study found a statistically significant relationship between gender and reporting an ICT incident. There is also a significant relationship between gender and the priority level of ICT incident. Interestingly, there is no statistically significant relationship between gender and reading knowledge articles. Additionally, there is no statistically significant relationship between gender and reporting an ICT incident related to the knowledge article that was read by staff. The knowledge acquired from this study is useful to service desk management practice as it will help to inform the creation of future knowledge articles and ICT incident reporting processes.
Keywords: digital divide, ICT service desk practice, knowledge management systems, workplace learning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6393313 Finite Element Analysis and Feasibility of Simple Stochastic Modeling in the Analysis of Fissuring in Grains during Soaking
Authors: Jonathan H. Perez, Fumihiko Tanaka, Daisuke Hamanaka, Toshitaka Uchino
Abstract:
A finite element analysis was conducted to determine the effect of moisture diffusion and hygroscopic swelling in rice. A parallel simple stochastic modeling was performed to predict the number of grains cracked as a result of moisture absorption and hygroscopic swelling. Rice grains were soaked in thermally (25 oC) controlled water and then tested for compressive stress. The destructive compressive stress tests revealed through compressive stress calculation that the peak force required to cause cracking in grains soaked in water reduced with time as soaking duration was extended. Results of the experiment showed that several grains had their value of the predicted compressive stress below the von Mises stress and were interpreted as grains which become cracked and/or broke during soaking. The technique developed in this experiment will facilitate the approximation of the number of grains which will crack during soaking.
Keywords: Cracking, Finite element analysis, hygroscopic swelling, moisture diffusion, von Mises stress.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19183312 Statistical and Land Planning Study of Tourist Arrivals in Greece during 2005-2016
Authors: Dimitra Alexiou
Abstract:
During the last 10 years, in spite of the economic crisis, the number of tourists arriving in Greece has increased, particularly during the tourist season from April to October. In this paper, the number of annual tourist arrivals is studied to explore their preferences with regard to the month of travel, the selected destinations, as well the amount of money spent. The collected data are processed with statistical methods, yielding numerical and graphical results. From the computation of statistical parameters and the forecasting with exponential smoothing, useful conclusions are arrived at that can be used by the Greek tourism authorities, as well as by tourist organizations, for planning purposes for the coming years. The results of this paper and the computed forecast can also be used for decision making by private tourist enterprises that are investing in Greece. With regard to the statistical methods, the method of Simple Exponential Smoothing of time series of data is employed. The search for a best forecast for 2017 and 2018 provides the value of the smoothing coefficient. For all statistical computations and graphics Microsoft Excel is used.
Keywords: Tourism, statistical methods, exponential smoothing, land spatial planning, economy, Microsoft Excel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7073311 Liability Aspects Related to Genetically Modified Food under the Food Safety Legislation in India
Authors: S. K. Balashanmugam, Padmavati Manchikanti, S. R. Subramanian
Abstract:
The question of legal liability over injury arising out of the import and the introduction of GM food emerges as a crucial issue confronting to promote GM food and its derivatives. There is a greater possibility of commercialized GM food from the exporting country to enter importing country where status of approval shall not be same. This necessitates the importance of fixing a liability mechanism to discuss the damage, if any, occurs at the level of transboundary movement or at the market. There was a widespread consensus to develop the Cartagena Protocol on Biosafety and to give for a dedicated regime on liability and redress in the form of Nagoya Kuala Lumpur Supplementary Protocol on the Liability and Redress (‘N-KL Protocol’) at the international context. The national legal frameworks based on this protocol are not adequately established in the prevailing food legislations of the developing countries. The developing economy like India is willing to import GM food and its derivatives after the successful commercialization of Bt Cotton in 2002. As a party to the N-KL Protocol, it is indispensable for India to formulate a legal framework and to discuss safety, liability, and regulatory issues surrounding GM foods in conformity to the provisions of the Protocol. The liability mechanism is also important in the case where the risk assessment and risk management is still in implementing stage. Moreover, the country is facing GM infiltration issues with its neighbors Bangladesh. As a precautionary approach, there is a need to formulate rules and procedure of legal liability to discuss any kind of damage occurs at transboundary trade. In this context, the proposed work will attempt to analyze the liability regime in the existing Food Safety and Standards Act, 2006 from the applicability and domestic compliance and to suggest legal and policy options for regulatory authorities.Keywords: Commercialisation, food safety, FSSAI, genetically modified foods, India, liability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22333310 Physical and Mechanical Phenomena Associated with Rock Failure in Brazilian Disc Specimens
Authors: Hamid Reza Nejati, Amin Nazerigivi, Ahmad Reza Sayadi
Abstract:
Failure mechanism of rocks is one of the fundamental aspects to study rock engineering stability. Rock is a material that contains flaws, initial damage, micro-cracks, etc. Failure of rock structure is largely due to tensile stress and was influenced by various parameters. In the present study, the effect of brittleness and loading rate on the physical and mechanical phenomena produced in rock during loading sequences is considered. For this purpose, Acoustic Emission (AE) technique is used to monitor fracturing process of three rock types (onyx marble, sandstone and soft limestone) with different brittleness and sandstone samples under different loading rate. The results of experimental tests revealed that brittleness and loading rate have a significant effect on the mode and number of induced fracture in rocks. An increase in rock brittleness increases the frequency of induced cracks, and the number of tensile fracture decreases when loading rate increases.Keywords: Brittleness, loading rate, acoustic emission, tensile fracture, shear fracture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14203309 Computer Aided Design Solution Based on Genetic Algorithms for FMEA and Control Plan in Automotive Industry
Authors: Nadia Belu, Laurentiu M. Ionescu, Agnieszka Misztal
Abstract:
In this paper we propose a computer-aided solution with Genetic Algorithms in order to reduce the drafting of reports: FMEA analysis and Control Plan required in the manufacture of the product launch and improved knowledge development teams for future projects. The solution allows to the design team to introduce data entry required to FMEA. The actual analysis is performed using Genetic Algorithms to find optimum between RPN risk factor and cost of production. A feature of Genetic Algorithms is that they are used as a means of finding solutions for multi criteria optimization problems. In our case, along with three specific FMEA risk factors is considered and reduce production cost. Analysis tool will generate final reports for all FMEA processes. The data obtained in FMEA reports are automatically integrated with other entered parameters in Control Plan. Implementation of the solution is in the form of an application running in an intranet on two servers: one containing analysis and plan generation engine and the other containing the database where the initial parameters and results are stored. The results can then be used as starting solutions in the synthesis of other projects. The solution was applied to welding processes, laser cutting and bending to manufacture chassis for buses. Advantages of the solution are efficient elaboration of documents in the current project by automatically generating reports FMEA and Control Plan using multiple criteria optimization of production and build a solid knowledge base for future projects. The solution which we propose is a cheap alternative to other solutions on the market using Open Source tools in implementation.Keywords: Automotive industry, control plan, FMEA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28753308 Mutation Rate for Evolvable Hardware
Authors: Emanuele Stomeo, Tatiana Kalganova, Cyrille Lambert
Abstract:
Evolvable hardware (EHW) refers to a selfreconfiguration hardware design, where the configuration is under the control of an evolutionary algorithm (EA). A lot of research has been done in this area several different EA have been introduced. Every time a specific EA is chosen for solving a particular problem, all its components, such as population size, initialization, selection mechanism, mutation rate, and genetic operators, should be selected in order to achieve the best results. In the last three decade a lot of research has been carried out in order to identify the best parameters for the EA-s components for different “test-problems". However different researchers propose different solutions. In this paper the behaviour of mutation rate on (1+λ) evolution strategy (ES) for designing logic circuits, which has not been done before, has been deeply analyzed. The mutation rate for an EHW system modifies values of the logic cell inputs, the cell type (for example from AND to NOR) and the circuit output. The behaviour of the mutation has been analyzed based on the number of generations, genotype redundancy and number of logic gates used for the evolved circuits. The experimental results found provide the behaviour of the mutation rate to be used during evolution for the design and optimization of logic circuits. The researches on the best mutation rate during the last 40 years are also summarized.Keywords: Evolvable hardware, mutation rate, evolutionarycomputation, design of logic circuit.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15013307 Biogas from Cover Crops and Field Residues: Effects on Soil, Water, Climate and Ecological Footprint
Authors: Manfred Szerencsits, Christine Weinberger, Maximilian Kuderna, Franz Feichtinger, Eva Erhart, Stephan Maier
Abstract:
Cover or catch crops have beneficial effects for soil, water, erosion, etc. If harvested, they also provide feedstock for biogas without competition for arable land in regions, where only one main crop can be produced per year. On average gross energy yields of approx. 1300 m³ methane (CH4) ha-1 can be expected from 4.5 tonnes (t) of cover crop dry matter (DM) in Austria. Considering the total energy invested from cultivation to compression for biofuel use a net energy yield of about 1000 m³ CH4 ha-1 is remaining. With the straw of grain maize or Corn Cob Mix (CCM) similar energy yields can be achieved. In comparison to catch crops remaining on the field as green manure or to complete fallow between main crops the effects on soil, water and climate can be improved if cover crops are harvested without soil compaction and digestate is returned to the field in an amount equivalent to cover crop removal. In this way, the risk of nitrate leaching can be reduced approx. by 25% in comparison to full fallow. The risk of nitrous oxide emissions may be reduced up to 50% by contrast with cover crops serving as green manure. The effects on humus content and erosion are similar or better than those of cover crops used as green manure when the same amount of biomass was produced. With higher biomass production the positive effects increase even if cover crops are harvested and the only digestate is brought back to the fields. The ecological footprint of arable farming can be reduced by approx. 50% considering the substitution of natural gas with CH4 produced from cover crops.
Keywords: Biogas, cover crops, catch crops, land use competition, sustainable agriculture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13163306 Determination of Agricultural Characteristics of Smooth Bromegrass (Bromus inermis Leyss) Lines under Konya Regional Conditions
Authors: Abdullah Özköse, Ahmet Tamkoç
Abstract:
The present study was conducted to determine the yield and yield components of smooth bromegrass lines under the environmental conditions of the Konya region during the growing seasons between 2011 and 2013. The experiment was performed in the randomized complete block design (RCBD) with four replications. It was found that the selected lines had a statistically significant effect on all the investigated traits, except for the main stem length and the number of nodes in the main stem. According to the two-year average calculated for various parameters checked in the smooth bromegrass lines, the main stem length ranged from 71.6 cm to 79.1 cm, the main stem diameter from 2.12 mm from 2.70 mm, the number of nodes in the main stem from 3.2 to 3.7, the internode length from 11.6 cm to 18.9 cm, flag leaf length from 9.7 cm to 12.7 cm, flag leaf width from 3.58 cm to 6.04 mm, herbage yield from 221.3 kg da–1 to 354.7 kg da–1 and hay yield from 100.4 kg da–1 to 190.1 kg da–1. The study concluded that the smooth bromegrass lines differ in terms of yield and yield components. Therefore, it is very crucial to select suitable varieties of smooth bromegrass to obtain optimum yield.
Keywords: Semiarid region, smooth bromegrass, yield, yield components.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12273305 Dimension Reduction of Microarray Data Based on Local Principal Component
Authors: Ali Anaissi, Paul J. Kennedy, Madhu Goyal
Abstract:
Analysis and visualization of microarraydata is veryassistantfor biologists and clinicians in the field of diagnosis and treatment of patients. It allows Clinicians to better understand the structure of microarray and facilitates understanding gene expression in cells. However, microarray dataset is a complex data set and has thousands of features and a very small number of observations. This very high dimensional data set often contains some noise, non-useful information and a small number of relevant features for disease or genotype. This paper proposes a non-linear dimensionality reduction algorithm Local Principal Component (LPC) which aims to maps high dimensional data to a lower dimensional space. The reduced data represents the most important variables underlying the original data. Experimental results and comparisons are presented to show the quality of the proposed algorithm. Moreover, experiments also show how this algorithm reduces high dimensional data whilst preserving the neighbourhoods of the points in the low dimensional space as in the high dimensional space.
Keywords: Linear Dimension Reduction, Non-Linear Dimension Reduction, Principal Component Analysis, Biologists.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15743304 An Experimental Investigation on the Droplet Behavior Impacting a Hot Surface above the Leidenfrost Temperature
Authors: Khaleel Sami Hamdan, Dong-Eok Kim, Sang-Ki Moon
Abstract:
An appropriate model to predict the size of the droplets resulting from the break-up with the structures will help in a better understanding and modeling of the two-phase flow calculations in the simulation of a reactor core loss-of-coolant accident (LOCA). A droplet behavior impacting on a hot surface above the Leidenfrost temperature was investigated. Droplets of known size and velocity were impacted to an inclined plate of hot temperature, and the behavior of the droplets was observed by a high-speed camera. It was found that for droplets of Weber number higher than a certain value, the higher the Weber number of the droplet the smaller the secondary droplets. The COBRA-TF model over-predicted the measured secondary droplet sizes obtained by the present experiment. A simple model for the secondary droplet size was proposed using the mass conservation equation. The maximum spreading diameter of the droplets was also compared to previous correlations and a fairly good agreement was found. A better prediction of the heat transfer in the case of LOCA can be obtained with the presented model.
Keywords: Break-up, droplet, impact, inclined hot plate, Leidenfrost temperature, LOCA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23713303 Micro-Penetrator for Canadian Planetary Exploration
Authors: Michaela Skulinova, Wanping Zheng, Yan-Ru Hu, Yvan Soucy
Abstract:
Space exploration is a highly visible endeavour of humankind to seek profound answers to questions about the origins of our solar system, whether life exists beyond Earth, and how we could live on other worlds. Different platforms have been utilized in planetary exploration missions, such as orbiters, landers, rovers, and penetrators. Having low mass, good mechanical contact with the surface, ability to acquire high quality scientific subsurface data, and ability to be deployed in areas that may not be conducive to landers or rovers, Penetrators provide an alternative and complimentary solution that makes possible scientific exploration of hardly accessible sites (icy areas, gully sites, highlands etc.). The Canadian Space Agency (CSA) has put space exploration as one of the pillars of its space program, and established ExCo program to prepare Canada for future international planetary exploration. ExCo sets surface mobility as its focus and priority, and invests mainly in the development of rovers because of Canada's niche space robotics technology. Meanwhile, CSA is also investigating how micro-penetrators can help Canada to fulfill its scientific objectives for planetary exploration. This paper presents a review of the micro-penetrator technologies, past missions, and lessons learned. It gives a detailed analysis of the technical challenges of micro-penetrators, such as high impact survivability, high precision guidance navigation and control, thermal protection, communications, and etc. Then, a Canadian perspective of a possible micro-penetrator mission is given, including Canadian scientific objectives and priorities, potential instruments, and flight opportunities.Keywords: micro-penetrator, CSA, planetary exploration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26823302 Deterministic Random Number Generator Algorithm for Cryptosystem Keys
Authors: Adi A. Maaita, Hamza A. A. Al_Sewadi
Abstract:
One of the crucial parameters of digital cryptographic systems is the selection of the keys used and their distribution. The randomness of the keys has a strong impact on the system’s security strength being difficult to be predicted, guessed, reproduced, or discovered by a cryptanalyst. Therefore, adequate key randomness generation is still sought for the benefit of stronger cryptosystems. This paper suggests an algorithm designed to generate and test pseudo random number sequences intended for cryptographic applications. This algorithm is based on mathematically manipulating a publically agreed upon information between sender and receiver over a public channel. This information is used as a seed for performing some mathematical functions in order to generate a sequence of pseudorandom numbers that will be used for encryption/decryption purposes. This manipulation involves permutations and substitutions that fulfill Shannon’s principle of “confusion and diffusion”. ASCII code characters were utilized in the generation process instead of using bit strings initially, which adds more flexibility in testing different seed values. Finally, the obtained results would indicate sound difficulty of guessing keys by attackers.Keywords: Cryptosystems, Information Security agreement, Key distribution, Random numbers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34313301 Low Molecular Weight Heparin during Pregnancy
Authors: Sihana Ahmeti Lika, Merita Dauti, Ledjan Malaj
Abstract:
The objective of this study is to analyze the prophylactic usage of low molecular weight heparin (LMWH) along pregnancy and the correlation between their usage and month/week of pregnancy, in the Department of Gynecology and Obstetrics, at Clinical Hospital in Tetovo. A retrospective study was undertaken during 01 January – 31 December 2012. Over of one year, the total number of patients was 4636. Among the 1447 (32.21%) pregnant women, 298 (20.59%) of them were prescribed LMWH. The majority of patients given LMWH, 119 (39.93%) were diagnosed hypercoagulable. The age group with the highest attendance was 25- 35, 141 patients (47.32%). For 195 (65.44%) patients, this was their first pregnancy. Earliest stage of using LMWH was the second month of pregnancy 4 (1.34%) cases. The most common patients were 70 women along the seventh month (23.49%), followed by 68 in the ninth month of pregnancy (22.81%). Women in the 28th gestational week, were found to be the most affected, a total of 55 (78.57%) were in that week. Clexane 2000 and Fraxiparine 0.3 were the most common for which low molecular weight heparin was prescribed. The number of patients which received Clexane 2000 was 84 (28.19%), followed by those with Fraxiparine 0.3 81 (27.18%). The administration of LMWH is associated with long hospitalization (median 14,6 days).
Keywords: Hypercoagulable state, low molecular weight heparin, month of pregnancy, pregnant women.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26423300 The Evaluation of Subclinical Hypothyroidism in Children with Morbid Obesity
Authors: Mustafa M. Donma, Orkide Donma
Abstract:
Cardiovascular (CV) pathology is one of the expected consequences of excessive fat gain. The role of zinc (Zn) in thyroid hormone metabolism (THM) is a matter of debate. Both thyroid stimulating hormone (TSH) and Zn levels are subject to variation in obese individuals. Zn participates in THM. It is closely related to TSH. Since thyroid hormones are required for Zn absorption, hypothyroidism can lead to Zn deficiency and vice versa. Zn exhibits protective effects on CV health and it is inversely correlated with CV markers in childhood obesity. The association between subclinical hypothyroidism (SCHT) and metabolic disorders is under investigation due to its clinical importance. SCHT is defined as the elevated serum TSH levels in the presence of normal free thyroxin (T4) concentrations. The aim of this study is to evaluate the associations between TSH levels and Zn concentrations in SCHT cases detected in morbid obese (MO) children with and without metabolic syndrome (MetS) [(MOMetS+ and MOMetS-)], respectively. 42 children were present in each study group. Informed consent forms were obtained. Tekrdag Namik Kemal University Faculty of Medicine Non-Interventional Clinical Investigations Ethical Committee approved the study protocol. World Health Organization criteria were used for obesity classification. Children with age and sex-dependent body mass index percentile values above 99 were defined as MO. Children exhibiting at least two of MetS criteria were included in MOMetS+ group. Elevated fasting blood glucose, elevated triglycerides (TRG)/decreased high density lipoprotein-cholesterol (HDL-C) concentrations, elevated blood pressure values in addition to central obesity were listed as MetS criteria. Anthropometric measures were recorded. Routine biochemical analyses were performed. In MOMetS- group 13, in MOMetS+ group 15 children were with SCHT. Statistical analyses were performed. p < 0.05 was accepted as statistically significant. In MOMetS- and MOMetS+ groups, TSH levels were 4.1 ± 2.9 mU/L and 4.6 ± 3.1 mU/L, respectively. Corresponding values for SCHT cases were 7.3 ± 3.1 mU/L and 8.0 ± 2.7 mU/L. Free T4 levels were within normal limits. Zn concentrations were negatively correlated with TSH levels in both groups. Significant negative correlation calculated in MOMetS+ group (r = -0.909; p < 0.001) was much stronger than that found in MOMetS- group (r = -0.706; p < 0.05). This strong correlation (r = -0.909; p < 0.001) calculated for cases with SCHT in MOMetS+ group was much lower in the same group (r = -0.793; p < 0.001) when all cases were considered. In conclusion, the presence of strong correlations between TSH and Zn in SCHT in both MOMetS- and MOMetS+ groups have pointed out that MO children were under the threat of CV pathologies. The detection of the much stronger correlation in MOMetS+ group in comparison with the correlation found in MOMetS- group was the indicator of greater CV risk due to the presence of MetS. In MOMetS+ group, correlation in SCHT cases found higher than correlation calculated for all cases confirmed much higher CV risk due to the contribution of SCHT.
Keywords: Cardiovascular risk, child morbid obesity, subclinical hypothyroidism, zinc.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2393299 Effect of Impact Angle on Erosive Abrasive Wear of Ductile and Brittle Materials
Authors: Ergin Kosa, Ali Göksenli
Abstract:
Erosion and abrasion are wear mechanisms reducing the lifetime of machine elements like valves, pump and pipe systems. Both wear mechanisms are acting at the same time, causing a “Synergy” effect, which leads to a rapid damage of the surface. Different parameters are effective on erosive abrasive wear rate. In this study effect of particle impact angle on wear rate and wear mechanism of ductile and brittle materials was investigated. A new slurry pot was designed for experimental investigation. As abrasive particle, silica sand was used. Particle size was ranking between 200- 500 μm. All tests were carried out in a sand-water mixture of 20% concentration for four hours. Impact velocities of the particles were 4.76 m/s. As ductile material steel St 37 with Vickers Hardness Number (VHN) of 245 and quenched St 37 with 510 VHN was used as brittle material. After wear tests, morphology of the eroded surfaces were investigated for better understanding of the wear mechanisms acting at different impact angles by using Scanning Electron Microscope. The results indicated that wear rate of ductile material was higher than brittle material. Maximum wear rate was observed by ductile material at a particle impact angle of 300 and decreased further by an increase in attack angle. Maximum wear rate by brittle materials was by impact angle of 450 and decreased further up to 900. Ploughing was the dominant wear mechanism by ductile material. Microcracks on the surface were detected by ductile materials, which are nucleation centers for crater formation. Number of craters decreased and depth of craters increased by ductile materials by attack angle higher than 300. Deformation wear mechanism was observed by brittle materials. Number and depth of pits decreased by brittle materials by impact angles higher than 450. At the end it is concluded that wear rate could not be directly related to impact angle of particles due to the different reaction of ductile and brittle materials.Keywords: Erosive wear, particle impact angle, silica sand, wear rate, ductile-brittle material.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30233298 Embedded Semantic Segmentation Network Optimized for Matrix Multiplication Accelerator
Authors: Jaeyoung Lee
Abstract:
Autonomous driving systems require high reliability to provide people with a safe and comfortable driving experience. However, despite the development of a number of vehicle sensors, it is difficult to always provide high perceived performance in driving environments that vary from time to season. The image segmentation method using deep learning, which has recently evolved rapidly, provides high recognition performance in various road environments stably. However, since the system controls a vehicle in real time, a highly complex deep learning network cannot be used due to time and memory constraints. Moreover, efficient networks are optimized for GPU environments, which degrade performance in embedded processor environments equipped simple hardware accelerators. In this paper, a semantic segmentation network, matrix multiplication accelerator network (MMANet), optimized for matrix multiplication accelerator (MMA) on Texas instrument digital signal processors (TI DSP) is proposed to improve the recognition performance of autonomous driving system. The proposed method is designed to maximize the number of layers that can be performed in a limited time to provide reliable driving environment information in real time. First, the number of channels in the activation map is fixed to fit the structure of MMA. By increasing the number of parallel branches, the lack of information caused by fixing the number of channels is resolved. Second, an efficient convolution is selected depending on the size of the activation. Since MMA is a fixed, it may be more efficient for normal convolution than depthwise separable convolution depending on memory access overhead. Thus, a convolution type is decided according to output stride to increase network depth. In addition, memory access time is minimized by processing operations only in L3 cache. Lastly, reliable contexts are extracted using the extended atrous spatial pyramid pooling (ASPP). The suggested method gets stable features from an extended path by increasing the kernel size and accessing consecutive data. In addition, it consists of two ASPPs to obtain high quality contexts using the restored shape without global average pooling paths since the layer uses MMA as a simple adder. To verify the proposed method, an experiment is conducted using perfsim, a timing simulator, and the Cityscapes validation sets. The proposed network can process an image with 640 x 480 resolution for 6.67 ms, so six cameras can be used to identify the surroundings of the vehicle as 20 frame per second (FPS). In addition, it achieves 73.1% mean intersection over union (mIoU) which is the highest recognition rate among embedded networks on the Cityscapes validation set.
Keywords: Edge network, embedded network, MMA, matrix multiplication accelerator and semantic segmentation network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 467