Search results for: testing techniques
8793 Air-Coupled Ultrasonic Testing for Non-Destructive Evaluation of Various Aerospace Composite Materials by Laser Vibrometry
Authors: J. Vyas, R. Kazys, J. Sestoke
Abstract:
Air-coupled ultrasonic is the contactless ultrasonic measurement approach which has become widespread for material characterization in Aerospace industry. It is always essential for the requirement of lightest weight, without compromising the durability. To archive the requirements, composite materials are widely used. This paper yields analysis of the air-coupled ultrasonics for composite materials such as CFRP (Carbon Fibre Reinforced Polymer) and GLARE (Glass Fiber Metal Laminate) and honeycombs for the design of modern aircrafts. Laser vibrometry could be the key source of characterization for the aerospace components. The air-coupled ultrasonics fundamentals, including principles, working modes and transducer arrangements used for this purpose is also recounted in brief. The emphasis of this paper is to approach the developed NDT techniques based on the ultrasonic guided waves applications and the possibilities of use of laser vibrometry in different materials with non-contact measurement of guided waves. 3D assessment technique which employs the single point laser head using, automatic scanning relocation of the material to assess the mechanical displacement including pros and cons of the composite materials for aerospace applications with defects and delaminations.Keywords: air-coupled ultrasonics, contactless measurement, laser interferometry, NDT, ultrasonic guided waves
Procedia PDF Downloads 2398792 Study of the Efficiency of a Synthetic Wax for Corrosion Protection of Steel in Aggressive Environments
Authors: Laidi Babouri
Abstract:
The remarkable properties of steel, such as hardness and impact resistance, motivate their use in the automotive manufacturing industry. However, due to the very vulnerable environmental conditions of use, the steel that makes up the car body can corrode. This situation is motivating more and more automobile manufacturers to develop research to develop processes minimizing the rate of degradation of the physicomechanical properties of these materials. The present work falls within this perspective; it presents the results of a research study focused on the use of synthetic wax for the protection of steel, type XES (DC04), against corrosion in aggressive environments. The media used in this study are an acid medium with a pH=5.6, a 3% chloride medium, and a dry medium. Evaluation of the protective power of synthetic wax in different environments was carried out using mass loss techniques (immersion), completed by electrochemical techniques (stationary and transient). The results of the immersion of the steel samples, with a surface area of (1.44 cm²), in the various media, for a period of 30 days, using the immersion technique, showed high protective efficiency of synthetic wax in acidic and saline environments, with a lesser degree in a dry environment. Moreover, the study of the protective power, using electrochemical techniques, confirmed the results obtained in static mode (loss of mass), the protective efficiency of synthetic wax, against the corrosion of steel, in different environments, which reaches a maximum rate of 99.87% in a saline environment.Keywords: corrosion, steel, industrial wax, environment, mass loss, electrochemical techniques
Procedia PDF Downloads 758791 Comparing the Detection of Autism Spectrum Disorder within Males and Females Using Machine Learning Techniques
Authors: Joseph Wolff, Jeffrey Eilbott
Abstract:
Autism Spectrum Disorders (ASD) are a spectrum of social disorders characterized by deficits in social communication, verbal ability, and interaction that can vary in severity. In recent years, researchers have used magnetic resonance imaging (MRI) to help detect how neural patterns in individuals with ASD differ from those of neurotypical (NT) controls for classification purposes. This study analyzed the classification of ASD within males and females using functional MRI data. Functional connectivity (FC) correlations among brain regions were used as feature inputs for machine learning algorithms. Analysis was performed on 558 cases from the Autism Brain Imaging Data Exchange (ABIDE) I dataset. When trained specifically on females, the algorithm underperformed in classifying the ASD subset of our testing population. Although the subject size was relatively smaller in the female group, the manual matching of both male and female training groups helps explain the algorithm’s bias, indicating the altered sex abnormalities in functional brain networks compared to typically developing peers. These results highlight the importance of taking sex into account when considering how generalizations of findings on males with ASD apply to females.Keywords: autism spectrum disorder, machine learning, neuroimaging, sex differences
Procedia PDF Downloads 2098790 Analysis of Automotive Sensor for Engine Knock System
Authors: Miroslav Gutten, Jozef Jurcik, Daniel Korenciak, Milan Sebok, Matej Kuceraa
Abstract:
This paper deals with the phenomenon of the undesirable detonation combustion in internal combustion engines. A control unit of the engine monitors these detonations using piezoelectric knock sensors. With the control of these sensors the detonations can be objectively measured just outside the car. If this component provides small amplitude of the output voltage it could happen that there would have been in the areas of the engine ignition combustion. The paper deals with the design of a simple device for the detection of this disorder. A construction of the testing device for the knock sensor suitable for diagnostics of knock combustion in internal combustion engines will be presented. The output signal of presented sensor will be described by Bessel functions. Using the first voltage extremes on the characteristics it is possible to create a reference for the evaluation of the polynomial residue. It should be taken into account that the velocity of sound in air is 330 m/s. This sound impinges on the walls of the combustion chamber and is detected by the sensor. The resonant frequency of the clicking of the motor is usually in the range from 5 kHz to 15 kHz. The sensor worked in the field to 37 kHz, which shall be taken into account on an own sensor resonance.Keywords: diagnostics, knock sensor, measurement, testing device
Procedia PDF Downloads 4478789 Techniques for Seismic Strengthening of Historical Monuments from Diagnosis to Implementation
Authors: Mircan Kaya
Abstract:
A multi-disciplinary approach is required in any intervention project for historical monuments. Due to the complexity of their geometry, the variable and unpredictable characteristics of original materials used in their creation, heritage structures are peculiar. Their histories are often complex, and they require correct diagnoses to decide on the techniques of intervention. This approach should not only combine technical aspects but also historical research that may help discover phenomena involving structural issues, and acquire a knowledge of the structure on its concept, method of construction, previous interventions, process of damage, and its current state. In addition to the traditional techniques like bed joint reinforcement, the repairing, strengthening and restoration of historical buildings may require several other modern methods which may be described as innovative techniques like pre-stressing and post-tensioning, use of shape memory alloy devices and shock transmission units, shoring, drilling, and the use of stainless steel or titanium. Regardless of the method to be incorporated in the strengthening process, which can be traditional or innovative, it is crucial to recognize that structural strengthening is the process of upgrading the structural system of the existing building with the aim of improving its performance under existing and additional loads like seismic loads. This process is much more complex than dealing with a new construction, owing to the fact that there are several unknown factors associated with the structural system. Material properties, load paths, previous interventions, existing reinforcement are especially important matters to be considered. There are several examples of seismic strengthening with traditional and innovative techniques around the world, which will be discussed in this paper in detail, including their pros and cons. Ultimately, however, the main idea underlying the philosophy of a successful intervention with the most appropriate techniques of strengthening a historic monument should be decided by a proper assessment of the specific needs of the building.Keywords: bed joint reinforcement, historical monuments, post-tensioning, pre-stressing, seismic strengthening, shape memory alloy devices, shock transmitters, tie rods
Procedia PDF Downloads 2648788 An Artificial Intelligence Framework to Forecast Air Quality
Authors: Richard Ren
Abstract:
Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms
Procedia PDF Downloads 1278787 Identification of Promising Infant Clusters to Obtain Improved Block Layout Designs
Authors: Mustahsan Mir, Ahmed Hassanin, Mohammed A. Al-Saleh
Abstract:
The layout optimization of building blocks of unequal areas has applications in many disciplines including VLSI floorplanning, macrocell placement, unequal-area facilities layout optimization, and plant or machine layout design. A number of heuristics and some analytical and hybrid techniques have been published to solve this problem. This paper presents an efficient high-quality building-block layout design technique especially suited for solving large-size problems. The higher efficiency and improved quality of optimized solutions are made possible by introducing the concept of Promising Infant Clusters in a constructive placement procedure. The results presented in the paper demonstrate the improved performance of the presented technique for benchmark problems in comparison with published heuristic, analytic, and hybrid techniques.Keywords: block layout problem, building-block layout design, CAD, optimization, search techniques
Procedia PDF Downloads 3868786 Mothers’ Experiences of Continuing Their Pregnancy after Prenatally Receiving a Diagnosis of Down Syndrome
Authors: Sevinj Asgarova
Abstract:
Within the last few decades, major advances in the field of prenatal testing have transpired yet little research regarding the experiences of mothers who chose to continue their pregnancies after prenatally receiving a diagnosis of Down Syndrome (DS) has been undertaken. Using social constructionism and interpretive description, this retrospective research study explores this topic from the point of view of the mothers involved and provides insight as to how the experience could be improved. Using purposive sampling, 23 mothers were recruited from British Columbia (n=11) and Ontario (n=12) in Canada. Data retrieved through semi-structured in-depth interviews were analyzed using inductive, constant comparative analysis, the major analytical techniques of interpretive description. Four primary phases emerged from the data analysis 1) healthcare professional-mothers communications, 2) initial emotional response, 3) subsequent decision-making and 4) an adjustment and reorganization of lifestyle to the preparation for the birth of the child. This study validates the individualized and contextualized nature of mothers’ decisions as influenced by multiple factors, with moral values/spiritual beliefs being significant. The mothers’ ability to cope was affected by the information communicated to them about their unborn baby’s diagnosis and the manner in which that information was delivered to them. Mothers used emotional coping strategies, dependent upon support from partners, family, and friends, as well as from other families who have children with DS. Additionally, they employed practical coping strategies, such as engaging in healthcare planning, seeking relevant information, and reimagining and reorganizing their lifestyle. Over time many families gained a sense of control over their situation and readjusted to the preparation for the birth of the child. Many mothers expressed the importance of maintaining positivity and hopefulness with respect to positive outcomes and opportunities for their children. The comprehensive information generated through this study will also provide healthcare professionals with relevant information to assist them in understanding the informational and emotional needs of these mothers. This should lead to an improvement in their practice and enhance their ability to intervene appropriately and effectively, better offering improved support to parents dealing with a diagnosis of DS for their child.Keywords: continuing affected pregnancy, decision making, disability, down syndrome, eugenic social attitudes, inequalities, life change events, prenatal care, prenatal testing, qualitative research, social change, social justice
Procedia PDF Downloads 1038785 Impact of Masonry Joints on Detection of Humidity Distribution in Aerated Concrete Masonry Constructions by Electric Impedance Spectrometry Measurements
Authors: Sanita Rubene, Martins Vilnitis, Juris Noviks
Abstract:
Aerated concrete is a load bearing construction material, which has high heat insulation parameters. Walls can be erected from aerated concrete masonry constructions and in perfect circumstances additional heat insulation is not required. The most common problem in aerated concrete heat insulation properties is the humidity distribution throughout the cross section of the masonry elements as well as proper and conducted drying process of the aerated concrete construction because only dry aerated concrete masonry constructions can reach high heat insulation parameters. In order to monitor drying process of the masonry and detect humidity distribution throughout the cross section of aerated concrete masonry construction application of electrical impedance spectrometry is applied. Further test results and methodology of this non-destructive testing method is described in this paper.Keywords: aerated concrete, electrical impedance spectrometry, humidity distribution, non-destructive testing
Procedia PDF Downloads 3298784 GIS for Simulating Air Traffic by Applying Different Multi-radar Positioning Techniques
Authors: Amara Rafik, Bougherara Maamar, Belhadj Aissa Mostefa
Abstract:
Radar data is one of the many data sources used by ATM Air Traffic Management systems. These data come from air navigation radar antennas. These radars intercept signals emitted by the various aircraft crossing the controlled airspace and calculate the position of these aircraft and retransmit their positions to the Air Traffic Management System. For greater reliability, these radars are positioned in such a way as to allow their coverage areas to overlap. An aircraft will therefore be detected by at least one of these radars. However, the position coordinates of the same aircraft and sent by these different radars are not necessarily identical. Therefore, the ATM system must calculate a single position (radar track) which will ultimately be sent to the control position and displayed on the air traffic controller's monitor. There are several techniques for calculating the radar track. Furthermore, the geographical nature of the problem requires the use of a Geographic Information System (GIS), i.e. a geographical database on the one hand and geographical processing. The objective of this work is to propose a GIS for traffic simulation which reconstructs the evolution over time of aircraft positions from a multi-source radar data set and by applying these different techniques.Keywords: ATM, GIS, radar data, air traffic simulation
Procedia PDF Downloads 858783 Determination of the Local Elastic Moduli of Shungite by Laser Ultrasonic Spectroscopy
Authors: Elena B. Cherepetskaya, Alexander A.Karabutov, Vladimir A. Makarov, Elena A. Mironova, Ivan A. Shibaev
Abstract:
In our study, the object of laser ultrasonic testing was plane-parallel plate of shungit (length 41 mm, width 31 mm, height 15 mm, medium exchange density 2247 kg/m3). We used laser-ultrasonic defectoscope with wideband opto-acoustic transducer in our investigation of the velocities of longitudinal and shear elastic ultrasound waves. The duration of arising elastic pulses was less than 100 ns. Under known material thickness, the values of the velocities were determined by the time delay of the pulses reflected from the bottom surface of the sample with respect to reference pulses. The accuracy of measurement was 0.3% in the case of longitudinal wave velocity and 0.5% in the case of shear wave velocity (scanning pitch along the surface was 2 mm). On the base of found velocities of elastic waves, local elastic moduli of shungit (Young modulus, shear modulus and Poisson's ratio) were uniquely determined.Keywords: laser ultrasonic testing , local elastic moduli, shear wave velocity, shungit
Procedia PDF Downloads 3088782 A Survey on Data-Centric and Data-Aware Techniques for Large Scale Infrastructures
Authors: Silvina Caíno-Lores, Jesús Carretero
Abstract:
Large scale computing infrastructures have been widely developed with the core objective of providing a suitable platform for high-performance and high-throughput computing. These systems are designed to support resource-intensive and complex applications, which can be found in many scientific and industrial areas. Currently, large scale data-intensive applications are hindered by the high latencies that result from the access to vastly distributed data. Recent works have suggested that improving data locality is key to move towards exascale infrastructures efficiently, as solutions to this problem aim to reduce the bandwidth consumed in data transfers, and the overheads that arise from them. There are several techniques that attempt to move computations closer to the data. In this survey we analyse the different mechanisms that have been proposed to provide data locality for large scale high-performance and high-throughput systems. This survey intends to assist scientific computing community in understanding the various technical aspects and strategies that have been reported in recent literature regarding data locality. As a result, we present an overview of locality-oriented techniques, which are grouped in four main categories: application development, task scheduling, in-memory computing and storage platforms. Finally, the authors include a discussion on future research lines and synergies among the former techniques.Keywords: data locality, data-centric computing, large scale infrastructures, cloud computing
Procedia PDF Downloads 2598781 Towards Modern Approaches of Intelligence Measurement for Clinical and Educational Practices
Authors: Alena Kulikova, Tatjana Kanonire
Abstract:
Intelligence research is one of the oldest fields of psychology. Many factors have made a research on intelligence, defined as reasoning and problem solving [1, 2], a very acute and urgent problem. Thus, it has been repeatedly shown that intelligence is a predictor of academic, professional, and social achievement in adulthood (for example, [3]); Moreover, intelligence predicts these achievements better than any other trait or ability [4]. The individual level, a comprehensive assessment of intelligence is a necessary criterion for the diagnosis of various mental conditions. For example, it is a necessary condition for psychological, medical and pedagogical commissions when deciding on educational needs and the most appropriate educational programs for school children. Assessment of intelligence is crucial in clinical psychodiagnostic and needs high-quality intelligence measurement tools. Therefore, it is not surprising that the development of intelligence tests is an essential part of psychological science and practice. Many modern intelligence tests have a long history and have been used for decades, for example, the Stanford-Binet test or the Wechsler test. However, the vast majority of these tests are based on the classic linear test structure, in which all respondents receive all tasks (see, for example, a critical review by [5]). This understanding of the testing procedure is a legacy of the pre-computer era, in which blank testing was the only diagnostic procedure available [6] and has some significant limitations that affect the reliability of the data obtained [7] and increased time costs. Another problem with measuring IQ is that classical line-structured tests do not fully allow to measure respondent's intellectual progress [8], which is undoubtedly a critical limitation. Advances in modern psychometrics allow for avoiding the limitations of existing tools. However, as in any rapidly developing industry, at the moment, psychometrics does not offer ready-made and straightforward solutions and requires additional research. In our presentation we would like to discuss the strengths and weaknesses of the current approaches to intelligence measurement and highlight “points of growth” for creating a test in accordance with modern psychometrics. Whether it is possible to create the instrument that will use all achievements of modern psychometric and remain valid and practically oriented. What would be the possible limitations for such an instrument? The theoretical framework and study design to create and validate the original Russian comprehensive computer test for measuring the intellectual development in school-age children will be presented.Keywords: Intelligence, psychometrics, psychological measurement, computerized adaptive testing, multistage testing
Procedia PDF Downloads 808780 Rapid Processing Techniques Applied to Sintered Nickel Battery Technologies for Utility Scale Applications
Authors: J. D. Marinaccio, I. Mabbett, C. Glover, D. Worsley
Abstract:
Through use of novel modern/rapid processing techniques such as screen printing and Near-Infrared (NIR) radiative curing, process time for the sintering of sintered nickel plaques, applicable to alkaline nickel battery chemistries, has been drastically reduced from in excess of 200 minutes with conventional convection methods to below 2 minutes using NIR curing methods. Steps have also been taken to remove the need for forming gas as a reducing agent by implementing carbon as an in-situ reducing agent, within the ink formulation.Keywords: batteries, energy, iron, nickel, storage
Procedia PDF Downloads 4398779 Estimation of Grinding Force and Material Characterization of Ceramic Matrix Composite
Authors: Lakshminarayanan, Vijayaraghavan, Krishnamurthy
Abstract:
The ever-increasing demand for high efficiency in automotive and aerospace applications requires new materials to suit to high temperature applications. The Ceramic Matrix Composites nowadays find its applications for high strength and high temperature environments. In this paper, Al2O3 and Sic ceramic materials are taken in particulate form as matrix and reinforcement respectively. They are blended together in Ball Milling and compacted in Cold Compaction Machine by powder metallurgy technique. Scanning Electron Microscope images are taken for the samples in order to find out proper blending of powders. Micro harness testing is also carried out for the samples in Vickers Micro Hardness Testing Equipment. Surface grinding of the samples is also carried out in Surface Grinding Machine in order to find out grinding force estimates. The surface roughness of the grounded samples is also taken in Surface Profilometer. These are yielding promising results.Keywords: ceramic matrix composite, cold compaction, material characterization, particulate and surface grinding
Procedia PDF Downloads 2428778 Scalable Cloud-Based LEO Satellite Constellation Simulator
Authors: Karim Sobh, Khaled El-Ayat, Fady Morcos, Amr El-Kadi
Abstract:
Distributed applications deployed on LEO satellites and ground stations require substantial communication between different members in a constellation to overcome the earth coverage barriers imposed by GEOs. Applications running on LEO constellations suffer the earth line-of-sight blockage effect. They need adequate lab testing before launching to space. We propose a scalable cloud-based net-work simulation framework to simulate problems created by the earth line-of-sight blockage. The framework utilized cloud IaaS virtual machines to simulate LEO satellites and ground stations distributed software. A factorial ANOVA statistical analysis is conducted to measure simulator overhead on overall communication performance. The results showed a very low simulator communication overhead. Consequently, the simulation framework is proposed as a candidate for testing LEO constellations with distributed software in the lab before space launch.Keywords: LEO, cloud computing, constellation, satellite, network simulation, netfilter
Procedia PDF Downloads 3868777 Testing the Moderating Effect of Sub Ethnic on Household Investment Behaviour
Authors: Widayat Widayat
Abstract:
Nowday, in the modern investment era, household behavior on investment is a topic that is quite warm. The development of the modern investment, indicated by the emergence of a variety of investment instruments, such as stocks, bonds and various forms of derivatives, affected on the complexity of choosing an investment, especially for traditional societies. Various studies show that there is more than one factor acting as a behavioral antesenden decide to choose an investment instrument. One of the factors, which contribute in determining the investment option is ethnic. Society with a particular sub-culture tend to prefer investing their particular instrument. This is because they have the values, norms and different social environmental. This article is designed to test the impact of sub-cultures between Osing-Java as moderator, in investing. The study was conducted in Banyuwangi, East Java Province of Indonesia. Data were collected using questionnaires, which is given to the head of the household respondents were selected as samples. Sample of households selected by multistage sampling method. The data have been collected processed using SmartPLS software and testing moderating effects using grouped sample test. The result showed that sub-ethnic and has a significant role in determining the investment.Keywords: investment behaviour, household, moderating, sub ethnic
Procedia PDF Downloads 3718776 Design and Implementation a Platform for Adaptive Online Learning Based on Fuzzy Logic
Authors: Budoor Al Abid
Abstract:
Educational systems are increasingly provided as open online services, providing guidance and support for individual learners. To adapt the learning systems, a proper evaluation must be made. This paper builds the evaluation model Fuzzy C Means Adaptive System (FCMAS) based on data mining techniques to assess the difficulty of the questions. The following steps are implemented; first using a dataset from an online international learning system called (slepemapy.cz) the dataset contains over 1300000 records with 9 features for students, questions and answers information with feedback evaluation. Next, a normalization process as preprocessing step was applied. Then FCM clustering algorithms are used to adaptive the difficulty of the questions. The result is three cluster labeled data depending on the higher Wight (easy, Intermediate, difficult). The FCM algorithm gives a label to all the questions one by one. Then Random Forest (RF) Classifier model is constructed on the clustered dataset uses 70% of the dataset for training and 30% for testing; the result of the model is a 99.9% accuracy rate. This approach improves the Adaptive E-learning system because it depends on the student behavior and gives accurate results in the evaluation process more than the evaluation system that depends on feedback only.Keywords: machine learning, adaptive, fuzzy logic, data mining
Procedia PDF Downloads 1968775 Isothermal Solid-Phase Amplification System for Detection of Yersinia pestis
Authors: Olena Mayboroda, Angel Gonzalez Benito, Jonathan Sabate Del Rio, Marketa Svobodova, Sandra Julich, Herbert Tomaso, Ciara K. O'Sullivan, Ioanis Katakis
Abstract:
DNA amplification is required for most molecular diagnostic applications but conventional PCR has disadvantages for field testing. Isothermal amplification techniques are being developed to respond to this problem. One of them is the Recombinase Polymerase Amplification (RPA) that operates at isothermal conditions without sacrificing specificity and sensitivity in easy-to-use formats. In this work RPA was used for the optical detection of solid-phase amplification of the potential biowarfare agent Yersinia pestis. Thiolated forward primers were immobilized on the surface of maleimide-activated microtitre plates for the quantitative detection of synthetic and genomic DNA, with elongation occurring only in the presence of the specific template DNA and solution phase reverse primers. Quantitative detection was achieved via the use of biotinylated reverse primers and post-amplification addition of streptavidin-HRP conjugate. The overall time of amplification and detection was less than 1 hour at a constant temperature of 37oC. Single-stranded and double-stranded DNA sequences were detected achieving detection limits of 4.04*10-13 M and 3.14*10-16 M, respectively. The system demonstrated high specificity with negligible responses to non-specific targets.Keywords: recombinase polymerase amplification, Yersinia pestis, solid-phase detection, ELONA
Procedia PDF Downloads 3038774 Monte Carlo Simulations of LSO/YSO for Dose Evaluation in Photon Beam Radiotherapy
Authors: H. Donya
Abstract:
Monte Carlo (MC) techniques play a fundamental role in radiotherapy. A two non-water-equivalent of different media were used to evaluate the dose in water. For such purpose, Lu2SiO5 (LSO) and Y2SiO5 (YSO) orthosilicates scintillators are chosen for MC simulation using Penelope code. To get higher efficiency in dose calculation, variance reduction techniques are discussed. Overall results of this investigation ensured that the LSO/YSO bi-media a good combination to tackle over-response issue in dynamic photon radiotherapy.Keywords: Lu2SiO5 (LSO) and Y2SiO5 (YSO) orthosilicates, Monte Carlo, correlated sampling, radiotherapy
Procedia PDF Downloads 4078773 Requirement Engineering and Software Product Line Scoping Paradigm
Authors: Ahmed Mateen, Zhu Qingsheng, Faisal Shahzad
Abstract:
Requirement Engineering (RE) is a part being created for programming structure during the software development lifecycle. Software product line development is a new topic area within the domain of software engineering. It also plays important role in decision making and it is ultimately helpful in rising business environment for productive programming headway. Decisions are central to engineering processes and they hold them together. It is argued that better decisions will lead to better engineering. To achieve better decisions requires that they are understood in detail. In order to address the issues, companies are moving towards Software Product Line Engineering (SPLE) which helps in providing large varieties of products with minimum development effort and cost. This paper proposed a new framework for software product line and compared with other models. The results can help to understand the needs in SPL testing, by identifying points that still require additional investigation. In our future scenario, we will combine this model in a controlled environment with industrial SPL projects which will be the new horizon for SPL process management testing strategies.Keywords: requirements engineering, software product lines, scoping, process structure, domain specific language
Procedia PDF Downloads 2258772 Optimization Techniques of Doubly-Fed Induction Generator Controller Design for Reliability Enhancement of Wind Energy Conversion Systems
Authors: Om Prakash Bharti, Aanchal Verma, R. K. Saket
Abstract:
The Doubly-Fed Induction Generator (DFIG) is suggested for Wind Energy Conversion System (WECS) to extract wind power. DFIG is preferably employed due to its robustness towards variable wind and rotor speed. DFIG has the adaptable property because the system parameters are smoothly dealt with, including real power, reactive power, DC-link voltage, and the transient and dynamic responses, which are needed to analyze constantly. The analysis becomes more prominent during any unusual condition in the electrical power system. Hence, the study and improvement in the system parameters and transient response performance of DFIG are required to be accomplished using some controlling techniques. For fulfilling the task, the present work implements and compares the optimization methods for the design of the DFIG controller for WECS. The bio-inspired optimization techniques are applied to get the optimal controller design parameters for DFIG-based WECS. The optimized DFIG controllers are then used to retrieve the transient response performance of the six-order DFIG model with a step input. The results using MATLAB/Simulink show the betterment of the Firefly algorithm (FFA) over other control techniques when compared with the other controller design methods.Keywords: doubly-fed induction generator, wind turbine, wind energy conversion system, induction generator, transfer function, proportional, integral, derivatives
Procedia PDF Downloads 938771 Testing Psychopathy as a Unified Theory of Crime and the Psychometric properties of the Youth Psychopathic Traits Inventory - Short Version among South African Youth
Authors: Leon Holtzhausen, Emma Campbell
Abstract:
This study aimed to explore the psychometric properties of the Youth Psychopathic Traits Inventory- short version (YPI-S) and the applicability of Psychopathy as a Unified Theory of Crime among 213 young adults in South Africa. The deviant behaviour variety scale and the YPI-S were used in this study. Results from factor analysis and reliability measures indicated the YPI-S seemed to have good psychometric properties when applied to the South African sample, however applicability of the behavioural dimension was a challenge. The results related to the association between deviant behaviours and psychopathic traits suggested that Psychopathy as a Unified Theory of Crime could be applied in the South African context. It is however important to note that future research should explore how the relevant scales could be culturally and contextually adapted for better psychometric outcomes.Keywords: testing psychopathy, adverse childhood experiences, youth psychopathic traits inventory, young adults
Procedia PDF Downloads 708770 A Parallel Cellular Automaton Model of Tumor Growth for Multicore and GPU Programming
Authors: Manuel I. Capel, Antonio Tomeu, Alberto Salguero
Abstract:
Tumor growth from a transformed cancer-cell up to a clinically apparent mass spans through a range of spatial and temporal magnitudes. Through computer simulations, Cellular Automata (CA) can accurately describe the complexity of the development of tumors. Tumor development prognosis can now be made -without making patients undergo through annoying medical examinations or painful invasive procedures- if we develop appropriate CA-based software tools. In silico testing mainly refers to Computational Biology research studies of application to clinical actions in Medicine. To establish sound computer-based models of cellular behavior, certainly reduces costs and saves precious time with respect to carrying out experiments in vitro at labs or in vivo with living cells and organisms. These aim to produce scientifically relevant results compared to traditional in vitro testing, which is slow, expensive, and does not generally have acceptable reproducibility under the same conditions. For speeding up computer simulations of cellular models, specific literature shows recent proposals based on the CA approach that include advanced techniques, such the clever use of supporting efficient data structures when modeling with deterministic stochastic cellular automata. Multiparadigm and multiscale simulation of tumor dynamics is just beginning to be developed by the concerned research community. The use of stochastic cellular automata (SCA), whose parallel programming implementations are open to yield a high computational performance, are of much interest to be explored up to their computational limits. There have been some approaches based on optimizations to advance in multiparadigm models of tumor growth, which mainly pursuit to improve performance of these models through efficient memory accesses guarantee, or considering the dynamic evolution of the memory space (grids, trees,…) that holds crucial data in simulations. In our opinion, the different optimizations mentioned above are not decisive enough to achieve the high performance computing power that cell-behavior simulation programs actually need. The possibility of using multicore and GPU parallelism as a promising multiplatform and framework to develop new programming techniques to speed-up the computation time of simulations is just starting to be explored in the few last years. This paper presents a model that incorporates parallel processing, identifying the synchronization necessary for speeding up tumor growth simulations implemented in Java and C++ programming environments. The speed up improvement that specific parallel syntactic constructs, such as executors (thread pools) in Java, are studied. The new tumor growth parallel model is proved using implementations with Java and C++ languages on two different platforms: chipset Intel core i-X and a HPC cluster of processors at our university. The parallelization of Polesczuk and Enderling model (normally used by researchers in mathematical oncology) proposed here is analyzed with respect to performance gain. We intend to apply the model and overall parallelization technique presented here to solid tumors of specific affiliation such as prostate, breast, or colon. Our final objective is to set up a multiparadigm model capable of modelling angiogenesis, or the growth inhibition induced by chemotaxis, as well as the effect of therapies based on the presence of cytotoxic/cytostatic drugs.Keywords: cellular automaton, tumor growth model, simulation, multicore and manycore programming, parallel programming, high performance computing, speed up
Procedia PDF Downloads 2448769 The Role of Sustainable Development in the Design and Planning of Smart Cities Using GIS Techniques: Models of Arab Cities
Authors: Ahmed M. Jihad
Abstract:
The paper presents the concept of sustainable development, and the role of geographic techniques in the design, planning and presentation of maps of smart cities with geographical vision, and the identification of programs and tools, and models of maps of Arab cities, is the problem of research in how to apply, process and experience these programs? What is the role of geographic techniques in planning and mapping the optimal place for these cities? The paper proposes an addition to the designs of Iraqi cities, as it can be developed in the future to serve as a model for interactive smart cities by developing its services. The importance of this paper stems from the concept of sustainable development dynamic which has become a method of development imposed by the present era in rapid development to achieve social balance and specialized programs in draw paper argues that ensuring sustainable development is achieved through the use of information technology. The paper will follow the theoretical presentation of the importance of the concept of development, design tools and programs. The paper follows the method of analysis of modern systems (System Analysis Approach) through the latest programs will provide results can be said that the new Iraqi cities can be developed with smart technologies, like some of the Arab and European cities that were newly created through the introduction of international investment, and therefore Plans can be made to select the best programs in manufacturing and producing maps and smart cities in the future.Keywords: geographic techniques, planning the cities, smart cities, sustainable development
Procedia PDF Downloads 2108768 The Application of Nuclear Energy for Sustainable Agriculture and Food Security: A Review
Authors: Gholamreza Farrokhi, Behzad Sani
Abstract:
The goals of sustainable agricultural are development, improved nutrition, and food security. Sustainable agriculture must be developed that will meet today’s needs for food and other products, as well as preserving the vital natural resource base that will allow future generations to meet their needs. Sustainable development requires international cooperation and the effective use of technology. Access to sustainable sources of food will remain a preeminent challenge in the decades to come. Based upon current practice and consumption, agricultural production will have to increase by about 70% by 2050 to meet demand. Nuclear techniques are used in developing countries to increase production sustainably by breeding improved crops, enhancing livestock reproduction and nutrition, as well as controlling animal and plant pests and diseases. Post-harvest losses can be reduced and safety increased with nuclear technology. Soil can be evaluated with nuclear techniques to conserve and improve soil productivity and water management.Keywords: food safety, food security, nuclear techniques, sustainable agriculture, sustainable future
Procedia PDF Downloads 3578767 Experimental Modal Analysis of Reinforced Concrete Square Slabs
Authors: M. S. Ahmed, F. A. Mohammad
Abstract:
The aim of this paper is to perform experimental modal analysis (EMA) of reinforced concrete (RC) square slabs. EMA is the process of determining the modal parameters (Natural Frequencies, damping factors, modal vectors) of a structure from a set of frequency response functions FRFs (curve fitting). Although experimental modal analysis (or modal testing) has grown steadily in popularity since the advent of the digital FFT spectrum analyzer in the early 1970’s, studying all members and materials using such method have not yet been well documented. Therefore, in this work, experimental tests were conducted on RC square specimens (0.6m x 0.6m with 40 mm). Experimental analysis is based on freely supported boundary condition. Moreover, impact testing as a fast and economical means of finding the modes of vibration of a structure was used during the experiments. In addition, Pico Scope 6 device and MATLAB software were used to acquire data, analyze and plot Frequency Response Function (FRF). The experimental natural frequencies which were extracted from measurements exhibit good agreement with analytical predictions. It is showed that EMA method can be usefully employed to perform the dynamic behavior of RC slabs.Keywords: natural frequencies, mode shapes, modal analysis, RC slabs
Procedia PDF Downloads 4088766 Effective Supply Chain Coordination with Hybrid Demand Forecasting Techniques
Authors: Gurmail Singh
Abstract:
Effective supply chain is the main priority of every organization which is the outcome of strategic corporate investments with deliberate management action. Value-driven supply chain is defined through development, procurement and by configuring the appropriate resources, metrics and processes. However, responsiveness of the supply chain can be improved by proper coordination. So the Bullwhip effect (BWE) and Net stock amplification (NSAmp) values were anticipated and used for the control of inventory in organizations by both discrete wavelet transform-Artificial neural network (DWT-ANN) and Adaptive Network-based fuzzy inference system (ANFIS). This work presents a comparative methodology of forecasting for the customers demand which is non linear in nature for a multilevel supply chain structure using hybrid techniques such as Artificial intelligence techniques including Artificial neural networks (ANN) and Adaptive Network-based fuzzy inference system (ANFIS) and Discrete wavelet theory (DWT). The productiveness of these forecasting models are shown by computing the data from real world problems for Bullwhip effect and Net stock amplification. The results showed that these parameters were comparatively less in case of discrete wavelet transform-Artificial neural network (DWT-ANN) model and using Adaptive network-based fuzzy inference system (ANFIS).Keywords: bullwhip effect, hybrid techniques, net stock amplification, supply chain flexibility
Procedia PDF Downloads 1278765 A Slip Transmission through Alpha/Beta Boundaries in a Titanium Alloy (Ti-6Al-4V)
Authors: Rayan B. M. Ameen, Ian P. Jones, Yu Lung Chiu
Abstract:
Single alpha-beta colony micro-pillars have been manufactured from a polycrystalline commercial Ti-6Al-4V sample using Focused Ion Beam (FIB). Each pillar contained two alpha lamellae separated by a thin fillet of beta phase. A nano-indenter was then used to conduct uniaxial micro-compression tests on Ti alloy single crystals, using a diamond flat tip as a compression platen. By controlling the crystal orientation along the micro-pillar using Electron back scattering diffraction (EBSD) different slip systems have been selectively activated. The advantage of the micro-compression method over conventional mechanical testing techniques is the ability to localize a single crystal volume which is characterizable after deformation. By matching the stress-strain relations resulting from micro-compression experiments to TEM (Transmission Electron Microscopy) studies of slip transmission mechanisms through the α-β interfaces, some proper constitutive material parameters such as the role of these interfaces in determining yield, strain-hardening behaviour, initial dislocation density and the critical resolved shear stress are suggested.Keywords: α/β-Ti alloy, focused ion beam, micro-mechanical test, nano-indentation, transmission electron diffraction, plastic flow
Procedia PDF Downloads 3858764 Studies on Lucrative Process Layout for Medium Scale Industries
Authors: Balamurugan Baladhandapani, Ganesh Renganathan, V. R. Sanal Kumar
Abstract:
In this paper a comprehensive review on various factory layouts has been carried out for designing a lucrative process layout for medium scale industries. Industry data base reveals that the end product rejection rate is on the order of 10% amounting large profit loss. In order to avoid these rejection rates and to increase the quality product production an intermediate non-destructive testing facility (INDTF) has been recommended for increasing the overall profit. We observed through detailed case studies that while introducing INDTF to medium scale industries the expensive production process can be avoided to the defective products well before its final shape. Additionally, the defective products identified during the intermediate stage can be effectively utilized for other applications or recycling; thereby the overall wastage of the raw materials can be reduced and profit can be increased. We concluded that the prudent design of a factory layout through critical path method facilitating with INDTF will warrant profitable outcome.Keywords: intermediate non-destructive testing, medium scale industries, process layout design
Procedia PDF Downloads 502