Search results for: Course Delivery Methods.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4380

Search results for: Course Delivery Methods.

2730 Utilization of Laser-Ablation Based Analytical Methods for Obtaining Complete Chemical Information of Algae

Authors: Pavel Pořízka, David Prochazka, Karel Novotný, Ota Samek, ZdeněkPilát, Klára Procházková, and Jozef Kaiser

Abstract:

Themain goal of this article is to find efficient methods for elemental and molecular analysis of living microorganisms (algae) under defined environmental conditions and cultivation processes. The overall knowledge of chemical composition is obtained utilizing laser-based techniques, Laser- Induced Breakdown Spectroscopy (LIBS) for acquiring information about elemental composition and Raman Spectroscopy for gaining molecular information, respectively. Algal cells were suspended in liquid media and characterized using their spectra. Results obtained employing LIBS and Raman Spectroscopy techniques will help to elucidate algae biology (nutrition dynamics depending on cultivation conditions) and to identify algal strains, which have the potential for applications in metal-ion absorption (bioremediation) and biofuel industry. Moreover, bioremediation can be readily combined with production of 3rd generation biofuels. In order to use algae for efficient fuel production, the optimal cultivation parameters have to be determinedleading to high production of oil in selected cellswithout significant inhibition of the photosynthetic activity and the culture growth rate, e.g. it is necessary to distinguish conditions for algal strain containing high amount of higher unsaturated fatty acids. Measurements employing LIBS and Raman Spectroscopy were utilized in order to give information about alga Trachydiscusminutus with emphasis on the amount of the lipid content inside the algal cell and the ability of algae to withdraw nutrients from its environment and bioremediation (elemental composition), respectively. This article can serve as the reference for further efforts in describing complete chemical composition of algal samples employing laserablation techniques.

Keywords: Laser-Induced Breakdown Spectroscopy, Raman Spectroscopy, Algae, Algal strains, Bioremediation, Biofuels.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2250
2729 Bee Parameter Determination via Weighted Centriod Modified Simplex and Constrained Response Surface Optimisation Methods

Authors: P. Luangpaiboon

Abstract:

Various intelligences and inspirations have been adopted into the iterative searching process called as meta-heuristics. They intelligently perform the exploration and exploitation in the solution domain space aiming to efficiently seek near optimal solutions. In this work, the bee algorithm, inspired by the natural foraging behaviour of honey bees, was adapted to find the near optimal solutions of the transportation management system, dynamic multi-zone dispatching. This problem prepares for an uncertainty and changing customers- demand. In striving to remain competitive, transportation system should therefore be flexible in order to cope with the changes of customers- demand in terms of in-bound and outbound goods and technological innovations. To remain higher service level but lower cost management via the minimal imbalance scenario, the rearrangement penalty of the area, in each zone, including time periods are also included. However, the performance of the algorithm depends on the appropriate parameters- setting and need to be determined and analysed before its implementation. BEE parameters are determined through the linear constrained response surface optimisation or LCRSOM and weighted centroid modified simplex methods or WCMSM. Experimental results were analysed in terms of best solutions found so far, mean and standard deviation on the imbalance values including the convergence of the solutions obtained. It was found that the results obtained from the LCRSOM were better than those using the WCMSM. However, the average execution time of experimental run using the LCRSOM was longer than those using the WCMSM. Finally a recommendation of proper level settings of BEE parameters for some selected problem sizes is given as a guideline for future applications.

Keywords: Meta-heuristic, Bee Algorithm, Dynamic Multi-Zone Dispatching, Linear Constrained Response SurfaceOptimisation Method, Weighted Centroid Modified Simplex Method

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1373
2728 Assessment of Path Loss Prediction Models for Wireless Propagation Channels at L-Band Frequency over Different Micro-Cellular Environments of Ekiti State, Southwestern Nigeria

Authors: C. I. Abiodun, S. O. Azi, J. S. Ojo, P. Akinyemi

Abstract:

The design of accurate and reliable mobile communication systems depends majorly on the suitability of path loss prediction methods and the adaptability of the methods to various environments of interest. In this research, the results of the adaptability of radio channel behavior are presented based on practical measurements carried out in the 1800 MHz frequency band. The measurements are carried out in typical urban, suburban and rural environments in Ekiti State, Southwestern part of Nigeria. A total number of seven base stations of MTN GSM service located in the studied environments were monitored. Path loss and break point distances were deduced from the measured received signal strength (RSS) and a practical path loss model is proposed based on the deduced break point distances. The proposed two slope model, regression line and four existing path loss models were compared with the measured path loss values. The standard deviations of each model with respect to the measured path loss were estimated for each base station. The proposed model and regression line exhibited lowest standard deviations followed by the Cost231-Hata model when compared with the Erceg Ericsson and SUI models. Generally, the proposed two-slope model shows closest agreement with the measured values with a mean error values of 2 to 6 dB. These results show that, either the proposed two slope model or Cost 231-Hata model may be used to predict path loss values in mobile micro cell coverage in the well-considered environments. Information from this work will be useful for link design of microwave band wireless access systems in the region.

Keywords: Break-point distances, path loss models, path loss exponent, received signal strength.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 819
2727 Mitigation of Radiation Levels for Base Transceiver Stations based on ITU-T Recommendation K.70

Authors: Reyes C., Ramos B.

Abstract:

This essay presents applicative methods to reduce human exposure levels in the area around base transceiver stations in a environment with multiple sources based on ITU-T recommendation K.70. An example is presented to understand the mitigation techniques and their results and also to learn how they can be applied, especially in developing countries where there is not much research on non-ionizing radiations.

Keywords: Electromagnetic fields (EMF), human exposure limits, intentional radiator, cumulative exposure ratio, base transceiver station (BTS), radiation levels.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2702
2726 Modeling Stress-Induced Regulatory Cascades with Artificial Neural Networks

Authors: Maria E. Manioudaki, Panayiota Poirazi

Abstract:

Yeast cells live in a constantly changing environment that requires the continuous adaptation of their genomic program in order to sustain their homeostasis, survive and proliferate. Due to the advancement of high throughput technologies, there is currently a large amount of data such as gene expression, gene deletion and protein-protein interactions for S. Cerevisiae under various environmental conditions. Mining these datasets requires efficient computational methods capable of integrating different types of data, identifying inter-relations between different components and inferring functional groups or 'modules' that shape intracellular processes. This study uses computational methods to delineate some of the mechanisms used by yeast cells to respond to environmental changes. The GRAM algorithm is first used to integrate gene expression data and ChIP-chip data in order to find modules of coexpressed and co-regulated genes as well as the transcription factors (TFs) that regulate these modules. Since transcription factors are themselves transcriptionally regulated, a three-layer regulatory cascade consisting of the TF-regulators, the TFs and the regulated modules is subsequently considered. This three-layer cascade is then modeled quantitatively using artificial neural networks (ANNs) where the input layer corresponds to the expression of the up-stream transcription factors (TF-regulators) and the output layer corresponds to the expression of genes within each module. This work shows that (a) the expression of at least 33 genes over time and for different stress conditions is well predicted by the expression of the top layer transcription factors, including cases in which the effect of up-stream regulators is shifted in time and (b) identifies at least 6 novel regulatory interactions that were not previously associated with stress-induced changes in gene expression. These findings suggest that the combination of gene expression and protein-DNA interaction data with artificial neural networks can successfully model biological pathways and capture quantitative dependencies between distant regulators and downstream genes.

Keywords: gene modules, artificial neural networks, yeast, stress

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1465
2725 Comparison of Data Reduction Algorithms for Image-Based Point Cloud Derived Digital Terrain Models

Authors: M. Uysal, M. Yilmaz, I. Tiryakioğlu

Abstract:

Digital Terrain Model (DTM) is a digital numerical representation of the Earth's surface. DTMs have been applied to a diverse field of tasks, such as urban planning, military, glacier mapping, disaster management. In the expression of the Earth' surface as a mathematical model, an infinite number of point measurements are needed. Because of the impossibility of this case, the points at regular intervals are measured to characterize the Earth's surface and DTM of the Earth is generated. Hitherto, the classical measurement techniques and photogrammetry method have widespread use in the construction of DTM. At present, RADAR, LiDAR, and stereo satellite images are also used for the construction of DTM. In recent years, especially because of its superiorities, Airborne Light Detection and Ranging (LiDAR) has an increased use in DTM applications. A 3D point cloud is created with LiDAR technology by obtaining numerous point data. However recently, by the development in image mapping methods, the use of unmanned aerial vehicles (UAV) for photogrammetric data acquisition has increased DTM generation from image-based point cloud. The accuracy of the DTM depends on various factors such as data collection method, the distribution of elevation points, the point density, properties of the surface and interpolation methods. In this study, the random data reduction method is compared for DTMs generated from image based point cloud data. The original image based point cloud data set (100%) is reduced to a series of subsets by using random algorithm, representing the 75, 50, 25 and 5% of the original image based point cloud data set. Over the ANS campus of Afyon Kocatepe University as the test area, DTM constructed from the original image based point cloud data set is compared with DTMs interpolated from reduced data sets by Kriging interpolation method. The results show that the random data reduction method can be used to reduce the image based point cloud datasets to 50% density level while still maintaining the quality of DTM.

Keywords: DTM, unmanned aerial vehicle, UAV, random, Kriging.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 810
2724 A New Derivative-Free Quasi-Secant Algorithm For Solving Non-Linear Equations

Authors: F. Soleymani, M. Sharifi

Abstract:

Most of the nonlinear equation solvers do not converge always or they use the derivatives of the function to approximate the root of such equations. Here, we give a derivative-free algorithm that guarantees the convergence. The proposed two-step method, which is to some extent like the secant method, is accompanied with some numerical examples. The illustrative instances manifest that the rate of convergence in proposed algorithm is more than the quadratically iterative schemes.

Keywords: Non-linear equation, iterative methods, derivative-free, convergence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1777
2723 Body Composition Analysis of University Students by Anthropometry and Bioelectrical Impedance Analysis

Authors: Vinti Davar

Abstract:

Background: Worldwide, at least 2.8 million people die each year as a result of being overweight or obese, and 35.8 million (2.3%) of global DALYs are caused by overweight or obesity. Obesity is acknowledged as one of the burning public health problems reducing life expectancy and quality of life. The body composition analysis of the university population is essential in assessing the nutritional status, as well as the risk of developing diseases associated with abnormal body fat content so as to make nutritional recommendations. Objectives: The main aim was to determine the prevalence of obesity and overweight in University students using Anthropometric analysis and BIA methods. Material and Methods: In this cross-sectional study, 283 university students participated. The body composition analysis was undertaken by using mainly: i) Anthropometric Measurement: Height, Weight, BMI, waist circumference, hip circumference and skin fold thickness, ii) Bio-electrical impedance was used for analysis of body fat mass, fat percent and visceral fat which was measured by Tanita SC-330P Professional Body Composition Analyzer. The data so collected were compiled in MS Excel and analyzed for males and females using SPSS 16. Results and Discussion: The mean age of the male (n= 153) studied subjects was 25.37 ±2.39 years and females (n=130) was 22.53 ±2.31. The data of BIA revealed very high mean fat per cent of the female subjects i.e. 30.3±6.5 per cent whereas mean fat per cent of the male subjects was 15.60±6.02 per cent indicating a normal body fat range. The findings showed high visceral fat of both males (12.92±3.02) and females (16.86±4.98). BMI, BF% and WHR were higher among females, and BMI was higher among males. The most evident correlation was verified between BF% and WHR for female students (r=0.902; p<0.001). The correlation of BFM and BF% with thickness of triceps, sub scapular and abdominal skin folds and BMI was significant (P<0.001). Conclusion: The studied data made it obvious that there is a need to initiate lifestyle changing strategies especially for adult females and encourage them to improve their dietary intake to prevent incidence of noncommunicable diseases due to obesity and high fat percentage.

Keywords: Anthropometry, bioelectrical impedance, body fat percentage, obesity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2834
2722 Analytical Solutions of Kortweg-de Vries(KdV) Equation

Authors: Foad Saadi, M. Jalali Azizpour, S.A. Zahedi

Abstract:

The objective of this paper is to present a comparative study of Homotopy Perturbation Method (HPM), Variational Iteration Method (VIM) and Homotopy Analysis Method (HAM) for the semi analytical solution of Kortweg-de Vries (KdV) type equation called KdV. The study have been highlighted the efficiency and capability of aforementioned methods in solving these nonlinear problems which has been arisen from a number of important physical phenomenon.

Keywords: Variational Iteration Method (VIM), HomotopyPerturbation Method (HPM), Homotopy Analysis Method (HAM), KdV Equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2375
2721 Person Identification using Gait by Combined Features of Width and Shape of the Binary Silhouette

Authors: M.K. Bhuyan, Aragala Jagan.

Abstract:

Current image-based individual human recognition methods, such as fingerprints, face, or iris biometric modalities generally require a cooperative subject, views from certain aspects, and physical contact or close proximity. These methods cannot reliably recognize non-cooperating individuals at a distance in the real world under changing environmental conditions. Gait, which concerns recognizing individuals by the way they walk, is a relatively new biometric without these disadvantages. The inherent gait characteristic of an individual makes it irreplaceable and useful in visual surveillance. In this paper, an efficient gait recognition system for human identification by extracting two features namely width vector of the binary silhouette and the MPEG-7-based region-based shape descriptors is proposed. In the proposed method, foreground objects i.e., human and other moving objects are extracted by estimating background information by a Gaussian Mixture Model (GMM) and subsequently, median filtering operation is performed for removing noises in the background subtracted image. A moving target classification algorithm is used to separate human being (i.e., pedestrian) from other foreground objects (viz., vehicles). Shape and boundary information is used in the moving target classification algorithm. Subsequently, width vector of the outer contour of binary silhouette and the MPEG-7 Angular Radial Transform coefficients are taken as the feature vector. Next, the Principal Component Analysis (PCA) is applied to the selected feature vector to reduce its dimensionality. These extracted feature vectors are used to train an Hidden Markov Model (HMM) for identification of some individuals. The proposed system is evaluated using some gait sequences and the experimental results show the efficacy of the proposed algorithm.

Keywords: Gait Recognition, Gaussian Mixture Model, PrincipalComponent Analysis, MPEG-7 Angular Radial Transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1911
2720 Analysis of Complexes Pairing Performat Radical and Water

Authors: Sanaz Gharehzadeh Shirazi, Subira Gharehzadeh Shirazi, Fariba Jafari

Abstract:

The present article comprises a theoretical study of structures Performat radical (HCO3) with H2O molecule. We make use of ab initio quantum chemical methods. Unrestricted Hartee-Fock (UHF) with the basis set6-311+g(2df,2p) and density functional theory (B3LYP) with the basis set 6-311+g(2df,2p) and also we done atoms in molecules (AIM) theory for them. We have found four stable geometries the PerformatRadical(HCO3) with H2O.

Keywords: Hydrogen binding, Performat Radical, Water, Gaussian, Atoms in molecules (AIM) theory

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1445
2719 Antimicrobial and Aroma Finishing of Organic Cotton Knits Using Vetiver Oil Microcapsules for Health Care Textiles

Authors: K. J. Sannapapamma, H. Malligawad Lokanath, Sakeena Naikwadi

Abstract:

Eco-friendly textiles are gaining importance among the consumers and textile manufacturers in the healthcare sector due to increased environmental pollution which leads to several health and environmental hazards. Hence, the research was designed to cultivate and develop the organic cotton knit, to prepare and characterize the Vetiver oil microcapsules for textile finishing and to access the wash durability of finished knits. The cotton SAHANA variety grown under organic production systems was processed and spun into 30 single yarn dyed with four natural colorants (Arecanut slurry, Eucalyptus leaves, Pomegranate rind and Indigo) and eco dyed yarn was further used for development of single jersy knitted fabric. Vetiveria zizanioides is an aromatic grass which is being traditionally used in medicine and perfumery. Vetiver essential oil was used for preparation of microcapsules by interfacial polymerization technique subjected to Gas Chromatography Mass Spectrometry (GCMS), Fourier Transform Infrared Spectroscopy (FTIR), Thermo Gravimetric Analyzer (TGA) and Scanning Electron Microscope (SEM) for characterization of microcapsules. The knitted fabric was finished with vetiver oil microcapsules by exhaust and pad dry cure methods. The finished organic knit was assessed for laundering on antimicrobial efficiency and aroma intensity. GCMS spectral analysis showed that, diethyl phthalate (28%) was the major compound found in vetiver oil followed by isoaromadendrene epoxide (7.72%), beta-vetivenene (6.92%), solavetivone (5.58%), aromadenderene, azulene and khusimol. Bioassay explained that, the vetiver oil and diluted vetiver oil possessed greater zone of inhibition against S. aureus and E. coli than the coconut oil. FTRI spectra of vetiver oil and microcapsules possessed similar peaks viz., C-H, C=C & C꞊O stretching and additionally oil microcapsules possessed the peak of 3331.24 cm-1 at 91.14 transmittance was attributed to N-H stretches. TGA of oil microcapsules revealed that, there was a minimum weight loss (5.835%) recorded at 467.09°C compared to vetiver oil i.e., -3.026% at the temperature of 396.24°C. The shape of the microcapsules was regular and round, some were spherical in shape and few were rounded by small aggregates. Irrespective of methods of application, organic cotton knits finished with microcapsules by pad dry cure method showed maximum zone of inhibition compared to knits finished by exhaust method against S. aureus and E. coli. The antimicrobial activity of the finished samples was subjected to multiple washing which indicated that knits finished with pad dry cure method showed a zone of inhibition even after 20th wash and better aroma retention compared to knits finished with the exhaust method of application. Further, the group of respondents rated that the 5th washed samples had the greater aroma intensity in both the methods than the other samples. Thus, the vetiver microencapsulated organic cotton knits are free from hazardous chemicals and have multi-functional properties that can be suitable for medical and healthcare textiles.

Keywords: Exhaust and pad dry cure finishing, interfacial polymerization, organic cotton knits, vetiver oil microcapsules.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1829
2718 Comparison of Process Slaughtered on Beef Cattle Based on Level of Cortisol and Fourier Transform Infrared Spectroscopy (FTIR)

Authors: Pudji Astuti, C. P. C. Putro, C. M. Airin, L. Sjahfirdi, S. Widiyanto, H. Maheshwari

Abstract:

Stress of slaughter animals starting long before until at the time of process of slaughtering which cause misery and decrease of meat quality. Meanwhile, determination of animal stress using hormonal such as cortisol is expensive and less practical so that portable stress indicator for cows based on Fourier Transform Infrared Spectroscopy (FTIR) must be provided. The aims of this research are to find out the comparison process of slaughter between Rope Casting Local (RCL) and Restraining Box Method (RBM) by measuring of cortisol and wavelength in FTIR methods. Thirty two of male Ongole crossbred cattle were used in this experiment. Blood sampling was taken from jugular vein when they were rested and repeated when slaughtered. All of blood samples were centrifuged at 3000 rpm for 20 minutes to get serum, and then divided into two parts for cortisol assayed using ELISA and for measuring the wavelength using FTIR. The serum then measured at the wavelength between 4000-400 cm-1 using MB3000 FTIR. Band data absorption in wavelength of FTIR is analyzed descriptively by using FTIR Horizon MBTM. For RCL, average of serum cortisol when the animals rested were 11.47 ± 4.88 ng/mL, when the time of slaughter were 23.27 ± 7.84 ng/mL. For RBM, level of cortisol when rested animals were 13.67 ± 3.41 ng/mL and 53.47 ± 20.25 ng/mL during the slaughter. Based on student t-Test, there were significantly different between RBM and RCL methods when beef cattle were slaughtered (P<0.05), but no significantly different when animals were rested (P>0.05). Result of FTIR with the various of wavelength such as methyl group (=CH3 ) 2986cm-1, methylene (=CH2 ) 2827 cm-1, hydroxyl (- OH) 3371 cm-1, carbonyl (ketones) (C=O) 1636 cm-1, carboxyl (COO-1) 1408 cm-1, glucosa 1057 cm-1, urea 1011 cm-1have been obtained. It can be concluded that the RCL slaughtered method is better than the RBM method based on the increase of cortisol as an indicator of stress in beef cattle (P<0.05). FTIR is really possible to be used as stub of stress tool due to differentiate of resting and slaughter condition by recognizing the increase of absorption and the separation of component group at the wavelength.  

Keywords: Cows, cortisol, FTIR, RBM, RCL, stress indicator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2463
2717 Holistic Approach to Teaching Mathematics in Secondary School as a Means of Improving Students’ Comprehension of Study Material

Authors: Natalia Podkhodova, Olga Sheremeteva, Mariia Soldaeva

Abstract:

Creating favourable conditions for students’ comprehension of mathematical content is one of the primary problems in teaching mathematics in secondary school. The fact of comprehension includes the ability to build a working situational model and thus becomes an important means of solving mathematical problems. This paper describes a holistic approach to teaching mathematics designed to address the primary challenges of such teaching; specifically, the challenge of students’ comprehension. Essentially, this approach consists of (1) establishing links between the attributes of the notion: the sense, the meaning, and the term; (2) taking into account the components of student’s subjective experience—value-based emotions, contextual, procedural and communicative—during the educational process; (3) linking together different ways to present mathematical information; (4) identifying and leveraging the relationships between real, perceptual and conceptual (scientific) mathematical spaces by applying real-life situational modelling. The article describes approaches to the practical use of these foundational concepts. Identifying how proposed methods and techniques influence understanding of material used in teaching mathematics was the primary goal. The study included an experiment in which 256 secondary school students took part: 142 in the study group and 114 in the control group. All students in these groups had similar levels of achievement in math and studied math under the same curriculum. In the course of the experiment, comprehension of two topics — “Derivative” and “Trigonometric functions”—was evaluated. Control group participants were taught using traditional methods. Students in the study group were taught using the holistic method: under teacher’s guidance, they carried out assignments designed to establish linkages between notion’s characteristics, to convert information from one mode of presentation to another, as well as assignments that required the ability to operate with all modes of presentation. Identification, accounting for and transformation of subjective experience were associated with methods of stimulating the emotional value component of the studied mathematical content (discussions of lesson titles, assignments aimed to create study dominants, performing theme-related physical exercise ...) The use of techniques that forms inter-subject notions based on linkages between, perceptual real and mathematical conceptual spaces proved to be of special interest to the students. Results of the experiment were analysed by presenting students in each of the groups with a final test in each of the studied topics. The test included assignments that required building real situational models. Statistical analysis was used to aggregate test results. Pierson criterion x2 was used to reveal statistics significance of results (pass-fail the modelling test). Significant difference of results was revealed (p < 0.001), which allowed to conclude that students in the study group showed better comprehension of mathematical information than those in the control group. The total number of completed assignments of each student was analysed as well, with average results calculated for each group. Statistical significance of result differences against the quantitative criterion (number of completed assignments) was determined using Student’s t-test, which showed that students in the study group completed significantly more assignments than those in the control group (p = 0.0001). Authors thus come to the conclusion that suggested increase in the level of comprehension of study material took place as a result of applying implemented methods and techniques.

Keywords: Comprehension of mathematical content, holistic approach to teaching mathematics in secondary school, subjective experience, technology of the formation of inter-subject notions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 608
2716 Solid State Drive End to End Reliability Prediction, Characterization and Control

Authors: Mohd Azman Abdul Latif, Erwan Basiron

Abstract:

A flaw or drift from expected operational performance in one component (NAND, PMIC, controller, DRAM, etc.) may affect the reliability of the entire Solid State Drive (SSD) system. Therefore, it is important to ensure the required quality of each individual component through qualification testing specified using standards or user requirements. Qualification testing is time-consuming and comes at a substantial cost for product manufacturers. A highly technical team, from all the eminent stakeholders is embarking on reliability prediction from beginning of new product development, identify critical to reliability parameters, perform full-blown characterization to embed margin into product reliability and establish control to ensure the product reliability is sustainable in the mass production. The paper will discuss a comprehensive development framework, comprehending SSD end to end from design to assembly, in-line inspection, in-line testing and will be able to predict and to validate the product reliability at the early stage of new product development. During the design stage, the SSD will go through intense reliability margin investigation with focus on assembly process attributes, process equipment control, in-process metrology and also comprehending forward looking product roadmap. Once these pillars are completed, the next step is to perform process characterization and build up reliability prediction modeling. Next, for the design validation process, the reliability prediction specifically solder joint simulator will be established. The SSD will be stratified into Non-Operating and Operating tests with focus on solder joint reliability and connectivity/component latent failures by prevention through design intervention and containment through Temperature Cycle Test (TCT). Some of the SSDs will be subjected to the physical solder joint analysis called Dye and Pry (DP) and Cross Section analysis. The result will be feedbacked to the simulation team for any corrective actions required to further improve the design. Once the SSD is validated and is proven working, it will be subjected to implementation of the monitor phase whereby Design for Assembly (DFA) rules will be updated. At this stage, the design change, process and equipment parameters are in control. Predictable product reliability at early product development will enable on-time sample qualification delivery to customer and will optimize product development validation, effective development resource and will avoid forced late investment to bandage the end-of-life product failures. Understanding the critical to reliability parameters earlier will allow focus on increasing the product margin that will increase customer confidence to product reliability.

Keywords: e2e reliability prediction, SSD, TCT, Solder Joint Reliability, NUDD, connectivity issues, qualifications, characterization and control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 400
2715 2D Graphical Analysis of Wastewater Influent Capacity Time Series

Authors: Monika Chuchro, Maciej Dwornik

Abstract:

The extraction of meaningful information from image could be an alternative method for time series analysis. In this paper, we propose a graphical analysis of time series grouped into table with adjusted colour scale for numerical values. The advantages of this method are also discussed. The proposed method is easy to understand and is flexible to implement the standard methods of pattern recognition and verification, especially for noisy environmental data.

Keywords: graphical analysis, time series, seasonality, noisy environmental data

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1450
2714 Spin Coherent State Path Integral for the Interaction of Two-Level System with Time Dependent Non-Uniform Magnetic Field

Authors: Rekik Rima, Aouachria Mekki

Abstract:

We study the movement of a two-level atom in interaction with time dependent nonuniform magnetic filed using the path integral formalism. The propagator is first written in the standard form by replacing the spin by a unit vector aligned along the polar and azimuthal directions. Then it is determined exactly using perturbation methods. Thus the Rabi formula of the system are deduced.

Keywords: Path integral, Formalism, Propagator, Transition probability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2020
2713 A Posterior Predictive Model-Based Control Chart for Monitoring Healthcare

Authors: Yi-Fan Lin, Peter P. Howley, Frank A. Tuyl

Abstract:

Quality measurement and reporting systems are used in healthcare internationally. In Australia, the Australian Council on Healthcare Standards records and reports hundreds of clinical indicators (CIs) nationally across the healthcare system. These CIs are measures of performance in the clinical setting, and are used as a screening tool to help assess whether a standard of care is being met. Existing analysis and reporting of these CIs incorporate Bayesian methods to address sampling variation; however, such assessments are retrospective in nature, reporting upon the previous six or twelve months of data. The use of Bayesian methods within statistical process control for monitoring systems is an important pursuit to support more timely decision-making. Our research has developed and assessed a new graphical monitoring tool, similar to a control chart, based on the beta-binomial posterior predictive (BBPP) distribution to facilitate the real-time assessment of health care organizational performance via CIs. The BBPP charts have been compared with the traditional Bernoulli CUSUM (BC) chart by simulation. The more traditional “central” and “highest posterior density” (HPD) interval approaches were each considered to define the limits, and the multiple charts were compared via in-control and out-of-control average run lengths (ARLs), assuming that the parameter representing the underlying CI rate (proportion of cases with an event of interest) required estimation. Preliminary results have identified that the BBPP chart with HPD-based control limits provides better out-of-control run length performance than the central interval-based and BC charts. Further, the BC chart’s performance may be improved by using Bayesian parameter estimation of the underlying CI rate.

Keywords: Average run length, Bernoulli CUSUM chart, beta binomial posterior predictive distribution, clinical indicator, health care organization, highest posterior density interval.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 878
2712 Some Results on Preconditioned Modified Accelerated Overrelaxation Method

Authors: Guangbin Wang, Deyu Sun, Fuping Tan

Abstract:

In this paper, we present new preconditioned modified accelerated overrelaxation (MAOR) method for solving linear systems. We compare the spectral radii of the iteration matrices of the preconditioned and the original methods. The comparison results show that the preconditioned MAOR method converges faster than the MAOR method whenever the MAOR method is convergent. Finally, we give one numerical example to confirm our theoretical results.

Keywords: preconditioned, MAOR method, linear system, convergence, comparison.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1647
2711 Mechanisms of Internet Security Attacks

Authors: J. Dubois, P. Jreije

Abstract:

Internet security attack could endanger the privacy of World Wide Web users and the integrity of their data. The attack can be carried out on today's most secure systems- browsers, including Netscape Navigator and Microsoft Internet Explorer. There are too many types, methods and mechanisms of attack where new attack techniques and exploits are constantly being developed and discovered. In this paper, various types of internet security attack mechanisms are explored and it is pointed out that when different types of attacks are combined together, network security can suffer disastrous consequences.

Keywords: DoS, internet attacks, router attack, security, trojan, virus, worm, XSS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2109
2710 Acoustic Finite Element Analysis of a Slit Model with Consideration of Air Viscosity

Authors: M. Sasajima, M. Watanabe, T. Yamaguchi Y. Kurosawa, Y. Koike

Abstract:

In very narrow pathways, the speed of sound propagation and the phase of sound waves change due to the air viscosity. We have developed a new finite element method (FEM) that includes the effects of air viscosity for modeling a narrow sound pathway. This method is developed as an extension of the existing FEM for porous sound-absorbing materials. The numerical calculation results for several three-dimensional slit models using the proposed FEM are validated against existing calculation methods.

Keywords: Simulation, FEM, air viscosity, slit.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1631
2709 Multi-Objective Cellular Manufacturing System under Machines with Different Life-Cycle using Genetic Algorithm

Authors: N. Javadian, J. Rezaeian, Y. Maali

Abstract:

In this paper a multi-objective nonlinear programming model of cellular manufacturing system is presented which minimize the intercell movements and maximize the sum of reliability of cells. We present a genetic approach for finding efficient solutions to the problem of cell formation for products having multiple routings. These methods find the non-dominated solutions and according to decision makers prefer, the best solution will be chosen.

Keywords: Cellular Manufacturing, Genetic Algorithm, Multiobjective, Life-Cycle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1946
2708 The Role of Home Composting in Waste Management Cost Reduction

Authors: Nahid Hassanshahi, Ayoub Karimi-Jashni, Nasser Talebbeydokhti

Abstract:

Due to the economic and environmental benefits of producing less waste, the US Environmental Protection Agency (EPA) introduces source reduction as one of the most important means to deal with the problems caused by increased landfills and pollution. Waste reduction involves all waste management methods, including source reduction, recycling, and composting, which reduce waste flow to landfills or other disposal facilities. Source reduction of waste can be studied from two perspectives: avoiding waste production, or reducing per capita waste production, and waste deviation that indicates the reduction of waste transfer to landfills. The present paper has investigated home composting as a managerial solution for reduction of waste transfer to landfills. Home composting has many benefits. The use of household waste for the production of compost will result in a much smaller amount of waste being sent to landfills, which in turn will reduce the costs of waste collection, transportation and burial. Reducing the volume of waste for disposal and using them for the production of compost and plant fertilizer might help to recycle the material in a shorter time and to use them effectively in order to preserve the environment and reduce contamination. Producing compost in a home-based manner requires very small piece of land for preparation and recycling compared with other methods. The final product of home-made compost is valuable and helps to grow crops and garden plants. It is also used for modifying the soil structure and maintaining its moisture. The food that is transferred to landfills will spoil and produce leachate after a while. It will also release methane and greenhouse gases. But, composting these materials at home is the best way to manage degradable materials, use them efficiently and reduce environmental pollution. Studies have shown that the benefits of the sale of produced compost and the reduced costs of collecting, transporting, and burying waste can well be responsive to the costs of purchasing home compost machine and the cost of related trainings. Moreover, the process of producing home compost may be profitable within 4 to 5 years and as a result, it will have a major role in reducing waste management.

Keywords: Compost, home compost, reducing waste, waste management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 884
2707 A Comparative Study on Different Approaches to Evaluate Ship Equilibrium Point

Authors: Alessandro A. Zizzari, Francesca Calabrese, Giovanni Indiveri, Andrea Coraddu, Diego Villa

Abstract:

The aim of this paper is to present a comparative study on two different methods for the evaluation of the equilibrium point of a ship, core issue for designing an On Board Stability System (OBSS) module that, starting from geometry information of a ship hull, described by a discrete model in a standard format, and the distribution of all weights onboard calculates the ship floating conditions (in draught, heel and trim).

Keywords: Algorithms, Computer applications, Equilibrium, Marine applications, Stability System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2168
2706 Trainer Aircraft Selection Using Preference Analysis for Reference Ideal Solution (PARIS)

Authors: C. Ardil

Abstract:

This article presents a multiple criteria evaluation for a trainer aircraft selection problem using "preference analysis for reference ideal solution (PARIS)” approach. The available relevant literature points to the use of multiple criteria decision making analysis (MCDMA) methods for the problem of trainer aircraft selection, which often involves conflicting multiple criteria. Therefore, this MCDMA study aims to propose a robust systematic integrated framework focusing on the trainer aircraft selection problem. For this purpose, an integrated preference analysis approach based the mean weight and entropy weight procedures with PARIS, and TOPSIS was used for a MCDMA compensating solution. In this study, six trainer aircraft alternatives were evaluated according to six technical decision criteria, and data were collected from the current relevant literature. As a result, the King Air C90GTi alternative was identified as the most suitable trainer aircraft alternative. In order to verify the stability and accuracy of the results obtained, comparisons were made with existing MCDMA methods during the sensitivity and validity analysis process.The results of the application were further validated by applying the comparative analysis-based PARIS, and TOPSIS method. The proposed integrated MCDMA systematic structure is also expected to address the issues encountered in the aircraft selection process. Finally, the analysis results obtained show that the proposed MCDMA method is an effective and accurate tool that can help analysts make better decisions.

Keywords: aircraft, trainer aircraft selection, multiple criteria decision making, multiple criteria decision making analysis, mean weight, entropy weight, MCDMA, PARIS, TOPSIS

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 430
2705 Assessing Overall Thermal Conductance Value of Low-Rise Residential Home Exterior Above-Grade Walls Using Infrared Thermography Methods

Authors: Matthew D. Baffa

Abstract:

Infrared thermography is a non-destructive test method used to estimate surface temperatures based on the amount of electromagnetic energy radiated by building envelope components. These surface temperatures are indicators of various qualitative building envelope deficiencies such as locations and extent of heat loss, thermal bridging, damaged or missing thermal insulation, air leakage, and moisture presence in roof, floor, and wall assemblies. Although infrared thermography is commonly used for qualitative deficiency detection in buildings, this study assesses its use as a quantitative method to estimate the overall thermal conductance value (U-value) of the exterior above-grade walls of a study home. The overall U-value of exterior above-grade walls in a home provides useful insight into the energy consumption and thermal comfort of a home. Three methodologies from the literature were employed to estimate the overall U-value by equating conductive heat loss through the exterior above-grade walls to the sum of convective and radiant heat losses of the walls. Outdoor infrared thermography field measurements of the exterior above-grade wall surface and reflective temperatures and emissivity values for various components of the exterior above-grade wall assemblies were carried out during winter months at the study home using a basic thermal imager device. The overall U-values estimated from each methodology from the literature using the recorded field measurements were compared to the nominal exterior above-grade wall overall U-value calculated from materials and dimensions detailed in architectural drawings of the study home. The nominal overall U-value was validated through calendarization and weather normalization of utility bills for the study home as well as various estimated heat loss quantities from a HOT2000 computer model of the study home and other methods. Under ideal environmental conditions, the estimated overall U-values deviated from the nominal overall U-value between ±2% to ±33%. This study suggests infrared thermography can estimate the overall U-value of exterior above-grade walls in low-rise residential homes with a fair amount of accuracy.

Keywords: Emissivity, heat loss, infrared thermography, thermal conductance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 836
2704 Deoiling Hydrocyclones Flow Field-A Comparison between k-Epsilon and LES

Authors: Maysam Saidi, Reza Maddahian, Bijan Farhanieh

Abstract:

In this research a comparison between k-epsilon and LES model for a deoiling hydrocyclone is conducted. Flow field of hydrocyclone is obtained by three-dimensional simulations with OpenFOAM code. Potential of prediction for both methods of this complex swirl flow is discussed. Large eddy simulation method results have more similarity to experiment and its results are presented in figures from different hydrocyclone cross sections.

Keywords: Deoiling hydrocyclones, k-epsilon model, Largeeddy simulation, OpenFOAM

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2525
2703 Complexity of Component-based Development of Embedded Systems

Authors: M. Zheng, V. S. Alagar

Abstract:

The paper discusses complexity of component-based development (CBD) of embedded systems. Although CBD has its merits, it must be augmented with methods to control the complexities that arise due to resource constraints, timeliness, and run-time deployment of components in embedded system development. Software component specification, system-level testing, and run-time reliability measurement are some ways to control the complexity.

Keywords: Components, embedded systems, complexity, softwaredevelopment, traffic controller system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1499
2702 Application of Four-electrode Method to Analysis Resistance Characteristics of Conductive Concrete

Authors: Chun-Yao Lee, Siang-Ren Wang

Abstract:

The purpose of this paper is to discuss the influence of resistance characteristic on the high conductive concrete considering the various voltage and environment. The four-electrode method is applied to the tailor-made high conductive concrete with appropriate proportion. The curve of resistivity with the changes of voltage and environment is plotted and the changes of resistivity are explored. The result based on the methods reveals that resistivity is less affected by the temperature factor, and the four-electrode method would be an applicable measurement method on a site inspection.

Keywords: Conductive concrete, Resistivity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1604
2701 A New Algorithm to Stereo Correspondence Using Rank Transform and Morphology Based On Genetic Algorithm

Authors: Razagh Hafezi, Ahmad Keshavarz, Vida Moshfegh

Abstract:

This paper presents a novel algorithm of stereo correspondence with rank transform. In this algorithm we used the genetic algorithm to achieve the accurate disparity map. Genetic algorithms are efficient search methods based on principles of population genetic, i.e. mating, chromosome crossover, gene mutation, and natural selection. Finally morphology is employed to remove the errors and discontinuities.

Keywords: genetic algorithm, morphology, rank transform, stereo correspondence

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2173