Search results for: cyclic steps
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 685

Search results for: cyclic steps

175 PCR based Detection of Food Borne Pathogens

Authors: Archana Panchapakesan Iyer, Taha Abdullah Kumosani

Abstract:

Many high-risk pathogens that cause disease in humans are transmitted through various food items. Food-borne disease constitutes a major public health problem. Assessment of the quality and safety of foods is important in human health. Rapid and easy detection of pathogenic organisms will facilitate precautionary measures to maintain healthy food. The Polymerase Chain Reaction (PCR) is a handy tool for rapid detection of low numbers of bacteria. We have designed gene specific primers for most common food borne pathogens such as Staphylococci, Salmonella and E.coli. Bacteria were isolated from food samples of various food outlets and identified using gene specific PCRs. We identified Staphylococci, Salmonella and E.coli O157 using gene specific primers by rapid and direct PCR technique in various food samples. This study helps us in getting a complete picture of the various pathogens that threaten to cause and spread food borne diseases and it would also enable establishment of a routine procedure and methodology for rapid identification of food borne bacteria using the rapid technique of direct PCR. This study will also enable us to judge the efficiency of present food safety steps taken by food manufacturers and exporters.

Keywords: food borne pathogens, PCR, food safety, rapiddetection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2776
174 A Fast Neural Algorithm for Serial Code Detection in a Stream of Sequential Data

Authors: Hazem M. El-Bakry, Qiangfu Zhao

Abstract:

In recent years, fast neural networks for object/face detection have been introduced based on cross correlation in the frequency domain between the input matrix and the hidden weights of neural networks. In our previous papers [3,4], fast neural networks for certain code detection was introduced. It was proved in [10] that for fast neural networks to give the same correct results as conventional neural networks, both the weights of neural networks and the input matrix must be symmetric. This condition made those fast neural networks slower than conventional neural networks. Another symmetric form for the input matrix was introduced in [1-9] to speed up the operation of these fast neural networks. Here, corrections for the cross correlation equations (given in [13,15,16]) to compensate for the symmetry condition are presented. After these corrections, it is proved mathematically that the number of computation steps required for fast neural networks is less than that needed by classical neural networks. Furthermore, there is no need for converting the input data into symmetric form. Moreover, such new idea is applied to increase the speed of neural networks in case of processing complex values. Simulation results after these corrections using MATLAB confirm the theoretical computations.

Keywords: Fast Code/Data Detection, Neural Networks, Cross Correlation, real/complex values.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1589
173 Inverse Heat Conduction Analysis of Cooling on Run Out Tables

Authors: M. S. Gadala, Khaled Ahmed, Elasadig Mahdi

Abstract:

In this paper, we introduced a gradient-based inverse solver to obtain the missing boundary conditions based on the readings of internal thermocouples. The results show that the method is very sensitive to measurement errors, and becomes unstable when small time steps are used. The artificial neural networks are shown to be capable of capturing the whole thermal history on the run-out table, but are not very effective in restoring the detailed behavior of the boundary conditions. Also, they behave poorly in nonlinear cases and where the boundary condition profile is different. GA and PSO are more effective in finding a detailed representation of the time-varying boundary conditions, as well as in nonlinear cases. However, their convergence takes longer. A variation of the basic PSO, called CRPSO, showed the best performance among the three versions. Also, PSO proved to be effective in handling noisy data, especially when its performance parameters were tuned. An increase in the self-confidence parameter was also found to be effective, as it increased the global search capabilities of the algorithm. RPSO was the most effective variation in dealing with noise, closely followed by CRPSO. The latter variation is recommended for inverse heat conduction problems, as it combines the efficiency and effectiveness required by these problems.

Keywords: Inverse Analysis, Function Specification, Neural Net Works, Particle Swarm, Run Out Table.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1666
172 Parallel-computing Approach for FFT Implementation on Digital Signal Processor (DSP)

Authors: Yi-Pin Hsu, Shin-Yu Lin

Abstract:

An efficient parallel form in digital signal processor can improve the algorithm performance. The butterfly structure is an important role in fast Fourier transform (FFT), because its symmetry form is suitable for hardware implementation. Although it can perform a symmetric structure, the performance will be reduced under the data-dependent flow characteristic. Even though recent research which call as novel memory reference reduction methods (NMRRM) for FFT focus on reduce memory reference in twiddle factor, the data-dependent property still exists. In this paper, we propose a parallel-computing approach for FFT implementation on digital signal processor (DSP) which is based on data-independent property and still hold the property of low-memory reference. The proposed method combines final two steps in NMRRM FFT to perform a novel data-independent structure, besides it is very suitable for multi-operation-unit digital signal processor and dual-core system. We have applied the proposed method of radix-2 FFT algorithm in low memory reference on TI TMSC320C64x DSP. Experimental results show the method can reduce 33.8% clock cycles comparing with the NMRRM FFT implementation and keep the low-memory reference property.

Keywords: Parallel-computing, FFT, low-memory reference, TIDSP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2158
171 Malicious Route Defending Reliable-Data Transmission Scheme for Multi Path Routing in Wireless Network

Authors: S. Raja Ratna, R. Ravi

Abstract:

Securing the confidential data transferred via wireless network remains a challenging problem. It is paramount to ensure that data are accessible only by the legitimate users rather than by the attackers. One of the most serious threats to organization is jamming, which disrupts the communication between any two pairs of nodes. Therefore, designing an attack-defending scheme without any packet loss in data transmission is an important challenge. In this paper, Dependence based Malicious Route Defending DMRD Scheme has been proposed in multi path routing environment to prevent jamming attack. The key idea is to defend the malicious route to ensure perspicuous transmission. This scheme develops a two layered architecture and it operates in two different steps. In the first step, possible routes are captured and their agent dependence values are marked using triple agents. In the second step, the dependence values are compared by performing comparator filtering to detect malicious route as well as to identify a reliable route for secured data transmission. By simulation studies, it is observed that the proposed scheme significantly identifies malicious route by attaining lower delay time and route discovery time; it also achieves higher throughput.

Keywords: Attacker, Dependence, Jamming, Malicious.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1718
170 Building Information Modeling-Based Approach for Automatic Quantity Take-off and Cost Estimation

Authors: Lo Kar Yin, Law Ka Mei

Abstract:

Architectural, engineering, construction and operations (AECO) industry practitioners have been well adapting to the dynamic construction market from the fundamental training of its disciplines. As further triggered by the pandemic since 2019, great steps are taken in virtual environment and the best collaboration is strived with project teams without boundaries. With adoption of Building Information Modeling-based approach and qualitative analysis, this paper is to review quantity take-off (QTO) and cost estimation process through modeling techniques in liaison with suppliers, fabricators, subcontractors, contractors, designers, consultants and services providers in the construction industry value chain for automatic project cost budgeting, project cost control and cost evaluation on design options of in-situ reinforced-concrete construction and Modular Integrated Construction (MiC) at design stage, variation of works and cash flow/spending analysis at construction stage as far as practicable, with a view to sharing the findings for enhancing mutual trust and co-operation among AECO industry practitioners. It is to foster development through a common prototype of design and build project delivery method in NEC4 Engineering and Construction Contract (ECC) Options A and C.

Keywords: Building Information Modeling, cost estimation, quantity take-off, modeling techniques.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 632
169 A Study of Learning to Enhance Career Skills Consistent with Disruptive Innovation in the Creative Strategies for Advertising Course

Authors: Kornchanok Chidchaisuwan

Abstract:

This project is a study of learning activities of creating experience from actual work performance to enhance career skills and technological usage abilities for uses in advertising career work performance for undergraduate students who enroll in the Creative Strategies for Advertising Course. The instructional model consisted of two learning approaches: (1) simulation-based learning, which is the learning with the use of simulations of working in various sections of creative advertisement work with their own work process and steps as well as the virtual technology learning in advertising companies; and (2) project-based learning, which is the learning that the learners engage in actual work performance based on the process of creating and producing creative advertisement works to be present on new media channels. The results of learning management showed that the effects on the students in various aspects were as follows: (1) the students had experience in the advertising process at the higher level; and (2) the students had work performance skills from the actual work performance that enabled them to possess the abilities to create and present their own work; also, they had created more efficient work outcomes and disseminated them on new media channels at a better level.

Keywords: Technical literacy, career skill, experience, simulation-based learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 322
168 Selection the Optimum Cooling Scheme for Generators based on the Electro-Thermal Analysis

Authors: Diako Azizi, Ahmad Gholami, Vahid Abbasi

Abstract:

Optimal selection of electrical insulations in electrical machinery insures reliability during operation. From the insulation studies of view for electrical machines, stator is the most important part. This fact reveals the requirement for inspection of the electrical machine insulation along with the electro-thermal stresses. In the first step of the study, a part of the whole structure of machine in which covers the general characteristics of the machine is chosen, then based on the electromagnetic analysis (finite element method), the machine operation is simulated. In the simulation results, the temperature distribution of the total structure is presented simultaneously by using electro-thermal analysis. The results of electro-thermal analysis can be used for designing an optimal cooling system. In order to design, review and comparing the cooling systems, four wiring structures in the slots of Stator are presented. The structures are compared to each other in terms of electrical, thermal distribution and remaining life of insulation by using Finite Element analysis. According to the steps of the study, an optimization algorithm has been presented for selection of appropriate structure.

Keywords: Electrical field, field distribution, insulation, winding, finite element method, electro thermal

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703
167 Unsupervised Classification of DNA Barcodes Species Using Multi-Library Wavelet Networks

Authors: Abdesselem Dakhli, Wajdi Bellil, Chokri Ben Amar

Abstract:

DNA Barcode provides good sources of needed information to classify living species. The classification problem has to be supported with reliable methods and algorithms. To analyze species regions or entire genomes, it becomes necessary to use the similarity sequence methods. A large set of sequences can be simultaneously compared using Multiple Sequence Alignment which is known to be NP-complete. However, all the used methods are still computationally very expensive and require significant computational infrastructure. Our goal is to build predictive models that are highly accurate and interpretable. In fact, our method permits to avoid the complex problem of form and structure in different classes of organisms. The empirical data and their classification performances are compared with other methods. Evenly, in this study, we present our system which is consisted of three phases. The first one, is called transformation, is composed of three sub steps; Electron-Ion Interaction Pseudopotential (EIIP) for the codification of DNA Barcodes, Fourier Transform and Power Spectrum Signal Processing. Moreover, the second phase step is an approximation; it is empowered by the use of Multi Library Wavelet Neural Networks (MLWNN). Finally, the third one, is called the classification of DNA Barcodes, is realized by applying the algorithm of hierarchical classification.

Keywords: DNA Barcode, Electron-Ion Interaction Pseudopotential, Multi Library Wavelet Neural Networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1928
166 Down-Regulated Gene Expression of GKN1 and GKN2 as Diagnostic Markers for Gastric Cancer

Authors: Amer A. Hasan, Mehri Igci, Ersin Borazan, Rozhgar A. Khailany, Emine Bayraktar, Ahmet Arslan

Abstract:

Gastric Cancer (GC) has high morbidity and fatality rate in various countries. It is still one of the most frequent and deadly diseases. Gastrokine1 (GKN1) and gastrokine2 (GKN2) genes are highly expressed in the normal stomach epithelium and play important roles in maintaining the integrity and homeostasis of stomach mucosal epithelial cells. In this study, 47 paired samples that were grouped according to the types of gastric cancer and the clinical characteristics of the patients, including gender and average of age. They were investigated with gene expression analysis and mutation screening by monitoring RT-PCR, SSCP and nucleotide sequencing techniques. Both GKN1 and GKN2 genes were observed significantly reduced found by (Wilcoxon signed rank test; p<0.05). As a result of gene screening, no mutation (no different genotype) was detected. It is considered that gene mutations are not the cause of gastrokines inactivation. In conclusion, the mRNA expression level of GKN1 and GKN2 genes statistically was decreased regardless the gender, age, or cancer type of patients. Reduced of gastrokine genes seem to occur at the initial steps of gastric cancer development.

Keywords: Diagnostic biomarker, gastric cancer, nucleotide sequencing, semi-quantitative RT-PCR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1422
165 Supervisor Controller-Based Colored Petri Nets for Deadlock Control and Machine Failures in Automated Manufacturing Systems

Authors: Husam Kaid, Abdulrahman Al-Ahmari, Zhiwu Li

Abstract:

This paper develops a robust deadlock control technique for shared and unreliable resources in automated manufacturing systems (AMSs) based on structural analysis and colored Petri nets, which consists of three steps. The first step involves using strict minimal siphon control to create a live (deadlock-free) system that does not consider resource failure. The second step uses an approach based on colored Petri net, in which all monitors designed in the first step are merged into a single monitor. The third step addresses the deadlock control problems caused by resource failures. For all resource failures in the Petri net model a common recovery subnet based on colored petri net is proposed. The common recovery subnet is added to the obtained system at the second step to make the system reliable. The proposed approach is evaluated using an AMS from the literature. The results show that the proposed approach can be applied to an unreliable complex Petri net model, has a simpler structure and less computational complexity, and can obtain one common recovery subnet to model all resource failures.

Keywords: Automated manufacturing system, colored Petri net, deadlock, siphon.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 415
164 A Computer Aided Detection (CAD) System for Microcalcifications in Mammograms - MammoScan mCaD

Authors: Kjersti Engan, Thor Ole Gulsrud, Karl Fredrik Fretheim, Barbro Furebotten Iversen, Liv Eriksen

Abstract:

Clusters of microcalcifications in mammograms are an important sign of breast cancer. This paper presents a complete Computer Aided Detection (CAD) scheme for automatic detection of clustered microcalcifications in digital mammograms. The proposed system, MammoScan μCaD, consists of three main steps. Firstly all potential microcalcifications are detected using a a method for feature extraction, VarMet, and adaptive thresholding. This will also give a number of false detections. The goal of the second step, Classifier level 1, is to remove everything but microcalcifications. The last step, Classifier level 2, uses learned dictionaries and sparse representations as a texture classification technique to distinguish single, benign microcalcifications from clustered microcalcifications, in addition to remove some remaining false detections. The system is trained and tested on true digital data from Stavanger University Hospital, and the results are evaluated by radiologists. The overall results are promising, with a sensitivity > 90 % and a low false detection rate (approx 1 unwanted pr. image, or 0.3 false pr. image).

Keywords: mammogram, microcalcifications, detection, CAD, MammoScan μCaD, VarMet, dictionary learning, texture, FTCM, classification, adaptive thresholding

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1772
163 Automated Testing of Workshop Robot Behavior

Authors: Arne Hitzmann, Philipp Wentscher, Alexander Gabel, Reinhard Gerndt

Abstract:

Autonomous mobile robots can be found in a wide field of applications. Their types range from household robots over workshop robots to autonomous cars and many more. All of them undergo a number of testing steps during development, production and maintenance. This paper describes an approach to improve testing of robot behavior. It was inspired by the RoboCup @work competition that itself reflects a robotics benchmark for industrial robotics. There, scaled down versions of mobile industrial robots have to navigate through a workshop-like environment or operation area and have to perform tasks of manipulating and transporting work pieces. This paper will introduce an approach of automated vision-based testing of the behavior of the so called youBot robot, which is the most widely used robot platform in the RoboCup @work competition. The proposed system allows automated testing of multiple tries of the robot to perform a specific missions and it allows for the flexibility of the robot, e.g. selecting different paths between two tasks within a mission. The approach is based on a multi-camera setup using, off the shelf cameras and optical markers. It has been applied for test-driven development (TDD) and maintenance-like verification of the robot behavior and performance.

Keywords: Supervisory control, Testing, Markers, Mono Vision, Automation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2349
162 Validation and Selection between Machine Learning Technique and Traditional Methods to Reduce Bullwhip Effects: a Data Mining Approach

Authors: Hamid R. S. Mojaveri, Seyed S. Mousavi, Mojtaba Heydar, Ahmad Aminian

Abstract:

The aim of this paper is to present a methodology in three steps to forecast supply chain demand. In first step, various data mining techniques are applied in order to prepare data for entering into forecasting models. In second step, the modeling step, an artificial neural network and support vector machine is presented after defining Mean Absolute Percentage Error index for measuring error. The structure of artificial neural network is selected based on previous researchers' results and in this article the accuracy of network is increased by using sensitivity analysis. The best forecast for classical forecasting methods (Moving Average, Exponential Smoothing, and Exponential Smoothing with Trend) is resulted based on prepared data and this forecast is compared with result of support vector machine and proposed artificial neural network. The results show that artificial neural network can forecast more precisely in comparison with other methods. Finally, forecasting methods' stability is analyzed by using raw data and even the effectiveness of clustering analysis is measured.

Keywords: Artificial Neural Networks (ANN), bullwhip effect, demand forecasting, Support Vector Machine (SVM).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1969
161 Analytical Comparison of Conventional Algorithms with Vedic Algorithm for Digital Multiplier

Authors: Akhilesh G. Naik, Dipankar Pal

Abstract:

In today’s scenario, the complexity of digital signal processing (DSP) applications and various microcontroller architectures have been increasing to such an extent that the traditional approaches to multiplier design in most processors are becoming outdated for being comparatively slow. Modern processing applications require suitable pipelined approaches, and therefore, algorithms that are friendlier with pipelined architectures. Traditional algorithms like Wallace Tree, Radix-4 Booth, Radix-8 Booth, Dadda architectures have been proven to be comparatively slow for pipelined architectures. These architectures, therefore, need to be optimized or combined with other architectures amongst them to enhance its performances and to be made suitable for pipelined hardware/architectures. Recently, Vedic algorithm mathematically has proven to be efficient by appearing to be less complex and with fewer steps for its output establishment and have assumed renewed importance. This paper describes and shows how the Vedic algorithm can be better suited for pipelined architectures and also can be combined with traditional architectures and algorithms for enhancing its ability even further. In this paper, we also established that for complex applications on DSP and other microcontroller architectures, using Vedic approach for multiplication proves to be the best available and efficient option.

Keywords: Wallace tree, Radix-4 Booth, Radix-8 Booth, Dadda, Vedic, Single-Stage Karatsuba, Looped Karatsuba.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 787
160 Implementation of CMMS Software for a Maintenance Plan in a Manufacturing Industry

Authors: Abimbola O. Aniki, Esther T. Akinlabi

Abstract:

This paper proposes an effective maintenance method by considering the implementation of the Computerized Maintenance Management System (CMMS) software to plan a maintenance activity in a manufacturing industry. Globally, maintenance is a very important activity in the manufacturing sector to prolong the life span of equipment and machinery; it is also applicable to all household items. It is obvious and well known that apart from giving long life to equipment, it reduces the substantial financial losses for repairs and save the production downtime. In some cases, appropriate maintenance of plant equipment and machinery reduces the tendencies of injuries to personnel in the job floor. But before the maintenance process can be carried out, proper and effective work order planning and scheduling must be in place in other to achieve the set goals and objectives of a maintenance shop. Brief reviews of common planning tools which include the Computerized Maintenance Management System (CMMS) are presented. An interesting outline of analyses on planning and scheduling for effective job planning in a typical manufacturing industry using the CMMS is also presented and discussed. Finally, the steps to adhere to in making job planning effective in a manufacturing industry are also highlighted.

Keywords: Advanced Downtime Analysis Programme (ADAP), Computerized Maintenance Management System (CMMS), Corrective Maintenance (CM), Executing Department (ED), Maintenance Department (MD), Preventive Maintenance (PM).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3382
159 Detecting Email Forgery using Random Forests and Naïve Bayes Classifiers

Authors: Emad E Abdallah, A.F. Otoom, ArwaSaqer, Ola Abu-Aisheh, Diana Omari, Ghadeer Salem

Abstract:

As emails communications have no consistent authentication procedure to ensure the authenticity, we present an investigation analysis approach for detecting forged emails based on Random Forests and Naïve Bays classifiers. Instead of investigating the email headers, we use the body content to extract a unique writing style for all the possible suspects. Our approach consists of four main steps: (1) The cybercrime investigator extract different effective features including structural, lexical, linguistic, and syntactic evidence from previous emails for all the possible suspects, (2) The extracted features vectors are normalized to increase the accuracy rate. (3) The normalized features are then used to train the learning engine, (4) upon receiving the anonymous email (M); we apply the feature extraction process to produce a feature vector. Finally, using the machine learning classifiers the email is assigned to one of the suspects- whose writing style closely matches M. Experimental results on real data sets show the improved performance of the proposed method and the ability of identifying the authors with a very limited number of features.

Keywords: Digital investigation, cybercrimes, emails forensics, anonymous emails, writing style, and authorship analysis

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5210
158 Lean Manufacturing: Systematic Layout Planning Application to an Assembly Line Layout of a Welding Industry

Authors: Fernando Augusto Ullmann Tobe, Moacyr Amaral Domingues, Figueiredo, Stephany Rie Yamamoto Gushiken

Abstract:

The purpose of this paper is to present the process of elaborating the layout of an assembly line of a welding industry using the principles of lean manufacturing as the main driver. The objective of this paper is relevant since the current layout of the assembly line causes non-productive times for operators, being related to the lean waste of unnecessary movements. The methodology used for the project development was Project-based Learning (PBL), which is an active way of learning focused on real problems. The process of selecting the methodology for layout planning was developed considering three criteria to evaluate the most relevant one for this paper's goal. As a result of this evaluation, Systematic Layout Planning was selected, and three steps were added to it – Value Stream Mapping for the current situation and after layout changed and the definition of lean tools and layout type. This inclusion was to consider lean manufacturing in the layout redesign of the industry. The layout change resulted in an increase in the value-adding time of operations carried out in the sector, reduction in movement times between previous and final assemblies, and in cost savings regarding the man-hour value of the employees, which can be invested in productive hours instead of movement times.

Keywords: Assembly line, layout, lean manufacturing, systematic layout planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 765
157 Numerical Simulation of Natural Gas Dispersion from Low Pressure Pipelines

Authors: Omid Adibi, Nategheh Najafpour, Bijan Farhanieh, Hossein Afshin

Abstract:

Gas release from the pipelines is one of the main factors in the gas industry accidents. Released gas ejects from the pipeline as a free jet and in the growth process, the fuel gets mixed with the ambient air. Accordingly, an accidental spark will release the chemical energy of the mixture with an explosion. Gas explosion damages the equipment and endangers the life of staffs. So due to importance of safety in gas industries, prevision of accident can reduce the number of the casualties. In this paper, natural gas leakages from the low pressure pipelines are studied in two steps: 1) the simulation of mixing process and identification of flammable zones and 2) the simulation of wind effects on the mixing process. The numerical simulations were performed by using the finite volume method and the pressure-based algorithm. Also, for the grid generation the structured method was used. The results show that, in just 6.4 s after accident, released natural gas could penetrate to 40 m in vertical and 20 m in horizontal direction. Moreover, the results show that the wind speed is a key factor in dispersion process. In fact, the wind transports the flammable zones into the downstream. Hence, to improve the safety of the people and human property, it is preferable to construct gas facilities and buildings in the opposite side of prevailing wind direction.

Keywords: Flammable zones, gas pipelines, numerical simulation, wind effects.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1106
156 Highly Accurate Target Motion Compensation Using Entropy Function Minimization

Authors: Amin Aghatabar Roodbary, Mohammad Hassan Bastani

Abstract:

One of the defects of stepped frequency radar systems is their sensitivity to target motion. In such systems, target motion causes range cell shift, false peaks, Signal to Noise Ratio (SNR) reduction and range profile spreading because of power spectrum interference of each range cell in adjacent range cells which induces distortion in High Resolution Range Profile (HRRP) and disrupt target recognition process. Thus Target Motion Parameters (TMPs) effects compensation should be employed. In this paper, such a method for estimating TMPs (velocity and acceleration) and consequently eliminating or suppressing the unwanted effects on HRRP based on entropy minimization has been proposed. This method is carried out in two major steps: in the first step, a discrete search method has been utilized over the whole acceleration-velocity lattice network, in a specific interval seeking to find a less-accurate minimum point of the entropy function. Then in the second step, a 1-D search over velocity is done in locus of the minimum for several constant acceleration lines, in order to enhance the accuracy of the minimum point found in the first step. The provided simulation results demonstrate the effectiveness of the proposed method.

Keywords: ATR, HRRP, motion compensation, SFW, TMP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 610
155 Study on the Optimization of Completely Batch Water-using Network with Multiple Contaminants Considering Flow Change

Authors: Jian Du, Shui Hong Hong, Lu Meng, Qing Wei Meng

Abstract:

This work addresses the problem of optimizing completely batch water-using network with multiple contaminants where the flow change caused by mass transfer is taken into consideration for the first time. A mathematical technique for optimizing water-using network is proposed based on source-tank-sink superstructure. The task is to obtain the freshwater usage, recycle assignments among water-using units, wastewater discharge and a steady water-using network configuration by following steps. Firstly, operating sequences of water-using units are determined by time constraints. Next, superstructure is simplified by eliminating the reuse and recycle from water-using units with maximum concentration of key contaminants. Then, the non-linear programming model is solved by GAMS (General Algebra Model System) for minimum freshwater usage, maximum water recycle and minimum wastewater discharge. Finally, numbers of operating periods are calculated to acquire the steady network configuration. A case study is solved to illustrate the applicability of the proposed approach.

Keywords: Completely batch process, flow change, multiple contaminants, water-using network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1398
154 The Appropriateness of Antibiotic Prescribing within Dundee Dental Hospital

Authors: Salma Ainine, Colin Ritchie, Tracey McFee

Abstract:

Background: The societal impact of antibiotic resistance is a major public health concern. The increase in incidence of resistant bacteria can ultimately be fatal. Objective: To analyse the appropriateness of antibiotic prescribing in Dundee Dental Hospital, ultimately improving the safety and quality of patient care. Methods: Two examiners independently crosschecked approximately fifty consecutive prescriptions, and corresponding patient case notes, for three data collection cycles between August 2014 – September 2015. The Scottish Dental Clinical Effectiveness Program (SDCEP) Drug Prescribing for Dentistry guidelines was the standard utilised. The criteria: clinical justification, regime justification and review arrangements was measured, and compared to the standard. Results: Cycle one revealed 42% of antibiotic prescriptions were appropriate. Interventions included: multiple staff meetings, introduction of a checklist attached to the prescription pack, and production of patient leaflets explaining indications for antibiotics. Cycle two and three revealed 44%, and 30% compliance, respectively. Conclusion: The results of the audit have yet to meet target standards set out in prescribing guidelines. However, steps are being taken and change has occurred on a cultural level.

Keywords: Antibiotic resistance, antibiotic stewardship, dental infection and hygiene standards.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1927
153 Effective Scheduling of Semiconductor Manufacturing using Simulation

Authors: Ingy A. El-Khouly, Khaled S. El-Kilany, Aziz E. El-Sayed

Abstract:

The process of wafer fabrication is arguably the most technologically complex and capital intensive stage in semiconductor manufacturing. This large-scale discrete-event process is highly reentrant, and involves hundreds of machines, restrictions, and processing steps. Therefore, production control of wafer fabrication facilities (fab), specifically scheduling, is one of the most challenging problems that this industry faces. Dispatching rules have been extensively applied to the scheduling problems in semiconductor manufacturing. Moreover, lot release policies are commonly used in this manufacturing setting to further improve the performance of such systems and reduce its inherent variability. In this work, simulation is used in the scheduling of re-entrant flow shop manufacturing systems with an application in semiconductor wafer fabrication; where, a simulation model has been developed for the Intel Five-Machine Six Step Mini-Fab using the ExtendTM simulation environment. The Mini-Fab has been selected as it captures the challenges involved in scheduling the highly re-entrant semiconductor manufacturing lines. A number of scenarios have been developed and have been used to evaluate the effect of different dispatching rules and lot release policies on the selected performance measures. Results of simulation showed that the performance of the Mini-Fab can be drastically improved using a combination of dispatching rules and lot release policy.

Keywords: Dispatching rules, lot release policy, re-entrant flowshop, semiconductor manufacturing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2503
152 High-Frequency Monitoring Results of a Piled Raft Foundation under Wind Loading

Authors: Laurent Pitteloud, Jörg Meier

Abstract:

Piled raft foundations represent an efficient and reliable technique for transferring high vertical and horizontal loads to the subsoil. Piled raft foundations were success­fully implemented for several high-rise buildings world­wide over the last decades. For the structural design of this foundation type the stiffnesses of both the piles and the raft have to be deter­mined for the static (e.g. dead load, live load) and the dynamic load cases (e.g. earthquake). In this context the question often arises, to which proportion wind loads are to be considered as dynamic loads. Usually a piled raft foundation has to be monitored in order to verify the design hypotheses. As an additional benefit, the analysis of this monitoring data may lead to a better under­standing of the behaviour of this foundation type for future projects in similar subsoil conditions. In case the measurement frequency is high enough, one may also draw conclusions on the effect of wind loading on the piled raft foundation. For a 41-storey office building in Basel, Switzerland, the preliminary design showed that a piled raft foundation was the best solution to satisfy both design requirements, as well as economic aspects. A high-frequency monitoring of the foundation including pile loads, vertical stresses under the raft, as well as pore water pressures was performed over 5 years. In windy situations the analysis of the measure­ments shows that the pile load increment due to wind consists of a static and a cyclic load term. As piles and raft react with different stiffnesses under static and dynamic loading, these measure­ments are useful for the correct definition of stiffnesses of future piled raft foundations. This paper outlines the design strategy and the numerical modelling of the aforementioned piled raft foundation. The measurement results are presented and analysed. Based on the findings, comments and conclusions on the definition of pile and raft stiffnesses for vertical and wind loading are proposed.

Keywords: Dynamic loading, high-frequency monitoring, piled raft foundations, wind loading.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 757
151 Modeling of Electrokinetic Mixing in Lab on Chip Microfluidic Devices

Authors: Virendra J. Majarikar, Harikrishnan N. Unni

Abstract:

This paper sets to demonstrate a modeling of electrokinetic mixing employing electroosmotic stationary and time-dependent microchannel using alternate zeta patches on the lower surface of the micromixer in a lab on chip microfluidic device. Electroosmotic flow is amplified using different 2D and 3D model designs with alternate and geometric zeta potential values such as 25, 50, and 100 mV, respectively, to achieve high concentration mixing in the electrokinetically-driven microfluidic system. The enhancement of electrokinetic mixing is studied using Finite Element Modeling, and simulation workflow is accomplished with defined integral steps. It can be observed that the presence of alternate zeta patches can help inducing microvortex flows inside the channel, which in turn can improve mixing efficiency. Fluid flow and concentration fields are simulated by solving Navier-Stokes equation (implying Helmholtz-Smoluchowski slip velocity boundary condition) and Convection-Diffusion equation. The effect of the magnitude of zeta potential, the number of alternate zeta patches, etc. are analysed thoroughly. 2D simulation reveals that there is a cumulative increase in concentration mixing, whereas 3D simulation differs slightly with low zeta potential as that of the 2D model within the T-shaped micromixer for concentration 1 mol/m3 and 0 mol/m3, respectively. Moreover, 2D model results were compared with those of 3D to indicate the importance of the 3D model in a microfluidic design process.

Keywords: COMSOL, electrokinetic, electroosmotic, microfluidics, zeta potential.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1159
150 A New Intelligent, Dynamic and Real Time Management System of Sewerage

Authors: R. Tlili Yaakoubi, H. Nakouri, O. Blanpain, S. Lallahem

Abstract:

The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of this project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 19 to 100 %. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 40 % of total volume rejected to the natural environment and of 65 % in the number of discharges.

Keywords: Automation, optimization, paradigm, RTC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1454
149 The Mitigation Strategy Analysis of Kuosheng Nuclear Power Plant Spent Fuel Pool Using MELCOR2.1/SNAP

Authors: Y. Chiang, J. R. Wang, J. H. Yang, Y. S. Tseng, C. Shih, S. W. Chen

Abstract:

Kuosheng nuclear power plant (NPP) is a BWR/6 plant in Taiwan. There is more concern for the safety of Spent Fuel Pools (SFPs) in Taiwan after Fukushima event. In order to estimate the safety of Kuosheng NPP SFP, by using MELCOR2.1 and SNAP, the safety analysis of Kuosheng NPP SFP was performed combined with the mitigation strategy of NEI 06-12 report. There were several steps in this research. First, the Kuosheng NPP SFP models were established by MELCOR2.1/SNAP. Second, the Station Blackout (SBO) analysis of Kuosheng SFP was done by TRACE and MELCOR under the cooling system failure condition. The results showed that the calculations of MELCOR and TRACE were very similar in this case. Second, the mitigation strategy analysis was done with the MELCOR model by following the NEI 06-12 report. The results showed the effectiveness of NEI 06-12 strategy in Kuosheng NPP SFP. Finally, a sensitivity study of SFP quenching was done to check the differences of different water injection time and the phenomena during the quenching. The results showed that if the cladding temperature was over 1600 K, the water injection may have chance to cause the accident more severe with more hydrogen generation. It was because of the oxidation heat and the “Breakaway” effect of the zirconium-water reaction. An animation model built by SNAP was also shown in this study.

Keywords: MELCOR, SNAP, spent fuel pool, quenching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 907
148 A TRIZ-based Approach to Generation of Service-supporting Product Concepts

Authors: Seungkyum Kim, Yongtae Park

Abstract:

Recently, business environment and customer needs have become rapidly changing, hence it is very difficult to fulfill sophisticated customer needs by product or service innovation only. In practice, to cope with this problem, various manufacturing companies have developed services to combine with their products. Along with this, many academic studies on PSS (Product Service System) which is the integrated system of products and services have been conducted from the viewpoint of manufacturers. On the other hand, service providers are also attempting to develop service-supporting products to increase their service competitiveness and provide differentiated value. However, there is a lack of research based on the service-centric point of view. Accordingly, this paper proposes a concept generation method for service-supporting product development from the service-centric point of view. This method is designed to be executed in five consecutive steps: situation analysis, problem definition, problem resolution, solution evaluation, and concept generation. In the proposed approach, some tools of TRIZ (Theory of Solving Inventive Problem) such as ISQ (Innovative Situation Questionnaire) and 40 inventive principles are employed in order to define problems of the current services and solve them by generating service-supporting product concepts. This research contributes to the development of service-supporting products and service-centric PSSs.

Keywords: TRIZ, PSS (Product Service System), service-supporting product, concept generation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1895
147 Timescape-Based Panoramic View for Historic Landmarks

Authors: H. Ali, A. Whitehead

Abstract:

Providing a panoramic view of famous landmarks around the world offers artistic and historic value for historians, tourists, and researchers. Exploring the history of famous landmarks by presenting a comprehensive view of a temporal panorama merged with geographical and historical information presents a unique challenge of dealing with images that span a long period, from the 1800’s up to the present. This work presents the concept of temporal panorama through a timeline display of aligned historic and modern images for many famous landmarks. Utilization of this panorama requires a collection of hundreds of thousands of landmark images from the Internet comprised of historic images and modern images of the digital age. These images have to be classified for subset selection to keep the more suitable images that chronologically document a landmark’s history. Processing of historic images captured using older analog technology under various different capturing conditions represents a big challenge when they have to be used with modern digital images. Successful processing of historic images to prepare them for next steps of temporal panorama creation represents an active contribution in cultural heritage preservation through the fulfillment of one of UNESCO goals in preservation and displaying famous worldwide landmarks.

Keywords: Cultural heritage, image registration, image subset selection, registered image similarity, temporal panorama, timescapes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 997
146 Novel D- glucose Based Glycomonomers Synthesis and Characterization

Authors: M.S. Mazăre, A. M. Pană, L. M. Ştefan, M. Silion, M. Bălan, G. Bandur, L. M. Rusnac

Abstract:

In the last decade, carbohydrates have attracted great attention as renewable resources for the chemical industry. Carbohydrates are abundantly found in nature in the form of monomers, oligomers and polymers, or as components of biopolymers and other naturally occurring substances. As natural products, they play important roles in conferring certain physical, chemical, and biological properties to their carrier molecules.The synthesis of this particular carbohydrate glycomonomer is part of our work to obtain biodegradable polymers. Our current paper describes the synthesis and characterization of a novel carbohydrate glycomonomer starting from D-glucose, in several synthesis steps, that involve the protection/deprotection of the D-glucose ring via acetylation, tritylation, then selective deprotection of the aromaticaliphatic protective group, in order to obtain 1,2,3,4-tetra-O-acetyl- 6-O-allyl-β-D-glucopyranose. The glycomonomer was then obtained by the allylation in drastic conditions of 1,2,3,4-tetra-O-acetyl-6-Oallyl- β-D-glucopyranose with allylic alcohol in the presence of stannic chloride, in methylene chloride, at room temperature. The proposed structure of the glycomonomer, 2,3,4-tri-O-acetyl-1,6-di- O-allyl-β-D-glucopyranose, was confirmed by FTIR, NMR and HPLC-MS spectrometry. This glycomonomer will be further submitted to copolymerization with certain acrylic or methacrylic monomers in order to obtain competitive plastic materials for applications in the biomedical field.

Keywords: allylation, D-glucose, glycomonomer, trityl chloride

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2004