Search results for: Image registration techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3814

Search results for: Image registration techniques

1174 EEG Signal Processing Methods to Differentiate Mental States

Authors: Sun H. Hwang, Young E. Lee, Yunhan Ga, Gilwon Yoon

Abstract:

EEG is a very complex signal with noises and other bio-potential interferences. EOG is the most distinct interfering signal when EEG signals are measured and analyzed. It is very important how to process raw EEG signals in order to obtain useful information. In this study, the EEG signal processing techniques such as EOG filtering and outlier removal were examined to minimize unwanted EOG signals and other noises. The two different mental states of resting and focusing were examined through EEG analysis. A focused state was induced by letting subjects to watch a red dot on the white screen. EEG data for 32 healthy subjects were measured. EEG data after 60-Hz notch filtering were processed by a commercially available EOG filtering and our presented algorithm based on the removal of outliers. The ratio of beta wave to theta wave was used as a parameter for determining the degree of focusing. The results show that our algorithm was more appropriate than the existing EOG filtering.

Keywords: EEG, focus, mental state, outlier, signal processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543
1173 Model Predictive Control and Proportional-Integral-Derivative Control of Quadcopters: A Comparative Analysis

Authors: Anel Hasić, Naser Prljača

Abstract:

In the domain of autonomous or piloted flights, the accurate control of quadrotor trajectories is of paramount significance for large numbers of tasks. These adaptable aerial platforms find applications that span from high-precision aerial photography and surveillance to demanding search and rescue missions. Among the fundamental challenges confronting quadrotor operation is the demand for accurate following of desired flight paths. To address this control challenge, among others, two celebrated well-established control strategies have emerged as noteworthy contenders: Model Predictive Control (MPC) and Proportional-Integral-Derivative (PID) control. In this work, we focus on the extensive examination of MPC and PID control techniques by using comprehensive simulation studies in MATLAB/Simulink. Intensive simulation results demonstrate the performance of the studied control algorithms.

Keywords: MATLAB, MPC, Model Predictive Control, PID, Proportional-Integral-Derivative, quadcopter, Simulink.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32
1172 Continuum-Based Modelling Approaches for Cell Mechanics

Authors: Yogesh D. Bansod, Jiri Bursa

Abstract:

The quantitative study of cell mechanics is of paramount interest, since it regulates the behaviour of the living cells in response to the myriad of extracellular and intracellular mechanical stimuli. The novel experimental techniques together with robust computational approaches have given rise to new theories and models, which describe cell mechanics as combination of biomechanical and biochemical processes. This review paper encapsulates the existing continuum-based computational approaches that have been developed for interpreting the mechanical responses of living cells under different loading and boundary conditions. The salient features and drawbacks of each model are discussed from both structural and biological points of view. This discussion can contribute to the development of even more precise and realistic computational models of cell mechanics based on continuum approaches or on their combination with microstructural approaches, which in turn may provide a better understanding of mechanotransduction in living cells.

Keywords: Cell mechanics, computational models, continuum approach, mechanical models.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2956
1171 Multi-Criteria Spatial Analysis for the Localization of Production Structures. Analytic Hierarchy Process and Geographical Information Systems in the Case of Expanding an Industrial Area

Authors: Gianluigi De Mare, Pierluigi Morano, Antonio Nesticò

Abstract:

Among the numerous economic evaluation techniques currently available, Multi-criteria Spatial Analysis lends itself to solving localization problems of property complexes and, in particular, production plants. The methodology involves the use of Geographical Information Systems (GIS) and the mapping overlay technique, which overlaps the different information layers of a territory in order to obtain an overview of the parameters that characterize it. This first phase is used to detect possible settlement surfaces of a new agglomeration, subsequently selected through Analytic Hierarchy Process (AHP), so as to choose the best alternative. The result ensures the synthesis of a multidimensional profile that expresses both the quantitative and qualitative effects. Each criterion can be given a different weight.

Keywords: Multi-criteria Spatial Analysis, Analytic Hierarchy Process, Geographical Information Systems, localization of industrial areas.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2039
1170 Alertness States Classification By SOM and LVQ Neural Networks

Authors: K. Ben Khalifa, M.H. Bédoui, M. Dogui, F. Alexandre

Abstract:

Several studies have been carried out, using various techniques, including neural networks, to discriminate vigilance states in humans from electroencephalographic (EEG) signals, but we are still far from results satisfactorily useable results. The work presented in this paper aims at improving this status with regards to 2 aspects. Firstly, we introduce an original procedure made of the association of two neural networks, a self organizing map (SOM) and a learning vector quantization (LVQ), that allows to automatically detect artefacted states and to separate the different levels of vigilance which is a major breakthrough in the field of vigilance. Lastly and more importantly, our study has been oriented toward real-worked situation and the resulting model can be easily implemented as a wearable device. It benefits from restricted computational and memory requirements and data access is very limited in time. Furthermore, some ongoing works demonstrate that this work should shortly results in the design and conception of a non invasive electronic wearable device.

Keywords: Electroencephalogram interpretation, artificialneural networks, vigilance states, hardware implementation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1476
1169 Unsupervised Feature Learning by Pre-Route Simulation of Auto-Encoder Behavior Model

Authors: Youngjae Jin, Daeshik Kim

Abstract:

This paper describes a cycle accurate simulation results of weight values learned by an auto-encoder behavior model in terms of pre-route simulation. Given the results we visualized the first layer representations with natural images. Many common deep learning threads have focused on learning high-level abstraction of unlabeled raw data by unsupervised feature learning. However, in the process of handling such a huge amount of data, the learning method’s computation complexity and time limited advanced research. These limitations came from the fact these algorithms were computed by using only single core CPUs. For this reason, parallel-based hardware, FPGAs, was seen as a possible solution to overcome these limitations. We adopted and simulated the ready-made auto-encoder to design a behavior model in VerilogHDL before designing hardware. With the auto-encoder behavior model pre-route simulation, we obtained the cycle accurate results of the parameter of each hidden layer by using MODELSIM. The cycle accurate results are very important factor in designing a parallel-based digital hardware. Finally this paper shows an appropriate operation of behavior model based pre-route simulation. Moreover, we visualized learning latent representations of the first hidden layer with Kyoto natural image dataset.

Keywords: Auto-encoder, Behavior model simulation, Digital hardware design, Pre-route simulation, Unsupervised feature learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2690
1168 Comparative Study of Filter Characteristics as Statistical Vocal Correlates of Clinical Psychiatric State in Human

Authors: Thaweesak Yingthawornsuk, Chusak Thanawattano

Abstract:

Acoustical properties of speech have been shown to be related to mental states of speaker with symptoms: depression and remission. This paper describes way to address the issue of distinguishing depressed patients from remitted subjects based on measureable acoustics change of their spoken sound. The vocal-tract related frequency characteristics of speech samples from female remitted and depressed patients were analyzed via speech processing techniques and consequently, evaluated statistically by cross-validation with Support Vector Machine. Our results comparatively show the classifier's performance with effectively correct separation of 93% determined from testing with the subjectbased feature model and 88% from the frame-based model based on the same speech samples collected from hospital visiting interview sessions between patients and psychiatrists.

Keywords: Depression, SVM, Vocal Extract, Vocal Tract

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1541
1167 Ensembling Classifiers – An Application toImage Data Classification from Cherenkov Telescope Experiment

Authors: Praveen Boinee, Alessandro De Angelis, Gian Luca Foresti

Abstract:

Ensemble learning algorithms such as AdaBoost and Bagging have been in active research and shown improvements in classification results for several benchmarking data sets with mainly decision trees as their base classifiers. In this paper we experiment to apply these Meta learning techniques with classifiers such as random forests, neural networks and support vector machines. The data sets are from MAGIC, a Cherenkov telescope experiment. The task is to classify gamma signals from overwhelmingly hadron and muon signals representing a rare class classification problem. We compare the individual classifiers with their ensemble counterparts and discuss the results. WEKA a wonderful tool for machine learning has been used for making the experiments.

Keywords: Ensembles, WEKA, Neural networks [NN], SupportVector Machines [SVM], Random Forests [RF].

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1765
1166 The Use of Degradation Measures to Design Reliability Test Plans

Authors: Stephen V. Crowder, Jonathan W. Lane

Abstract:

With short production development times, there is an increased need to demonstrate product reliability relatively quickly with minimal testing. In such cases there may be few if any observed failures. Thus it may be difficult to assess reliability using the traditional reliability test plans that measure only time (or cycles) to failure. For many components, degradation measures will contain important information about performance and reliability. These measures can be used to design a minimal test plan, in terms of number of units placed on test and duration of the test, necessary to demonstrate a reliability goal. In this work we present a case study involving an electronic component subject to degradation. The data, consisting of 42 degradation paths of cycles to failure, are first used to estimate a reliability function. Bootstrapping techniques are then used to perform power studies and develop a minimal reliability test plan for future production of this component. 

Keywords: Degradation Measure, Time to Failure Distribution, Bootstrap.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1878
1165 Logistics Support as a Key Success Factor in Gastronomy

Authors: Hanna Ziętara

Abstract:

Gastronomy is one of the oldest forms of commercial activity. It is currently one of the most popular and still dynamically developing branches of business. Socio-economic changes, its widespread occurrence, new techniques or culinary styles affect the almost unlimited possibilities of its development. Importantly, regardless of the form of business adopted, foodservice is strongly related to logistics processes, and areas of foodservice that are closely linked to logistics are of strategic importance. Any inefficiency in logistics processes results in reduced chances for success and achieving competitive advantage by companies belonging to the catering industry. The aim of the paper is to identify the areas of logistic support, occurring in the catering business, and affecting the scope of the logistic processes implemented. The aim of the paper is implemented through a plural homogeneous approach, based on direct observation, text analysis of current documents, and in-depth free targeted interviews.

Keywords: Gastronomy, competitive advantage, logistics, logistics support.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 569
1164 Adomian’s Decomposition Method to Functionally Graded Thermoelastic Materials with Power Law

Authors: Hamdy M. Youssef, Eman A. Al-Lehaibi

Abstract:

This paper presents an iteration method for the numerical solutions of a one-dimensional problem of generalized thermoelasticity with one relaxation time under given initial and boundary conditions. The thermoelastic material with variable properties as a power functional graded has been considered. Adomian’s decomposition techniques have been applied to the governing equations. The numerical results have been calculated by using the iterations method with a certain algorithm. The numerical results have been represented in figures, and the figures affirm that Adomian’s decomposition method is a successful method for modeling thermoelastic problems. Moreover, the empirical parameter of the functional graded, and the lattice design parameter have significant effects on the temperature increment, the strain, the stress, the displacement.

Keywords: Adomian, Decomposition Method, Generalized Thermoelasticity, algorithm, empirical parameter, lattice design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 553
1163 Design and Development of a 3D Printed Myoelectric-Controlled Prosthesis Hand Using sEMG Sensor

Authors: Sher Shermin Azmiri Khan, Syeda Jannatul Ferdous, Sushmita Chakraborty

Abstract:

Over the last decades, biomedical engineering prosthetics become one of the most essential grounds. Prosthetic hands are rapidly evolving. Therefore, for designing prosthetic components, it is essential to improve quality such as make it affordable and improve patient comfort and mobility by making them lightweight and easy to wear. In this paper, we proposed a myoelectric controlled prosthesis hand. We can fabricate and manufacture customized cost-effective, small volumes of 3D printed hand which is interesting. The total weight of an adult hand is about 1000 gm including a battery. The prosthetic hand is built up with low-cost materials and techniques, the cost of manufacturing will be approximately US$145. The hand can grip objects of different shapes and sizes. The 3D printed hand can rotate its wrist like a human hand. The prosthetic hand is capable of showing some types of human gestures.

Keywords: Prosthetic Hand, sEMG, 3D printing, Arduino.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 579
1162 Conceptual Multidimensional Model

Authors: Manpreet Singh, Parvinder Singh, Suman

Abstract:

The data is available in abundance in any business organization. It includes the records for finance, maintenance, inventory, progress reports etc. As the time progresses, the data keep on accumulating and the challenge is to extract the information from this data bank. Knowledge discovery from these large and complex databases is the key problem of this era. Data mining and machine learning techniques are needed which can scale to the size of the problems and can be customized to the application of business. For the development of accurate and required information for particular problem, business analyst needs to develop multidimensional models which give the reliable information so that they can take right decision for particular problem. If the multidimensional model does not possess the advance features, the accuracy cannot be expected. The present work involves the development of a Multidimensional data model incorporating advance features. The criterion of computation is based on the data precision and to include slowly change time dimension. The final results are displayed in graphical form.

Keywords: Multidimensional, data precision.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1458
1161 Adaptive Gaussian Mixture Model for Skin Color Segmentation

Authors: Reza Hassanpour, Asadollah Shahbahrami, Stephan Wong

Abstract:

Skin color based tracking techniques often assume a static skin color model obtained either from an offline set of library images or the first few frames of a video stream. These models can show a weak performance in presence of changing lighting or imaging conditions. We propose an adaptive skin color model based on the Gaussian mixture model to handle the changing conditions. Initial estimation of the number and weights of skin color clusters are obtained using a modified form of the general Expectation maximization algorithm, The model adapts to changes in imaging conditions and refines the model parameters dynamically using spatial and temporal constraints. Experimental results show that the method can be used in effectively tracking of hand and face regions.

Keywords: Face detection, Segmentation, Tracking, Gaussian Mixture Model, Adaptation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2415
1160 Investigating Causes of Pavement Deterioration in Khartoum State, Sudan

Authors: Magdi M. E. Zumrawi

Abstract:

It is quite essential to investigate the causes of pavement deterioration in order to select the proper maintenance technique. The objective of this study was to identify factors cause deterioration of recently constructed roads in Khartoum state. A comprehensive literature concerning the factors of road deterioration, common road defects and their causes were reviewed. Three major road projects with different deterioration reasons were selected for this study. The investigation involved field survey and laboratory testing on those projects to examine the existing pavement conditions. The results revealed that the roads investigated experienced severe failures in the forms of cracks, potholes, and rutting in the wheel path. The causes of those failures were found mainly linked to poor drainage, traffic overloading, expansive subgrade soils, and the use of low quality materials in construction. Based on the results, recommendations were provided to help highway engineers in selecting the most effective repair techniques for specific kinds of distresses.

Keywords: Pavement, deterioration, causes, failures.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4976
1159 Line Balancing in the Hard Disk Drive Process Using Simulation Techniques

Authors: Teerapun Saeheaw, Nivit Charoenchai, Wichai Chattinnawat

Abstract:

Simulation model is an easy way to build up models to represent real life scenarios, to identify bottlenecks and to enhance system performance. Using a valid simulation model may give several advantages in creating better manufacturing design in order to improve the system performances. This paper presents result of implementing a simulation model to design hard disk drive manufacturing process by applying line balancing to improve both productivity and quality of hard disk drive process. The line balance efficiency showed 86% decrease in work in process, output was increased by an average of 80%, average time in the system was decreased 86% and waiting time was decreased 90%.

Keywords: line balancing, arena, hard disk drive process, simulation, work in process (WIP)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2164
1158 Improving the Effectiveness of Software Testing through Test Case Reduction

Authors: R. P. Mahapatra, Jitendra Singh

Abstract:

This paper proposes a new technique for improving the efficiency of software testing, which is based on a conventional attempt to reduce test cases that have to be tested for any given software. The approach utilizes the advantage of Regression Testing where fewer test cases would lessen time consumption of the testing as a whole. The technique also offers a means to perform test case generation automatically. Compared to one of the techniques in the literature where the tester has no option but to perform the test case generation manually, the proposed technique provides a better option. As for the test cases reduction, the technique uses simple algebraic conditions to assign fixed values to variables (Maximum, minimum and constant variables). By doing this, the variables values would be limited within a definite range, resulting in fewer numbers of possible test cases to process. The technique can also be used in program loops and arrays.

Keywords: Software Testing, Test Case Generation, Test CaseReduction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3017
1157 Study the Effect of Ultrasonic Irradiation and Surfactant/Fe ions Weight Ratio on Morphology and Particle Size of Magnetite Nanoparticles Synthesised by co-precipitation for Medical Application

Authors: S.Saloomeh Azimipour Meibod, Peyman Pourafshary, Hamid Reza Madaah Hosseini

Abstract:

A biocompatible ferrofluid have been prepared by coprecipitation of FeCl2.4H2O and FeCl3.6H2O under ultrasonic irradiation and with NaOH as alkaline agent. Cystein was also used as capping agent in the solution. Magnetic properties of the produced ferrofluid were then determined by VSM test and magnetite nanoparticles were characterized by XRD and TEM techniques. The effect of surfactant to Fe ion weight ratio was also studied during this project by using two different amount of Dextran. Results showed the presence of a biocompatible superparamagnetic ferrofluid including magnetite nanoparticles with particle size ranging under 20 nm. The increase in the surfactant content results in the narrowing of the size distribution and reduction of the particle size and more solution stability.

Keywords: Biocompatibility, Ferrofluid, Nanoparticle, Sizedistribution, Ultrasonic irradiation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1815
1156 Signature Identification Scheme Based on Iterated Function Systems

Authors: Nadia M. G. AL-Saidi

Abstract:

Since 1984 many schemes have been proposed for digital signature protocol, among them those that based on discrete log and factorizations. However a new identification scheme based on iterated function (IFS) systems are proposed and proved to be more efficient. In this study the proposed identification scheme is transformed into a digital signature scheme by using a one way hash function. It is a generalization of the GQ signature schemes. The attractor of the IFS is used to obtain public key from a private one, and in the encryption and decryption of a hash function. Our aim is to provide techniques and tools which may be useful towards developing cryptographic protocols. Comparisons between the proposed scheme and fractal digital signature scheme based on RSA setting, as well as, with the conventional Guillou-Quisquater signature, and RSA signature schemes is performed to prove that, the proposed scheme is efficient and with high performance.

Keywords: Digital signature, Fractal, Iterated function systems(IFS), Guillou-Quisquater (GQ) protocol, Zero-knowledge (ZK)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1514
1155 Issues and Architecture for Supporting Data Warehouse Queries in Web Portals

Authors: Minsoo Lee, Yoon-kyung Lee, Hyejung Yoon, Soo-kyung Song, Sujeong Cheong

Abstract:

Data Warehousing tools have become very popular and currently many of them have moved to Web-based user interfaces to make it easier to access and use the tools. The next step is to enable these tools to be used within a portal framework. The portal framework consists of pages having several small windows that contain individual data warehouse query results. There are several issues that need to be considered when designing the architecture for a portal enabled data warehouse query tool. Some issues need special techniques that can overcome the limitations that are imposed by the nature of data warehouse queries. Issues such as single sign-on, query result caching and sharing, customization, scheduling and authorization need to be considered. This paper discusses such issues and suggests an architecture to support data warehouse queries within Web portal frameworks.

Keywords: Data Warehousing tools, data warehousing queries, web portal frameworks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2121
1154 Statistical Optimization of Adsorption of a Harmful Dye from Aqueous Solution

Authors: M. Arun, A. Kannan

Abstract:

Textile industries cater to varied customer preferences and contribute substantially to the economy. However, these textile industries also produce a considerable amount of effluents. Prominent among these are the azo dyes which impart considerable color and toxicity even at low concentrations. Azo dyes are also used as coloring agents in food and pharmaceutical industry. Despite their applications, azo dyes are also notorious pollutants and carcinogens. Popular techniques like photo-degradation, biodegradation and the use of oxidizing agents are not applicable for all kinds of dyes, as most of them are stable to these techniques. Chemical coagulation produces a large amount of toxic sludge which is undesirable and is also ineffective towards a number of dyes. Most of the azo dyes are stable to UV-visible light irradiation and may even resist aerobic degradation. Adsorption has been the most preferred technique owing to its less cost, high capacity and process efficiency and the possibility of regenerating and recycling the adsorbent. Adsorption is also most preferred because it may produce high quality of the treated effluent and it is able to remove different kinds of dyes. However, the adsorption process is influenced by many variables whose inter-dependence makes it difficult to identify optimum conditions. The variables include stirring speed, temperature, initial concentration and adsorbent dosage. Further, the internal diffusional resistance inside the adsorbent particle leads to slow uptake of the solute within the adsorbent. Hence, it is necessary to identify optimum conditions that lead to high capacity and uptake rate of these pollutants. In this work, commercially available activated carbon was chosen as the adsorbent owing to its high surface area. A typical azo dye found in textile effluent waters, viz. the monoazo Acid Orange 10 dye (CAS: 1936-15-8) has been chosen as the representative pollutant. Adsorption studies were mainly focused at obtaining equilibrium and kinetic data for the batch adsorption process at different process conditions. Studies were conducted at different stirring speed, temperature, adsorbent dosage and initial dye concentration settings. The Full Factorial Design was the chosen statistical design framework for carrying out the experiments and identifying the important factors and their interactions. The optimum conditions identified from the experimental model were validated with actual experiments at the recommended settings. The equilibrium and kinetic data obtained were fitted to different models and the model parameters were estimated. This gives more details about the nature of adsorption taking place. Critical data required to design batch adsorption systems for removal of Acid Orange 10 dye and identification of factors that critically influence the separation efficiency are the key outcomes from this research.

Keywords: Acid Orange 10, Activated carbon, Optimum conditions, Statistical design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1353
1153 Applications of Artificial Neural Network to Building Statistical Models for Qualifying and Indexing Radiation Treatment Plans

Authors: Pei-Ju Chao, Tsair-Fwu Lee, Wei-Luen Huang, Long-Chang Chen, Te-Jen Su, Wen-Ping Chen

Abstract:

The main goal in this paper is to quantify the quality of different techniques for radiation treatment plans, a back-propagation artificial neural network (ANN) combined with biomedicine theory was used to model thirteen dosimetric parameters and to calculate two dosimetric indices. The correlations between dosimetric indices and quality of life were extracted as the features and used in the ANN model to make decisions in the clinic. The simulation results show that a trained multilayer back-propagation neural network model can help a doctor accept or reject a plan efficiently. In addition, the models are flexible and whenever a new treatment technique enters the market, the feature variables simply need to be imported and the model re-trained for it to be ready for use.

Keywords: neural network, dosimetric index, radiation treatment, tumor

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1690
1152 A Strategic Evaluation Approach for Defining the Maturity of Manufacturing Technologies

Authors: G. Reinhart, S. Schindler

Abstract:

Due to dynamic evolution, the ability of a manufacturing technology to produce a special product is changing. Therefore, it is essential to monitor the established techniques and processes to detect whether a company-s production will fit future circumstances. Concerning the manufacturing technology planning process, companies must decide when to change to a new technology for maintaining and increasing competitive advantages. In this context, the maturity assessment of the focused technologies is crucial. This article presents an approach for defining the maturity of a manufacturing technology from a strategic point of view. The concept is based on the approach of technology readiness level (TRL) according to NASA (National Aeronautics and Space Administration), but also includes dynamic changes. Therefore, the model takes into account the concept of the technology life cycle. Furthermore, it enables a company to estimate the ideal date for implementation of a new manufacturing technology.

Keywords: Maturity Assessment, Manufacturing Technology Planning, Technology Life Cycle, Technology Readiness Level.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2492
1151 The Significance of the Radiography Technique in the Non-Destructive Evaluation of the Integrity and Reliability of Cast Interconnects

Authors: Keshav Pujeri, Pranesh Jain, Krutibas Panda

Abstract:

Significant changes in oil and gas drilling have emphasized the need to verify the integrity and reliability of drill stem components. Defects are inevitable in cast components, regardless of application; but if these defects go undetected, any severe defect could cause down-hole failure. One such defect is shrinkage porosity. Castings with lower level shrinkage porosity (CB levels 1 and 2) have scattered pores and do not occupy large volumes; so pressure testing and helium leak testing (HLT) are sufficient for qualifying the castings. However, castings with shrinkage porosity of CB level 3 and higher, behave erratically under pressure testing and HLT making these techniques insufficient for evaluating the castings- integrity. This paper presents a case study to highlight how the radiography technique is much more effective than pressure testing and HLT.

Keywords: Casting Defects, Interconnects, Leak Check, Pressure Test, Radiography.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3514
1150 Making Ends Meet: The Challenges of Investing in and Accounting for Sustainability

Authors: György Á. Horváth, Piroska Harazin

Abstract:

The transition to sustainable development requires considerable investments from stakeholders, both financial and immaterial. However, accounting for such investments often poses a challenge, as ventures with intangible or non-financial returns remain oblivious to conventional accounting techniques and risk assessment. That such investments may significantly contribute to the welfare of those affected may act as a driving force behind attempting to bridge this gap. This gains crucial importance as investments must be also backed by governments and administrations; entities whose budget depends on taxpayers- contributions and whose tasks are based on securing the welfare of their citizens. Besides economic welfare, citizens also require social and environmental wellbeing too. However, administrations must also safeguard that welfare is guaranteed not only to present, but to future generations too. With already strained budgets and the requirement of sustainable development, governments on all levels face the double challenge of making both of these ends meet.

Keywords: Accounting, Administration and Government, RiskAssessment, Sustainable Development

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1332
1149 Performance Evaluation of an Online Text-Based Strategy Game

Authors: Nazleeni S. Haron, Mohd K. Zaime , Izzatdin A. Aziz, Mohd H. Hasan

Abstract:

Text-based game is supposed to be a low resource consumption application that delivers good performances when compared to graphical-intensive type of games. But, nowadays, some of the online text-based games are not offering performances that are acceptable to the users. Therefore, an online text-based game called Star_Quest has been developed in order to analyze its behavior under different performance measurements. Performance metrics such as throughput, scalability, response time and page loading time are captured to yield the performance of the game. The techniques in performing the load testing are also disclosed to exhibit the viability of our work. The comparative assessment between the results obtained and the accepted level of performances are conducted as to determine the performance level of the game. The study reveals that the developed game managed to meet all the performance objectives set forth.

Keywords: Online text-based games, performance evaluation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1609
1148 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling

Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra

Abstract:

Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.

Keywords: Multi-temporal satellite image, urban growth, Non-stationarity, stochastic modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1504
1147 Conversion of Methanol to Propylene over a High Silica B-HZSM-5 Catalyst

Authors: Aina Xu, Hongfang Ma, Haitao Zhang, Weiyong Ying, Dingye Fang

Abstract:

Hydrothermally synthesized high silica borosilicates with the MFI structure was subjected to several characterization techniques. The effect of boron on the structure and acidity of HZSM-5 catalyst were studied by XRD, SEM, N2 adsorption, solid state NMR, NH3-TPD. It was confirmed that boron had entered the framework in the boron samples. The results also revealed that strong acidity was weakened and weak acidity was strengthened by the boron added zeolite framework compared with parent catalyst. The catalytic performance was carried out in a fixed bed at 460°C for methanol to propylene (MTP) reaction. The results of MTP reaction showed a great increment of the propylene selectivity and excellent stability for the B-HZSM-5. The catalyst exhibited about 81% selectivity to C2 = - C4 = olefins with 40% selectivity of propylene as major component at near 100% methanol conversion, and the stable performance in the studied period was 100h.

Keywords: Methanol to propylene, HZSM-5, boron.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3567
1146 High Performance Computing Using Out-of- Core Sparse Direct Solvers

Authors: Mandhapati P. Raju, Siddhartha Khaitan

Abstract:

In-core memory requirement is a bottleneck in solving large three dimensional Navier-Stokes finite element problem formulations using sparse direct solvers. Out-of-core solution strategy is a viable alternative to reduce the in-core memory requirements while solving large scale problems. This study evaluates the performance of various out-of-core sequential solvers based on multifrontal or supernodal techniques in the context of finite element formulations for three dimensional problems on a Windows platform. Here three different solvers, HSL_MA78, MUMPS and PARDISO are compared. The performance of these solvers is evaluated on a 64-bit machine with 16GB RAM for finite element formulation of flow through a rectangular channel. It is observed that using out-of-core PARDISO solver, relatively large problems can be solved. The implementation of Newton and modified Newton's iteration is also discussed.

Keywords: Out-of-core, PARDISO, MUMPS, Newton.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2144
1145 Regression Test Selection Technique for Multi-Programming Language

Authors: Walid S. Abd El-hamid, Sherif S. El-Etriby, Mohiy M. Hadhoud

Abstract:

Regression testing is a maintenance activity applied to modified software to provide confidence that the changed parts are correct and that the unchanged parts have not been adversely affected by the modifications. Regression test selection techniques reduce the cost of regression testing, by selecting a subset of an existing test suite to use in retesting modified programs. This paper presents the first general regression-test-selection technique, which based on code and allows selecting test cases for any programs written in any programming language. Then it handles incomplete program. We also describe RTSDiff, a regression-test-selection system that implements the proposed technique. The results of the empirical studied that performed in four programming languages java, C#, Cµ and Visual basic show that the efficiency and effective in reducing the size of test suit.

Keywords: Regression testing, testing, test selection, softwareevolution, software maintenance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1533