Search results for: standard methodology.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3021

Search results for: standard methodology.

2721 Influence of Replacement Used Reference Coordinate System for Georeferencing of the Old Map of Europe

Authors: Jakub Havlicek, Jiri Cajthaml

Abstract:

The article describes the effect of the replacement of the used reference coordinate system in the georeferencing of an old map of Europe. The map was georeferenced into three types of projection – the equal-area conic (original cartographic projection), cylindrical Plate Carrée and cylindrical Mercator map projection. The map was georeferenced by means of the affine and the second-order polynomial transformation. The resulting georeferenced raster datasets from the Plate Carrée and Mercator projection were projected into the equal-area conic projection by means of projection equations. The output is the comparison of drawn graphics, the magnitude of standard deviations for individual projections and types of transformation.

Keywords: Georeferencing, reference coordinate system, transformation, standard deviation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543
2720 Towards Development of Solution for Business Process-Oriented Data Analysis

Authors: M. Klimavicius

Abstract:

This paper proposes a modeling methodology for the development of data analysis solution. The Author introduce the approach to address data warehousing issues at the at enterprise level. The methodology covers the process of the requirements eliciting and analysis stage as well as initial design of data warehouse. The paper reviews extended business process model, which satisfy the needs of data warehouse development. The Author considers that the use of business process models is necessary, as it reflects both enterprise information systems and business functions, which are important for data analysis. The Described approach divides development into three steps with different detailed elaboration of models. The Described approach gives possibility to gather requirements and display them to business users in easy manner.

Keywords: Data warehouse, data analysis, business processmanagement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1388
2719 Combination of Standard Secondary Raw Materials and New Production Waste Materials in Green Concrete Technology

Authors: M. Tazky, R. Hela, P. Novosad, L. Osuska

Abstract:

This paper deals with the possibility of safe incorporation fluidised bed combustion fly ash (waste material) into cement matrix together with next commonly used secondary raw material, which is high-temperature fly ash. Both of these materials have a very high pozzolanic ability, and the right combination could bring important improvements in both the physico-mechanical properties and the better durability of a cement composite. This paper tries to determine the correct methodology for designing green concrete by using modern methods measuring rheology of fresh concrete and following hydration processes. The use of fluidised bed combustion fly ash in cement composite production as an admixture is not currently common, but there are some real possibilities for its potential. The most striking negative aspect is its chemical composition which supports the development of new product formation, influencing the durability of the composite. Another disadvantage is the morphology of grains, which have a negative effect on consistency. This raises the question of how this waste can be used in concrete production to emphasize its positive properties and eliminate negatives. The focal point of the experiment carried out on cement pastes was particularly on the progress of hydration processes, aiming for the possible acceleration of pozzolanic reactions of both types of fly ash.

Keywords: High-temperature fly ash, fluidised bed combustion fly ash, pozzolanic, CaO (calcium oxide), rheology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 779
2718 A Mapping Approach of Code Generation for Arinc653-Based Avionics Software

Authors: Lu Zou, Dianfu MA, Ying Wang, Xianqi Zhao

Abstract:

Avionic software architecture has transit from a federated avionics architecture to an integrated modular avionics (IMA) .ARINC 653 (Avionics Application Standard Software Interface) is a software specification for space and time partitioning in Safety-critical avionics Real-time operating systems. Methods to transform the abstract avionics application logic function to the executable model have been brought up, however with less consideration about the code generating input and output model specific for ARINC 653 platform and inner-task synchronous dynamic interaction order sequence. In this paper, we proposed an AADL-based model-driven design methodology to fulfill the purpose to automatically generating Cµ executable model on ARINC 653 platform from the ARINC653 architecture which defined as AADL653 in order to facilitate the development of the avionics software constructed on ARINC653 OS. This paper presents the mapping rules between the AADL653 elements and the elements in Cµ language, and define the code generating rules , designs an automatic C µ code generator .Then, we use a case to illustrate our approach. Finally, we give the related work and future research directions.

Keywords: IMA, ARINC653, AADL653, code generation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3030
2717 Modeling of Coagulation Process for the Removal of Carbofuran in Aqueous Solution

Authors: Roli Saini, Pradeep Kumar

Abstract:

A coagulation/flocculation process was adopted for the reduction of carbamate insecticide (carbofuran) from aqueous solution. Ferric chloride (FeCl3) was used as a coagulant to treat the carbofuran. To exploit the reduction efficiency of pesticide concentration and COD, the jar-test experiments were carried out and process was optimized through response surface methodology (RSM). The effects of two independent factors; i.e., FeCl3 dosage and pH on the reduction efficiency were estimated by using central composite design (CCD). The initial COD of the 30 mg/L concentrated solution was found to be 510 mg/L. Results exposed that the maximum reduction occurred at an optimal condition of FeCl3 = 80 mg/L, and pH = 5.0, from which the reduction of concentration and COD 75.13% and 65.34%, respectively. The present study also predicted that the obtained regression equations could be helpful as the theoretical basis for the coagulation process of pesticide wastewater.

Keywords: Carbofuran, coagulation, optimization, response surface methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1100
2716 Newton-Raphson State Estimation Solution Employing Systematically Constructed Jacobian Matrix

Authors: Nursyarizal Mohd Nor, Ramiah Jegatheesan, Perumal Nallagownden

Abstract:

Newton-Raphson State Estimation method using bus admittance matrix remains as an efficient and most popular method to estimate the state variables. Elements of Jacobian matrix are computed from standard expressions which lack physical significance. In this paper, elements of the state estimation Jacobian matrix are obtained considering the power flow measurements in the network elements. These elements are processed one-by-one and the Jacobian matrix H is updated suitably in a simple manner. The constructed Jacobian matrix H is integrated with Weight Least Square method to estimate the state variables. The suggested procedure is successfully tested on IEEE standard systems.

Keywords: State Estimation (SE), Weight Least Square (WLS), Newton-Raphson State Estimation (NRSE), Jacobian matrix H.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2460
2715 Comparative Study of Evolutionary Model and Clustering Methods in Circuit Partitioning Pertaining to VLSI Design

Authors: K. A. Sumitra Devi, N. P. Banashree, Annamma Abraham

Abstract:

Partitioning is a critical area of VLSI CAD. In order to build complex digital logic circuits its often essential to sub-divide multi -million transistor design into manageable Pieces. This paper looks at the various partitioning techniques aspects of VLSI CAD, targeted at various applications. We proposed an evolutionary time-series model and a statistical glitch prediction system using a neural network with selection of global feature by making use of clustering method model, for partitioning a circuit. For evolutionary time-series model, we made use of genetic, memetic & neuro-memetic techniques. Our work focused in use of clustering methods - K-means & EM methodology. A comparative study is provided for all techniques to solve the problem of circuit partitioning pertaining to VLSI design. The performance of all approaches is compared using benchmark data provided by MCNC standard cell placement benchmark net lists. Analysis of the investigational results proved that the Neuro-memetic model achieves greater performance then other model in recognizing sub-circuits with minimum amount of interconnections between them.

Keywords: VLSI, circuit partitioning, memetic algorithm, genetic algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1633
2714 Optimization of NaOH Thermo-Chemical Pretreatment to Enhance Solubilisation of Organic Food Waste by Response Surface Methodology

Authors: H. Junoh, K. Palanisamy, C. H. Yip, F. L. Pua

Abstract:

This study investigates the influence of low temperature thermo-chemical pretreatment of organic food waste on performance of COD solubilisation. Both temperature and alkaline agent were reported to have effect on solubilizing any possible biomass including organic food waste. The three independent variables considered in this pretreatment were temperature (50-90oC), pretreatment time (30-120 minutes) and alkaline concentration, sodium hydroxide, NaOH (0.7-15 g/L). The maximal condition obtained were 90oC, 15 g/L NaOH for 2 hours. Solubilisation has potential in enhancing methane production by providing high amount of soluble components at early stage during anaerobic digestion.

Keywords: Food waste, pretreatments, respond surface methodology, ANOVA, anaerobic digestion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2282
2713 A Framework for Improving Trade Contractors’ Productivity Tracking Methods

Authors: Sophia Hayes, Kenny L. Liang, Sahil Sharma, Austin Shema, Mahmoud Bader, Mohamed Elbarkouky

Abstract:

Despite being one of the most significant economic contributors of the country, Canada’s construction industry is lagging behind other sectors when it comes to labor productivity improvements. The construction industry is very collaborative as a general contractor, will hire trade contractors to perform most of a project’s work; meaning low productivity from one contractor can have a domino effect on the shared success of a project. To address this issue and encourage trade contractors to improve their productivity tracking methods, an investigative study was done on the productivity views and tracking methods of various trade contractors. Additionally, an in-depth review was done on four standard tracking methods used in the construction industry: cost codes, benchmarking, the job productivity measurement (JPM) standard, and WorkFace Planning (WFP). The four tracking methods were used as a baseline in comparing the trade contractors’ responses, determining gaps within their current tracking methods, and for making improvement recommendations. 15 interviews were conducted with different trades to analyze how contractors value productivity. The results of these analyses indicated that there seem to be gaps within the construction industry when it comes to an understanding of the purpose and value in productivity tracking. The trade contractors also shared their current productivity tracking systems; which were then compared to the four standard tracking methods used in the construction industry. Gaps were identified in their various tracking methods and using a framework; recommendations were made based on the type of trade on how to improve how they track productivity.

Keywords: Trade contractors’ productivity, productivity tracking, cost codes, benchmarking, job productivity measurement, JPM, workface planning WFP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 881
2712 A Novel Design Methodology for a 1.5 KW DC/DC Converter in EV and Hybrid EV Applications

Authors: Farhan Beg

Abstract:

This paper presents a method for the efficient implementation of a unidirectional or bidirectional DC/DC converter. The DC/DC converter is used essentially for energy exchange between the low voltage service battery and a high voltage battery commonly found in Electric Vehicle applications. In these applications, apart from cost, efficiency of design is an important characteristic. A useful way to reduce the size of electronic equipment in the electric vehicles is proposed in this paper. The technique simplifies the mechanical complexity and maximizes the energy usage using the latest converter control techniques. Moreover a bidirectional battery charger for hybrid electric vehicles is also implemented in this paper. Several simulations on the test system have been carried out in Matlab/Simulink environment. The results exemplify the robustness of the proposed design methodology in case of a 1.5 KW DC-DC converter.

Keywords: DC-DC converters, Electric Vehicles, Direct Current Control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3693
2711 Information Security Risk Management in IT-Based Process Virtualization: A Methodological Design Based on Action Research

Authors: Jefferson Camacho Mejía, Jenny Paola Forero Pachón, Luis Carlos Gómez Flórez

Abstract:

Action research is a qualitative research methodology, which leads the researcher to delve into the problems of a community in order to understand its needs in depth and finally, to propose actions that lead to a change of social paradigm. Although this methodology had its beginnings in the human sciences, it has attracted increasing interest and acceptance in the field of information systems research since the 1990s. The countless possibilities offered nowadays by the use of Information Technologies (IT) in the development of different socio-economic activities have meant a change of social paradigm and the emergence of the so-called information and knowledge society. According to this, governments, large corporations, small entrepreneurs and in general, organizations of all kinds are using IT to virtualize their processes, taking them from the physical environment to the digital environment. However, there is a potential risk for organizations related with exposing valuable information without an appropriate framework for protecting it. This paper shows progress in the development of a methodological design to manage the information security risks associated with the IT-based processes virtualization, by applying the principles of the action research methodology and it is the result of a systematic review of the scientific literature. This design consists of seven fundamental stages. These are distributed in the three stages described in the action research methodology: 1) Observe, 2) Analyze and 3) Take actions. Finally, this paper aims to offer an alternative tool to traditional information security management methodologies with a view to being applied specifically in the planning stage of IT-based process virtualization in order to foresee risks and to establish security controls before formulating IT solutions in any type of organization.

Keywords: Action research, information security, information technology, methodological design, process virtualization, risk management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 944
2710 Harmonizing Spatial Plans: A Methodology to Integrate Sustainable Mobility and Energy Plans to Promote Resilient City Planning

Authors: B. Sanchez, D. Zambrana-Vasquez, J. Fresner, C. Krenn, F. Morea, L. Mercatelli

Abstract:

Local administrations are facing established targets on sustainable development from different disciplines at the heart of different city departments. Nevertheless, some of these targets, such as CO2 reduction, relate to two or more disciplines, as it is the case of sustainable mobility and energy plans (SUMP & SECAP/SEAP). This opens up the possibility to efficiently cooperate among different city departments and to create and develop harmonized spatial plans by using available resources and together achieving more ambitious goals in cities. The steps of the harmonization processes developed result in the identification of areas to achieve common strategic objectives. Harmonization, in other words, helps different departments in local authorities to work together and optimize the use or resources by sharing the same vision, involving key stakeholders, and promoting common data assessment to better optimize the resources. A methodology to promote resilient city planning via the harmonization of sustainable mobility and energy plans is presented in this paper. In order to validate the proposed methodology, a representative city engaged in an innovation process in efficient spatial planning is used as a case study. The harmonization process of sustainable mobility and energy plans covers identifying matching targets between different fields, developing different spatial plans with dual benefit and common indicators guaranteeing the continuous improvement of the harmonized plans. The proposed methodology supports local administrations in consistent spatial planning, considering both energy efficiency and sustainable mobility. Thus, municipalities can use their human and economic resources efficiently. This guarantees an efficient upgrade of land use plans integrating energy and mobility aspects in order to achieve sustainability targets, as well as to improve the wellbeing of its citizens.

Keywords: Harmonized planning, spatial planning, sustainable energy, sustainable mobility, SECAP, SUMP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 776
2709 The Size Effects of Keyboards (Keycaps) on Computer Typing Tasks

Authors: Chih-Chun Lai, Jun-Yu Wang

Abstract:

Keyboard is the most important equipment for computer tasks. However, improper design of keyboard would cause some symptoms like ulnar and/or radial deviations. The research goal of this study was to investigate the optimal size(s) of keycaps to increase efficiency. As shown in the questionnaire pre-study with 49 participants aged from 20 to 44, the most commonly used keyboards were 101-key standard keyboards. Most of the keycap sizes (W×L) were 1.3×1.5 cm and 1.5×1.5 cm. The fingertip breadths of most participants were 1.2 cm. Therefore, in the main study with 18 participants, a standard keyboard with each set of the 3-sized (1.2×1.4 cm, 1.3×1.5 cm, and 1.5×1.5 cm) keycaps were used to investigate their typing efficiency, respectively. The results revealed that the differences between the operating times for using 1.3×1.5 cm and 1.2×1.4 cm keycaps was insignificant while operating times for using 1.5×1.5cm keycaps were significantly longer than for using 1.2×1.4 cm or 1.3×1.5 cm, respectively. As for typing error rate, there was no significant difference.

Keywords: Keyboard, Keycap size, Typing efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1538
2708 Analysis of Effect of Pre-Logic Factoring on Cell Based Combinatorial Logic Synthesis

Authors: Padmanabhan Balasubramanian, Bashetty Raghavendra

Abstract:

In this paper, an analysis is presented, which demonstrates the effect pre-logic factoring could have on an automated combinational logic synthesis process succeeding it. The impact of pre-logic factoring for some arbitrary combinatorial circuits synthesized within a FPGA based logic design environment has been analyzed previously. This paper explores a similar effect, but with the non-regenerative logic synthesized using elements of a commercial standard cell library. On an overall basis, the results obtained pertaining to the analysis on a variety of MCNC/IWLS combinational logic benchmark circuits indicate that pre-logic factoring has the potential to facilitate simultaneous power, delay and area optimized synthesis solutions in many cases.

Keywords: Algebraic factoring, Combinational logic synthesis, Standard cells, Low power, Delay optimization, Area reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1372
2707 Objective Evaluation of Mathematical Morphology Edge Detection on Computed Tomography (CT) Images

Authors: Emhimed Saffor, Abdelkader Salama

Abstract:

In this paper problem of edge detection in digital images is considered. Edge detection based on morphological operators was applied on two sets (brain & chest) ct images. Three methods of edge detection by applying line morphological filters with multi structures in different directions have been used. 3x3 filter for first method, 5x5 filter for second method, and 7x7 filter for third method. We had applied this algorithm on (13 images) under MATLAB program environment. In order to evaluate the performance of the above mentioned edge detection algorithms, standard deviation (SD) and peak signal to noise ratio (PSNR) were used for justification for all different ct images. The objective method and the comparison of different methods of edge detection,  shows that high values of both standard deviation and PSNR values of edge detection images were obtained. 

Keywords: Medical images, Matlab, Edge detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2632
2706 Identity Verification Using k-NN Classifiers and Autistic Genetic Data

Authors: Fuad M. Alkoot

Abstract:

DNA data have been used in forensics for decades. However, current research looks at using the DNA as a biometric identity verification modality. The goal is to improve the speed of identification. We aim at using gene data that was initially used for autism detection to find if and how accurate is this data for identification applications. Mainly our goal is to find if our data preprocessing technique yields data useful as a biometric identification tool. We experiment with using the nearest neighbor classifier to identify subjects. Results show that optimal classification rate is achieved when the test set is corrupted by normally distributed noise with zero mean and standard deviation of 1. The classification rate is close to optimal at higher noise standard deviation reaching 3. This shows that the data can be used for identity verification with high accuracy using a simple classifier such as the k-nearest neighbor (k-NN). 

Keywords: Biometrics, identity verification, genetic data, k-nearest neighbor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1113
2705 New Explicit Group Newton's Iterative Methods for the Solutions of Burger's Equation

Authors: Tan K. B., Norhashidah Hj. M. Ali

Abstract:

In this article, we aim to discuss the formulation of two explicit group iterative finite difference methods for time-dependent two dimensional Burger-s problem on a variable mesh. For the non-linear problems, the discretization leads to a non-linear system whose Jacobian is a tridiagonal matrix. We discuss the Newton-s explicit group iterative methods for a general Burger-s equation. The proposed explicit group methods are derived from the standard point and rotated point Crank-Nicolson finite difference schemes. Their computational complexity analysis is discussed. Numerical results are given to justify the feasibility of these two proposed iterative methods.

Keywords: Standard point Crank-Nicolson (CN), Rotated point Crank-Nicolson (RCN), Explicit Group (EG), Explicit Decoupled Group (EDG).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1598
2704 The Assessment of Reforms in Different Countries by Social-Economic Development Integral Index

Authors: Samson Davoyan, Tatevik Sahakyan

Abstract:

The purpose of this report is to suggest a new methodology for the assessment of the comparative efficiency of the reforms made in different countries by an integral index. We have highlighted the reforms made in post-crisis period in 21 former socialist countries. The integral index describes the social-economic development level. The integral index contains of six indexes: The Global Competitiveness Index, Doing Business, The Corruption Perception, The Index of Economic Freedom, The Human Development, and The Democracy Index, which are reported by different international organizations. With the help of our methodology we first summarized the above-mentioned 6 indexes and attained 1 general index, besides, our new method enables us to assess the comparative efficiency of the reforms made in different countries by analyzing them. The purpose is to reveal the opportunities and threats of socialeconomic reforms in different directions.

Keywords: Assessment, comparative, effectiveness, reforms

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1431
2703 Hybridizing Genetic Algorithm with Biased Chance Local Search

Authors: Mehdi Basikhasteh, Mohamad A. Movafaghpour

Abstract:

This paper explores university course timetabling problem. There are several characteristics that make scheduling and timetabling problems particularly difficult to solve: they have huge search spaces, they are often highly constrained, they require sophisticated solution representation schemes, and they usually require very time-consuming fitness evaluation routines. Thus standard evolutionary algorithms lack of efficiency to deal with them. In this paper we have proposed a memetic algorithm that incorporates the problem specific knowledge such that most of chromosomes generated are decoded into feasible solutions. Generating vast amount of feasible chromosomes makes the progress of search process possible in a time efficient manner. Experimental results exhibit the advantages of the developed Hybrid Genetic Algorithm than the standard Genetic Algorithm.

Keywords: University Course Timetabling, Memetic Algorithm, Biased Chance Assignment, Optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1693
2702 Surface Water Pollution by Open Refuse Dumpsite in North Central of Nigeria

Authors: Abimbola Motunrayo Folami, Ibironke Titilayo Enitan, Feroz Mohomed Swalaha

Abstract:

Water is a vital resource that is important in ensuring the growth and development of any country. To sustain the basic human needs and the demands for agriculture, industry, conservational and ecosystem, enough quality and quantity water is needed. Contamination of water resources is now a global and public health concern. Hence, this study assessed the water quality of Ndawuse River by measuring the physicochemical parameters and heavy metals concentrations of the river using standard methods. In total, 16 surface water samples were obtained from five locations along the river, from upstream to downstream as well as samples from the dumpsite. The results obtained were compared with the standard limits set by both the World Health Organization and the Federal Environmental Protection Agency for domestic purposes. The results of the measured parameters indicated that biological oxygen demand (85.88 mg/L), turbidity (44.51 NTU), Iron (0.014 - 3.511 mg /L) and chromium (0.078 - 0.14 mg /L) were all above the standard limits. The results further showed that the quality of surface water is being significantly affected by human activities around the Ndawuse River which could pose an adverse health risk to several communities that rely on this river as their primary source of water. Therefore, there is a need for strict enforcement of environmental laws to protect the aquatic ecosystem and to avoid long term cumulative exposure risk that heavy metals may pose on human health.

Keywords: Abuja, contaminants, heavy metals, Ndawuse River, Nigeria, surface water.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 596
2701 Determination of Volatile Organic Compounds in Human Breath by Optical Fiber Sensing

Authors: C. I. L. Justino, L. I. B. Silva, K. Duarte, A. C. Freitas, T. A. P. Rocha-Santos, A. C. Duarte

Abstract:

This work proposes an optical fiber system (OF) for sensing various volatile organic compounds (VOCs) in human breath for the diagnosis of some metabolic disorders as a non-invasive methodology. The analyzed VOCs are alkanes (i.e., ethane, pentane, heptane, octane, and decane), and aromatic compounds (i.e., benzene, toluene, and styrene). The OF displays high analytical performance since it provides near real-time responses, rapid analysis, and low instrumentation costs, as well as it exhibits useful linear range and detection limits; the developed OF sensor is also comparable to a reference methodology (gas chromatography-mass spectrometry) for the eight tested VOCs.

Keywords: Breath analysis, gas chromatography-mass spectrometry, optical fiber sensor, volatile organic compounds

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2294
2700 Six Sigma-Based Optimization of Shrinkage Accuracy in Injection Molding Processes

Authors: Sky Chou, Joseph C. Chen

Abstract:

This paper focuses on using six sigma methodologies to reach the desired shrinkage of a manufactured high-density polyurethane (HDPE) part produced by the injection molding machine. It presents a case study where the correct shrinkage is required to reduce or eliminate defects and to improve the process capability index Cp and Cpk for an injection molding process. To improve this process and keep the product within specifications, the six sigma methodology, design, measure, analyze, improve, and control (DMAIC) approach, was implemented in this study. The six sigma approach was paired with the Taguchi methodology to identify the optimized processing parameters that keep the shrinkage rate within the specifications by our customer. An L9 orthogonal array was applied in the Taguchi experimental design, with four controllable factors and one non-controllable/noise factor. The four controllable factors identified consist of the cooling time, melt temperature, holding time, and metering stroke. The noise factor is the difference between material brand 1 and material brand 2. After the confirmation run was completed, measurements verify that the new parameter settings are optimal. With the new settings, the process capability index has improved dramatically. The purpose of this study is to show that the six sigma and Taguchi methodology can be efficiently used to determine important factors that will improve the process capability index of the injection molding process.

Keywords: Injection molding, shrinkage, six sigma, Taguchi parameter design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1372
2699 Development of Split Air-Conditioning System using Chill Water as the Cooling Medium

Authors: Faezzan Madli, Zamri Noranai, Md. Norrizam Mohamad Jaat, Hamidon Salleh, Mohammad Zainal Md Yusof

Abstract:

Current air conditioning system is using refrigerant as the cooling medium. The main purpose of this study is to develop an air conditioning system using chill water as the cooling medium. In this system, chill water used to replace refrigerant as the cooling medium. This study is focus on the split type unit air conditioning system only. It will be involving some renovation on the indoor unit and freezer. The cooling capability of this system was validate by few series of testing, which conducted at standard 36m3 office room. Result of the testing found that 0.1 m3 of chill water is able to maintain the room temperature within standard up to 4 ~ 8 hours. It expected able to maintain room temperature up to 10 hour with some improvement.

Keywords: Chill water, air-condition, office room.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2865
2698 Variation of the Dynamic Characteristics of a Spindle with the Change of Bearing Preload

Authors: Shinji Oouchi, Hajime Nomura, Kung-Da Wu, Yong-Run Chen, Jui-Pin Hung

Abstract:

This paper presents the variation of the dynamic characteristics of a spindle with the change of bearing preload. The correlations between the variation of bearing preload and fundamental modal parameters were first examined by conducting vibration tests on physical spindle units. Experimental measurements show that the dynamic compliance and damping ratio associated with the dominating modes were affected to vary with variation of the bearing preload. When the bearing preload was slightly deviated from a standard value, the modal frequency and damping ability also vary to different extent, which further enable the spindle to perform with different compliance. For the spindle used in this study, a standard preload value set on bearings would enable the spindle to behave a higher stiffness as compared with others with a preload variation. This characteristic can be served as a reference to examine the variation of bearing preload of spindle in assemblage or operation.

Keywords: Dynamic compliance, Bearing preload, Modal damping.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2394
2697 Knowledge Acquisition for the Construction of an Evolving Ontology: Application to Augmented Surgery

Authors: Nora Taleb, Sellami Mokhtar, Michel Simonet

Abstract:

This work concerns the evolution and the maintenance of an ontological resource in relation with the evolution of the corpus of texts from which it had been built. The knowledge forming a text corpus, especially in dynamic domains, is in continuous evolution. When a change in the corpus occurs, the domain ontology must evolve accordingly. Most methods manage ontology evolution independently from the corpus from which it is built; in addition, they treat evolution just as a process of knowledge addition, not considering other knowledge changes. We propose a methodology for managing an evolving ontology from a text corpus that evolves over time, while preserving the consistency and the persistence of this ontology. Our methodology is based on the changes made on the corpus to reflect the evolution of the considered domain - augmented surgery in our case. In this context, the results of text mining techniques, as well as the ARCHONTE method slightly modified, are used to support the evolution process.

Keywords: Corpus, Evolution, Ontology

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1437
2696 Computer-aided Sequence Planning of Shearing Operations in Progressive Dies

Authors: Alan C. Lin, Dean K. Sheu

Abstract:

This paper aims to study the methodology of building the knowledge of planning adequate punches in order to complete the task of strip layout for shearing processes, using progressive dies. The proposed methodology uses die design rules and characteristics of different types of punches to classify them into five groups: prior use (the punches must be used first), posterior use (must be used last), compatible use (may be used together), sequential use (certain punches must precede some others) and simultaneous use (must be used together). With these five groups of punches, the searching space of feasible designs will be greatly reduced, and superimposition becomes a more effective method of punch layout. The superimposition scheme will generate many feasible solutions, an evaluation function based on number of stages, moment balancing and strip stability is developed for helping designers to find better solutions.

Keywords: Manufacturing systems, advances in metal forming, computer-aided design, progressive die.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2183
2695 Event Related Potentials in Terms of Visual and Auditory Stimuli

Authors: Seokbeen Lim, KyeongSeok Sim, DaKyeong Shin, Gilwon Yoon

Abstract:

Event-related potential (ERP) is one of the useful tools for investigating cognitive reactions. In this study, the potential of ERP components detected after auditory and visual stimuli was examined. Subjects were asked to respond upon stimuli that were of three categories; Target, Non-Target and Standard stimuli. The ERP after stimulus was measured. In the experiment of visual evoked potentials (VEPs), the subjects were asked to gaze at a center point on the monitor screen where the stimuli were provided by the reversal pattern of the checkerboard. In consequence of the VEP experiments, we observed consistent reactions. Each peak voltage could be measured when the ensemble average was applied. Visual stimuli had smaller amplitude and a longer latency compared to that of auditory stimuli. The amplitude was the highest with Target and the smallest with Standard in both stimuli.

Keywords: Auditory stimulus, EEG, event related potential, oddball task, visual stimulus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1246
2694 Gluten-Free Cookies Enriched with Blueberry Pomace: Optimization of Baking Process

Authors: Aleksandra Mišan, Bojana Šarić, Nataša Nedeljković, Mladenka Pestorić, Pavle Jovanov, Milica Pojić, Jelena Tomić, Bojana Filipčev, Miroslav Hadnađev, Anamarija Mandić

Abstract:

With the aim of improving nutritional profile and antioxidant capacity of gluten-free cookies, blueberry pomace, by-product of juice production, was processed into a new food ingredient by drying and grinding and used for a gluten-free cookie formulation. Since the quality of a baked product is highly influenced by the baking conditions, the objective of this work was to optimize the baking time and thickness of dough pieces, by applying Response Surface Methodology (RSM) in order to obtain the best technological quality of the cookies. The experiments were carried out according to a Central Composite Design (CCD) by selecting the dough thickness and baking time as independent variables, while hardness, color parameters (L*, a* and b* values), water activity, diameter and short/long ratio were response variables. According to the results of RSM analysis, the baking time of 13.74min and dough thickness of 4.08mm was found to be the optimal for the baking temperature of 170°C. As similar optimal parameters were obtained by previously conducted experiment based on sensory analysis, response surface methodology (RSM) can be considered as a suitable approach to optimize the baking process.

Keywords: Baking process, blueberry pomace, gluten-free cookies, Response Surface Methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2569
2693 Physical and Mechanical Properties of Particleboard from Bamboo Waste

Authors: Vanchai Laemlaksakul

Abstract:

This research was to evaluate a technical feasibility of making single-layer experimental particleboard panels from bamboo waste (Dendrocalamus asper Backer) by converting bamboo into strips, which are used to make laminated bamboo furniture. Variable factors were density (600, 700 and 800 kg/m3) and temperature of condition (25, 40 and 55 °C). The experimental panels were tested for their physical and mechanical properties including modulus of elasticity (MOE), modulus of rupture (MOR), internal bonding strength (IB), screw holding strength (SH) and thickness swelling values according to the procedures defined by Japanese Industrial Standard (JIS). The test result of mechanical properties showed that the MOR, MOE and IB values were not in the set criteria, except the MOR values at the density of 700 kg/m3 at 25 °C and at the density of 800 kg/m3 at 25 and 40 °C, the IB values at the density of 600 kg/m3, at 40 °C, and at the density of 800 kg/m3 at 55 °C. The SH values had the test result according to the set standard, except with the density of 600 kg/m3, at 40 and 55 °C. Conclusively, a valuable renewable biomass, bamboo waste could be used to manufacture boards.

Keywords: Particleboard, Urea Formaldehyde Resin, BambooWaste

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5636
2692 The Improvement of Environmental Protection through Motor Vehicle Noise Abatement

Authors: Z. Jovanovic, Z. Masonicic, S. Dragutinovic, Z. Sakota

Abstract:

In this paper, a methodology for noise reduction of motor vehicles in use is presented. The methodology relies on synergic model of noise generation as a function of time. The arbitrary number of motor vehicle noise sources act in concert yielding the generation of the overall noise level of motor vehicle thereafter. The number of noise sources participating in the overall noise level of motor vehicle is subjected to the constraint of the calculation of the acoustic potential of each noise source under consideration. It is the prerequisite condition for the calculation of the acoustic potential of the whole vehicle. The recast form of pertinent set of equations describing the synergic model is laid down and solved by dint of Gauss method. The bunch of results emerged and some of them i.e. those ensuing from model application to MDD FAP Priboj motor vehicle in use are particularly elucidated.

Keywords: Noise abatement, MV noise sources, noise source identification, muffler.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1288