Search results for: Markovian Decision Process based Adaptive Scheduling
13605 An Evaluation of Barriers to Implement Reverse Logistics: A Case Study of Indian Fastener Industry
Authors: D. Garg, S. Luthra, A. Haleem
Abstract:
Reverse logistics (RL) is supposed to be a systematic procedure that helps in improving the environmental hazards and maintain business sustainability for industries. Industries in Indian are now opting for adoption of RL techniques in business. But, RL practices are not popular in Indian industries because of many barriers for its successful implementation. Therefore, need arises to identify and evaluate the barriers to implement RL practices by taking an Indian industries perspective. Literature review approach and case study approach have been adapted to identify relevant barriers to implement RL practices. Further, Fuzzy Decision Making Trial and Evaluation Laboratory methodology has been brought into use for evaluating causal relationships among the barriers to implement RL practices. Seven barriers out of ten barriers have been categorized into the cause group and remaining into effect group. This research will help Indian industries to manage these barriers towards effective implementing RL practices.Keywords: Barriers, decision making trial and evaluation laboratory, fuzzy set theory, Indian industries, reverse logistics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 219013604 Ontology Population via NLP Techniques in Risk Management
Authors: Jawad Makki, Anne-Marie Alquier, Violaine Prince
Abstract:
In this paper we propose an NLP-based method for Ontology Population from texts and apply it to semi automatic instantiate a Generic Knowledge Base (Generic Domain Ontology) in the risk management domain. The approach is semi-automatic and uses a domain expert intervention for validation. The proposed approach relies on a set of Instances Recognition Rules based on syntactic structures, and on the predicative power of verbs in the instantiation process. It is not domain dependent since it heavily relies on linguistic knowledge. A description of an experiment performed on a part of the ontology of the PRIMA1 project (supported by the European community) is given. A first validation of the method is done by populating this ontology with Chemical Fact Sheets from Environmental Protection Agency2. The results of this experiment complete the paper and support the hypothesis that relying on the predicative power of verbs in the instantiation process improves the performance.Keywords: Information Extraction, Instance Recognition Rules, Ontology Population, Risk Management, Semantic analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 153513603 An Energy-Efficient Protocol with Static Clustering for Wireless Sensor Networks
Authors: Amir Sepasi Zahmati, Bahman Abolhassani, Ali Asghar Beheshti Shirazi, Ali Shojaee Bakhtiari
Abstract:
A wireless sensor network with a large number of tiny sensor nodes can be used as an effective tool for gathering data in various situations. One of the major issues in wireless sensor networks is developing an energy-efficient routing protocol which has a significant impact on the overall lifetime of the sensor network. In this paper, we propose a novel hierarchical with static clustering routing protocol called Energy-Efficient Protocol with Static Clustering (EEPSC). EEPSC, partitions the network into static clusters, eliminates the overhead of dynamic clustering and utilizes temporary-cluster-heads to distribute the energy load among high-power sensor nodes; thus extends network lifetime. We have conducted simulation-based evaluations to compare the performance of EEPSC against Low-Energy Adaptive Clustering Hierarchy (LEACH). Our experiment results show that EEPSC outperforms LEACH in terms of network lifetime and power consumption minimization.Keywords: Clustering methods, energy efficiency, routingprotocol, wireless sensor networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 272113602 The Mechanistic and Oxidative Study of Methomyl and Parathion Degradation by Fenton Process
Authors: Chihhao Fan, Ming-Chu Liao
Abstract:
The purpose of this study is to investigate the chemical degradation of the organophosphorus pesticide of parathion and carbamate insecticide of methomyl in the aqueous phase through Fenton process. With the employment of batch Fenton process, the degradation of the two selected pesticides at different pH, initial concentration, humic acid concentration, and Fenton reagent dosages was explored. The Fenton process was found effective to degrade parathion and methomyl. The optimal dosage of Fenton reagents (i.e., molar concentration ratio of H2O2 to Fe2+) at pH 7 for parathion degradation was equal to 3, which resulted in 50% removal of parathion. Similarly, the optimal dosage for methomyl degradation was 1, resulting in 80% removal of methomyl. This study also found that the presence of humic substances has enhanced pesticide degradation by Fenton process significantly. The mass spectroscopy results showed that the hydroxyl free radical may attack the single bonds with least energy of investigated pesticides to form smaller molecules which is more easily to degrade either through physio-chemical or bilolgical processes.Keywords: Fenton Process, humic acid, methomyl, parathion, pesticides
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 163013601 A Bathtub Curve from Nonparametric Model
Authors: Eduardo C. Guardia, Jose W. M. Lima, Afonso H. M. Santos
Abstract:
This paper presents a nonparametric method to obtain the hazard rate “Bathtub curve” for power system components. The model is a mixture of the three known phases of a component life, the decreasing failure rate (DFR), the constant failure rate (CFR) and the increasing failure rate (IFR) represented by three parametric Weibull models. The parameters are obtained from a simultaneous fitting process of the model to the Kernel nonparametric hazard rate curve. From the Weibull parameters and failure rate curves the useful lifetime and the characteristic lifetime were defined. To demonstrate the model the historic time-to-failure of distribution transformers were used as an example. The resulted “Bathtub curve” shows the failure rate for the equipment lifetime which can be applied in economic and replacement decision models.
Keywords: Bathtub curve, failure analysis, lifetime estimation, parameter estimation, Weibull distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 223413600 Drop Impact on a Vibrated, Heated Surface: Towards a Potential New Way of Elaborating Nuclear Fuel from Gel Microspheres
Authors: Méryl Brothier, Dominique Moulinier, Christophe Bertaux
Abstract:
The gel-supported precipitation (GSP) process can be used to make spherical particles (spherules) of nuclear fuel, particularly for very high temperature reactors (VHTR) and even for implementing the process called SPHEREPAC. In these different cases, the main characteristics are the sphericity of the particles to be manufactured and the control over their grain size. Nonetheless, depending on the specifications defined for these spherical particles, the GSP process has intrinsic limits, particularly when fabricating very small particles. This paper describes the use of secondary fragmentation (water, water/PVA and uranyl nitrate) on solid surfaces under varying temperature and vibration conditions to assess the relevance of using this new technique to manufacture very small spherical particles by means of a modified GSP process. The fragmentation mechanisms are monitored and analysed, before the trends for its subsequent optimised application are described.Keywords: Microsphere elaboration, nuclear fuel, droplet impact , gel-supported precipitation process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 159013599 Monitoring the Drying and Grinding Process during Production of Celitement through a NIR-Spectroscopy Based Approach
Authors: Carolin Lutz, Jörg Matthes, Patrick Waibel, Ulrich Precht, Krassimir Garbev, Günter Beuchle, Uwe Schweike, Peter Stemmermann, Hubert B. Keller
Abstract:
Online measurement of the product quality is a challenging task in cement production, especially in the production of Celitement, a novel environmentally friendly hydraulic binder. The mineralogy and chemical composition of clinker in ordinary Portland cement production is measured by X-ray diffraction (XRD) and X-ray fluorescence (XRF), where only crystalline constituents can be detected. But only a small part of the Celitement components can be measured via XRD, because most constituents have an amorphous structure. This paper describes the development of algorithms suitable for an on-line monitoring of the final processing step of Celitement based on NIR-data. For calibration intermediate products were dried at different temperatures and ground for variable durations. The products were analyzed using XRD and thermogravimetric analyses together with NIR-spectroscopy to investigate the dependency between the drying and the milling processes on one and the NIR-signal on the other side. As a result, different characteristic parameters have been defined. A short overview of the Celitement process and the challenging tasks of the online measurement and evaluation of the product quality will be presented. Subsequently, methods for systematic development of near-infrared calibration models and the determination of the final calibration model will be introduced. The application of the model on experimental data illustrates that NIR-spectroscopy allows for a quick and sufficiently exact determination of crucial process parameters.Keywords: Calibration model, celitement, cementitious material, NIR spectroscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 173013598 Quality-Driven Business Process Refactoring
Authors: María Fernández-Ropero, Ricardo Pérez-Castillo, Ismael Caballero, Mario Piattini
Abstract:
Appropriate description of business processes through standard notations has become one of the most important assets for organizations. Organizations must therefore deal with quality faults in business process models such as the lack of understandability and modifiability. These quality faults may be exacerbated if business process models are mined by reverse engineering, e.g., from existing information systems that support those business processes. Hence, business process refactoring is often used, which change the internal structure of business processes whilst its external behavior is preserved. This paper aims to choose the most appropriate set of refactoring operators through the quality assessment concerning understandability and modifiability. These quality features are assessed through well-proven measures proposed in the literature. Additionally, a set of measure thresholds are heuristically established for applying the most promising refactoring operators, i.e., those that achieve the highest quality improvement according to the selected measures in each case.Keywords: business process model, modifiability, refactoring, understandability
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 152813597 Evaluation of Optimum Performance of Lateral Intakes
Authors: Mohammad Reza Pirestani, Hamid Reza Vosoghifar, Pegah Jazayeri
Abstract:
In designing river intakes and diversion structures, it is paramount that the sediments entering the intake are minimized or, if possible, completely separated. Due to high water velocity, sediments can significantly damage hydraulic structures especially when mechanical equipment like pumps and turbines are used. This subsequently results in wasting water, electricity and further costs. Therefore, it is prudent to investigate and analyze the performance of lateral intakes affected by sediment control structures. Laboratory experiments, despite their vast potential and benefits, can face certain limitations and challenges. Some of these include: limitations in equipment and facilities, space constraints, equipment errors including lack of adequate precision or mal-operation, and finally, human error. Research has shown that in order to achieve the ultimate goal of intake structure design – which is to design longlasting and proficient structures – the best combination of sediment control structures (such as sill and submerged vanes) along with parameters that increase their performance (such as diversion angle and location) should be determined. Cost, difficulty of execution and environmental impacts should also be included in evaluating the optimal design. This solution can then be applied to similar problems in the future. Subsequently, the model used to arrive at the optimal design requires high level of accuracy and precision in order to avoid improper design and execution of projects. Process of creating and executing the design should be as comprehensive and applicable as possible. Therefore, it is important that influential parameters and vital criteria is fully understood and applied at all stages of choosing the optimal design. In this article, influential parameters on optimal performance of the intake, advantages and disadvantages, and efficiency of a given design are studied. Then, a multi-criterion decision matrix is utilized to choose the optimal model that can be used to determine the proper parameters in constructing the intake.
Keywords: Diversion structures lateral intake, multi criteria decision making, optimal design, sediment control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 223113596 Study on Clarification of the Core Technology in a Monozukuri Company
Authors: Nishiyama Toshiaki, Tadayuki Kyountani, Nguyen Huu Phuc, Shigeyuki Haruyama, Oke Oktavianty
Abstract:
It is important to clarify the company’s core technology in product development process to strengthen their power in providing technology that meets the customer requirement. QFD method is adopted to clarify the core technology through identifying the high element technologies that are related to the voice of customer, and offer the most delightful features for customer. AHP is used to determine the importance of evaluating factors. A case study was conducted by using this approach in Japan’s Monozukuri Company (so called manufacturing company) to clarify their core technology based on customer requirements.Keywords: QFD, product development process, core technology, AHP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 110013595 Adaptation of Iterative Methods to Solve Fuzzy Mathematical Programming Problems
Authors: Ricardo C. Silva, Luiza A. P. Cantao, Akebo Yamakami
Abstract:
Based on the fuzzy set theory this work develops two adaptations of iterative methods that solve mathematical programming problems with uncertainties in the objective function and in the set of constraints. The first one uses the approach proposed by Zimmermann to fuzzy linear programming problems as a basis and the second one obtains cut levels and later maximizes the membership function of fuzzy decision making using the bound search method. We outline similarities between the two iterative methods studied. Selected examples from the literature are presented to validate the efficiency of the methods addressed.Keywords: Fuzzy Theory, Nonlinear Optimization, Fuzzy Mathematics Programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162313594 On Speeding Up Support Vector Machines: Proximity Graphs Versus Random Sampling for Pre-Selection Condensation
Authors: Xiaohua Liu, Juan F. Beltran, Nishant Mohanchandra, Godfried T. Toussaint
Abstract:
Support vector machines (SVMs) are considered to be the best machine learning algorithms for minimizing the predictive probability of misclassification. However, their drawback is that for large data sets the computation of the optimal decision boundary is a time consuming function of the size of the training set. Hence several methods have been proposed to speed up the SVM algorithm. Here three methods used to speed up the computation of the SVM classifiers are compared experimentally using a musical genre classification problem. The simplest method pre-selects a random sample of the data before the application of the SVM algorithm. Two additional methods use proximity graphs to pre-select data that are near the decision boundary. One uses k-Nearest Neighbor graphs and the other Relative Neighborhood Graphs to accomplish the task.Keywords: Machine learning, data mining, support vector machines, proximity graphs, relative-neighborhood graphs, k-nearestneighbor graphs, random sampling, training data condensation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 191913593 Language and Retrieval Accuracy
Authors: Ahmed Abdelali, Jim Cowie, Hamdy S. Soliman
Abstract:
One of the major challenges in the Information Retrieval field is handling the massive amount of information available to Internet users. Existing ranking techniques and strategies that govern the retrieval process fall short of expected accuracy. Often relevant documents are buried deep in the list of documents returned by the search engine. In order to improve retrieval accuracy we examine the issue of language effect on the retrieval process. Then, we propose a solution for a more biased, user-centric relevance for retrieved data. The results demonstrate that using indices based on variations of the same language enhances the accuracy of search engines for individual users.Keywords: Information Search and Retrieval, LanguageVariants, Search Engine, Retrieval Accuracy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 147713592 Variable Regularization Parameter Normalized Least Mean Square Adaptive Filter
Authors: Young-Seok Choi
Abstract:
We present a normalized LMS (NLMS) algorithm with robust regularization. Unlike conventional NLMS with the fixed regularization parameter, the proposed approach dynamically updates the regularization parameter. By exploiting a gradient descent direction, we derive a computationally efficient and robust update scheme for the regularization parameter. In simulation, we demonstrate the proposed algorithm outperforms conventional NLMS algorithms in terms of convergence rate and misadjustment error.Keywords: Regularization, normalized LMS, system identification, robustness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 187613591 Dynamic Network Routing Method Based on Chromosome Learning
Authors: Xun Liang
Abstract:
In this paper, we probe into the traffic assignment problem by the chromosome-learning-based path finding method in simulation, which is to model the driver' behavior in the with-in-a-day process. By simply making a combination and a change of the traffic route chromosomes, the driver at the intersection chooses his next route. The various crossover and mutation rules are proposed with extensive examples.
Keywords: Chromosome learning, crossover, mutation, traffic path finding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 134813590 Standardization and Adaption Requirements in Production System Transplants
Authors: G. Schuh, T. Potente, D. Kupke, S. Ivanescu
Abstract:
As German companies roll out their standardized production systems to offshore manufacturing plants, they face the challenge of implementing them in different cultural environments. Studies show that the local adaptation is one of the key factors for a successful implementation. Thus the question arises of where the line between standardization and adaptation can be drawn. To answer this question the influence of culture on production systems is analysed in this paper. The culturally contingent components of production systems are identified. Also the contingency factors are classified according to their impact on the necessary adaptation changes and implementation effort. Culturally specific decision making, coordination, communication and motivation patterns require one-time changes in organizational and process design. The attitude towards rules requires more intense coaching and controlling. Lastly a framework is developed to depict standardization and adaption needs when transplanting production systems into different cultural environments.Keywords: Culture, influence of national culture on production systems, lean production, production systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 182513589 Momentum Accounting in Public Management: A Case Study in a Brazilian Navy-s Services Provider Military Organization
Authors: Rodrigo Barreiros Leal, Aracéli Cristina de Sousa Ferreira
Abstract:
This study examines the possibility to apply the theory of multidimensional accounting (momentum accounting) in a Brazilian Navy-s Services Provider Military Organization (Organização Militar Prestadora de Serviços - OMPS). In general, the core of the said theory is the fact that Accounting does not recognize the inertia of transactions occurring in an entity, and that occur repeatedly in some cases, regardless of the implementation of new actions by its managers. The study evaluates the possibility of greater use of information recorded in the financial statements of the unit of analysis, within the strategic decisions of the organization. As a research strategy, we adopted the case study. The results infer that it is possible to use the theory in the context of a multidimensional OMPS, promoting useful information for decision-making and thereby contributing to the strengthening of the necessary alignment of its administration with the current desires of the Brazilian society.
Keywords: Multidimensional Accounting, Public Management, Decision Making.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 287913588 Application of Griddization Management to Construction Hazard Management
Authors: Lingzhi Li, Jiankun Zhang, Tiantian Gu
Abstract:
Hazard management that can prevent fatal accidents and property losses is a fundamental process during the buildings’ construction stage. However, due to lack of safety supervision resources and operational pressures, the conduction of hazard management is poor and ineffective in China. In order to improve the quality of construction safety management, it is critical to explore the use of information technologies to ensure that the process of hazard management is efficient and effective. After exploring the existing problems of construction hazard management in China, this paper develops the griddization management model for construction hazard management. First, following the knowledge grid infrastructure, the griddization computing infrastructure for construction hazards management is designed which includes five layers: resource entity layer, information management layer, task management layer, knowledge transformation layer and application layer. This infrastructure will be as the technical support for realizing grid management. Second, this study divides the construction hazards into grids through city level, district level and construction site level according to grid principles. Last, a griddization management process including hazard identification, assessment and control is developed. Meanwhile, all stakeholders of construction safety management, such as owners, contractors, supervision organizations and government departments, should take the corresponding responsibilities in this process. Finally, a case study based on actual construction hazard identification, assessment and control is used to validate the effectiveness and efficiency of the proposed griddization management model. The advantage of this designed model is to realize information sharing and cooperative management between various safety management departments.
Keywords: Construction hazard, grid management, griddization computing, process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 157913587 A Generic Approach to Reuse Unified Modeling Language Components Following an Agile Process
Authors: Rim Bouhaouel, Naoufel Kraïem, Zuhoor Al Khanjari
Abstract:
Unified Modeling Language (UML) is considered as one of the widespread modeling language standardized by the Object Management Group (OMG). Therefore, the model driving engineering (MDE) community attempts to provide reuse of UML diagrams, and do not construct it from scratch. The UML model appears according to a specific software development process. The existing method generation models focused on the different techniques of transformation without considering the development process. Our work aims to construct an UML component from fragments of UML diagram basing on an agile method. We define UML fragment as a portion of a UML diagram, which express a business target. To guide the generation of fragments of UML models using an agile process, we need a flexible approach, which adapts to the agile changes and covers all its activities. We use the software product line (SPL) to derive a fragment of process agile method. This paper explains our approach, named RECUP, to generate UML fragments following an agile process, and overviews the different aspects. In this paper, we present the approach and we define the different phases and artifacts.Keywords: UML, component, fragment, agile, SPL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 91713586 GSA-Based Design of Dual Proportional Integral Load Frequency Controllers for Nonlinear Hydrothermal Power System
Authors: M. Elsisi, M. Soliman, M. A. S. Aboelela, W. Mansour
Abstract:
This paper considers the design of Dual Proportional- Integral (DPI) Load Frequency Control (LFC), using gravitational search algorithm (GSA). The design is carried out for nonlinear hydrothermal power system where generation rate constraint (GRC) and governor dead band are considered. Furthermore, time delays imposed by governor-turbine, thermodynamic process, and communication channels are investigated. GSA is utilized to search for optimal controller parameters by minimizing a time-domain based objective function. GSA-based DPI has been compared to Ziegler- Nichols based PI, and Genetic Algorithm (GA) based PI controllers in order to demonstrate the superior efficiency of the proposed design. Simulation results are carried for a wide range of operating conditions and system parameters variations.Keywords: Gravitational Search Algorithm (GSA), Load Frequency Control (LFC), Dual Proportional-Integral (DPI) controller.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 198513585 An Extended Eclectic Paradigm of Dunning: Impact of New International Business Processes
Authors: D. de Matías Batalla
Abstract:
This paper develops and extended eclectic paradigm to fit the firm internationalization process with the real international business world. The approach is based on Dunning´s, introducing new concepts like mode of entry, international joint venture o international mergers and acquisitions. At the same time is presented a model to describe the Spanish international mergers and acquisitions in order to determinate the most important factor that influence in this type of foreign direct investment.
Keywords: Dunning, eclectic paradigm, foreign direct investment, IJV, international business, international management, multinational firms, firm internationalization process, M&A.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 306913584 A Relational Case-Based Reasoning Framework for Project Delivery System Selection
Authors: Yang Cui, Yong Qiang Chen
Abstract:
An appropriate project delivery system (PDS) is crucial to the success of a construction projects. Case-based Reasoning (CBR) is a useful support for PDS selection. However, the traditional CBR approach represents cases as attribute-value vectors without taking relations among attributes into consideration, and could not calculate the similarity when the structures of cases are not strictly same. Therefore, this paper solves this problem by adopting the Relational Case-based Reasoning (RCBR) approach for PDS selection, considering both the structural similarity and feature similarity. To develop the feature terms of the construction projects, the criteria and factors governing PDS selection process are first identified. Then feature terms for the construction projects are developed. Finally, the mechanism of similarity calculation and a case study indicate how RCBR works for PDS selection. The adoption of RCBR in PDS selection expands the scope of application of traditional CBR method and improves the accuracy of the PDS selection system.
Keywords: Relational Cased-based Reasoning, Case-based Reasoning, Project delivery system, Selection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 199413583 A New Approach for Predicting and Optimizing Weld Bead Geometry in GMAW
Authors: Farhad Kolahan, Mehdi Heidari
Abstract:
Gas Metal Arc Welding (GMAW) processes is an important joining process widely used in metal fabrication industries. This paper addresses modeling and optimization of this technique using a set of experimental data and regression analysis. The set of experimental data has been used to assess the influence of GMAW process parameters in weld bead geometry. The process variables considered here include voltage (V); wire feed rate (F); torch Angle (A); welding speed (S) and nozzle-to-plate distance (D). The process output characteristics include weld bead height, width and penetration. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. In the next stage, the proposed model is embedded into a Simulated Annealing (SA) algorithm to optimize the GMAW process parameters. The objective is to determine a suitable set of process parameters that can produce desired bead geometry, considering the ranges of the process parameters. Computational results prove the effectiveness of the proposed model and optimization procedure.Keywords: Weld Bead Geometry, GMAW welding, Processparameters Optimization, Modeling, SA algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 218713582 Simulation of the Extensional Flow Mixing of Molten Aluminium and Fly Ash Nanoparticles
Authors: O. Ualibek, C. Spitas, V. Inglezakis, G. Itskos
Abstract:
This study presents simulations of an aluminium melt containing an initially non-dispersed fly ash nanoparticle phase. Mixing is affected predominantly by means of forced extensional flow via either straight or slanted orifices. The sensitivity to various process parameters is determined. The simulated process is used for the production of cast fly ash-aluminium nanocomposites. The possibilities for rod and plate stock grading in the context of a continuous casting process implementation are discussed.Keywords: Metal matrix composites, fly ash nanoparticles, aluminium 2024, agglomeration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 100213581 Evaluation Process for the Hardware Safety Integrity Level
Authors: Sung Kyu Kim, Yong Soo Kim
Abstract:
Safety instrumented systems (SISs) are becoming increasingly complex and the proportion of programmable electronic parts is growing. The IEC 61508 global standard was established to ensure the functional safety of SISs, but it was expressed in highly macroscopic terms. This study introduces an evaluation process for hardware safety integrity levels through failure modes, effects, and diagnostic analysis (FMEDA).FMEDA is widely used to evaluate safety levels, and it provides the information on failure rates and failure mode distributions necessary to calculate a diagnostic coverage factor for a given component. In our evaluation process, the components of the SIS subsystem are first defined in terms of failure modes and effects. Then, the failure rate and failure mechanism distribution are assigned to each component. The safety mode and detectability of each failure mode are determined for each component. Finally, the hardware safety integrity level is evaluated based on the calculated results.Keywords: Safety instrumented system; Safety integrity level; Failure modes, effects, and diagnostic analysis; IEC 61508.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 253113580 Contractor Selection in Saudi Arabia
Authors: M. A. Bajaber, M. A. Taha
Abstract:
Contractor selection in Saudi Arabia is very important due to the large construction boom and the contractor role to get over construction risks. The need for investigating contractor selection is due to the following reasons; large number of defaulted or failed projects (18%), large number of disputes attributed to contractor during the project execution stage (almost twofold), the extension of the General Agreement on Tariffs and Trade (GATT) into construction industry, and finally the few number of researches. The selection strategy is not perfect and considered as the reason behind irresponsible contractors. As a response, this research was conducted to review the contractor selection strategies as an integral part of a long advanced research to develop a good selection model. Many techniques can be used to form a selection strategy; multi criteria for optimizing decision, prequalification to discover contractor-s responsibility, bidding process for competition, third party guarantee to enhance the selection, and fuzzy techniques for ambiguities and incomplete information.
Keywords: Bidding, Construction industry, Contractor selection, Saudi Arabia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 314113579 Design Patterns for Emergency Management Processes
Authors: Tomáš Ludík, Jiří Barta, Josef Navrátil
Abstract:
Natural or human made disasters have a significant negative impact on the environment. At the same time there is an extensive effort to support management and decision making in emergency situations by information technologies. Therefore the purpose of the paper is to propose a design patterns applicable in emergency management, enabling better analysis and design of emergency management processes and therefore easier development and deployment of information systems in the field of emergency management. It will be achieved by detailed analysis of existing emergency management legislation, contingency plans and information systems. The result is a set of design patterns focused at emergency management processes that enable easier design of emergency plans or development of new information system. These results will have a major impact on the development of new information systems as well as to more effective and faster solving of emergencies.
Keywords: Analysis and Design, Business Process Modeling Notation, Contingency Plans, Design Patterns, Emergency Management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 309913578 Air Dispersion Model for Prediction Fugitive Landfill Gaseous Emission Impact in Ambient Atmosphere
Authors: Moustafa Osman Mohammed
Abstract:
This paper will explore formation of HCl aerosol at atmospheric boundary layers and encourages the uptake of environmental modeling systems (EMSs) as a practice evaluation of gaseous emissions (“framework measures”) from small and medium-sized enterprises (SMEs). The conceptual model predicts greenhouse gas emissions to ecological points beyond landfill site operations. It focuses on incorporation traditional knowledge into baseline information for both measurement data and the mathematical results, regarding parameters influence model variable inputs. The paper has simplified parameters of aerosol processes based on the more complex aerosol process computations. The simple model can be implemented to both Gaussian and Eulerian rural dispersion models. Aerosol processes considered in this study were (i) the coagulation of particles, (ii) the condensation and evaporation of organic vapors, and (iii) dry deposition. The chemical transformation of gas-phase compounds is taken into account photochemical formulation with exposure effects according to HCl concentrations as starting point of risk assessment. The discussion set out distinctly aspect of sustainability in reflection inputs, outputs, and modes of impact on the environment. Thereby, models incorporate abiotic and biotic species to broaden the scope of integration for both quantification impact and assessment risks. The later environmental obligations suggest either a recommendation or a decision of what is a legislative should be achieved for mitigation measures of landfill gas (LFG) ultimately.Keywords: Air dispersion model, landfill management, spatial analysis, environmental impact and risk assessment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155813577 Risk Evaluation of Information Technology Projects Based on Fuzzy Analytic Hierarchal Process
Authors: H. Iranmanesh, S. Nazari Shirkouhi, M. R. Skandari
Abstract:
Information Technology (IT) projects are always accompanied by various risks and because of high rate of failure in such projects, managing risks in order to neutralize or at least decrease their effects on the success of the project is strongly essential. In this paper, fuzzy analytical hierarchy process (FAHP) is exploited as a means of risk evaluation methodology to prioritize and organize risk factors faced in IT projects. A real case of IT projects, a project of design and implementation of an integrated information system in a vehicle producing company in Iran is studied. Related risk factors are identified and then expert qualitative judgments about these factors are acquired. Translating these judgments to fuzzy numbers and using them as an input to FAHP, risk factors are then ranked and prioritized by FAHP in order to make project managers aware of more important risks and enable them to adopt suitable measures to deal with these highly devastative risks.Keywords: Information technology projects, Risk evaluation, Analytic hierarchal process, fuzzy logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 194213576 Edge-end Pixel Extraction for Edge-based Image Segmentation
Authors: Mahinda P. Pathegama, Özdemir Göl
Abstract:
Extraction of edge-end-pixels is an important step for the edge linking process to achieve edge-based image segmentation. This paper presents an algorithm to extract edge-end pixels together with their directional sensitivities as an augmentation to the currently available mathematical models. The algorithm is implemented in the Java environment because of its inherent compatibility with web interfaces since its main use is envisaged to be for remote image analysis on a virtual instrumentation platform.
Keywords: edge-end pixels, image processing, imagesegmentation, pixel extraction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2154