Search results for: process developed data warehouse
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 40610

Search results for: process developed data warehouse

40580 Modeling and Simulation of Fluid Catalytic Cracking Process

Authors: Sungho Kim, Dae Shik Kim, Jong Min Lee

Abstract:

Fluid catalytic cracking (FCC) process is one of the most important process in modern refinery industry. This paper focuses on the fluid catalytic cracking (FCC) process. As the FCC process is difficult to model well, due to its non linearities and various interactions between its process variables, rigorous process modeling of whole FCC plant is demanded for control and plant-wide optimization of the plant. In this study, a process design for the FCC plant includes riser reactor, main fractionator, and gas processing unit was developed. A reactor model was described based on four-lumped kinetic scheme. Main fractionator, gas processing unit and other process units are designed to simulate real plant data, using a process flow sheet simulator, Aspen PLUS. The custom reactor model was integrated with the process flow sheet simulator to develop an integrated process model.

Keywords: fluid catalytic cracking, simulation, plant data, process design

Procedia PDF Downloads 498
40579 Application of Knowledge Discovery in Database Techniques in Cost Overruns of Construction Projects

Authors: Mai Ghazal, Ahmed Hammad

Abstract:

Cost overruns in construction projects are considered as worldwide challenges since the cost performance is one of the main measures of success along with schedule performance. To overcome this problem, studies were conducted to investigate the cost overruns' factors, also projects' historical data were analyzed to extract new and useful knowledge from it. This research is studying and analyzing the effect of some factors causing cost overruns using the historical data from completed construction projects. Then, using these factors to estimate the probability of cost overrun occurrence and predict its percentage for future projects. First, an intensive literature review was done to study all the factors that cause cost overrun in construction projects, then another review was done for previous researcher papers about mining process in dealing with cost overruns. Second, a proposed data warehouse was structured which can be used by organizations to store their future data in a well-organized way so it can be easily analyzed later. Third twelve quantitative factors which their data are frequently available at construction projects were selected to be the analyzed factors and suggested predictors for the proposed model.

Keywords: construction management, construction projects, cost overrun, cost performance, data mining, data warehousing, knowledge discovery, knowledge management

Procedia PDF Downloads 338
40578 Storage Assignment Strategies to Reduce Manual Picking Errors with an Emphasis on an Ageing Workforce

Authors: Heiko Diefenbach, Christoph H. Glock

Abstract:

Order picking, i.e., the order-based retrieval of items in a warehouse, is an important time- and cost-intensive process for many logistic systems. Despite the ongoing trend of automation, most order picking systems are still manual picker-to-parts systems, where human pickers walk through the warehouse to collect ordered items. Human work in warehouses is not free from errors, and order pickers may at times pick the wrong or the incorrect number of items. Errors can cause additional costs and significant correction efforts. Moreover, age might increase a person’s likelihood to make mistakes. Hence, the negative impact of picking errors might increase for an aging workforce currently witnessed in many regions globally. A significant amount of research has focused on making order picking systems more efficient. Among other factors, storage assignment, i.e., the assignment of items to storage locations (e.g., shelves) within the warehouse, has been subject to optimization. Usually, the objective is to assign items to storage locations such that order picking times are minimized. Surprisingly, there is a lack of research concerned with picking errors and respective prevention approaches. This paper hypothesize that the storage assignment of items can affect the probability of pick errors. For example, storing similar-looking items apart from one other might reduce confusion. Moreover, storing items that are hard to count or require a lot of counting at easy-to-access and easy-to-comprehend self heights might reduce the probability to pick the wrong number of items. Based on this hypothesis, the paper discusses how to incorporate error-prevention measures into mathematical models for storage assignment optimization. Various approaches with respective benefits and shortcomings are presented and mathematically modeled. To investigate the newly developed models further, they are compared to conventional storage assignment strategies in a computational study. The study specifically investigates how the importance of error prevention increases with pickers being more prone to errors due to age, for example. The results suggest that considering error-prevention measures for storage assignment can reduce error probabilities with only minor decreases in picking efficiency. The results might be especially relevant for an aging workforce.

Keywords: an aging workforce, error prevention, order picking, storage assignment

Procedia PDF Downloads 180
40577 Data Acquisition System for Automotive Testing According to the European Directive 2004/104/EC

Authors: Herminio Martínez-García, Juan Gámiz, Yolanda Bolea, Antoni Grau

Abstract:

This article presents an interactive system for data acquisition in vehicle testing according to the test process defined in automotive directive 2004/104/EC. The project has been designed and developed by authors for the Spanish company Applus-LGAI. The developed project will result in a new process, which will involve the creation of braking cycle test defined in the aforementioned automotive directive. It will also allow the analysis of new vehicle features that was not feasible, allowing an increasing interaction with the vehicle. Potential users of this system in the short term will be vehicle manufacturers and in a medium term the system can be extended to testing other automotive components and EMC tests.

Keywords: automotive process, data acquisition system, electromagnetic compatibility (EMC) testing, European Directive 2004/104/EC

Procedia PDF Downloads 314
40576 Assessing the Effectiveness of Warehousing Facility Management: The Case of Mantrac Ghana Limited

Authors: Kuhorfah Emmanuel Mawuli

Abstract:

Generally, for firms to enhance their operational efficiency of logistics, it is imperative to assess the logistics function. The cost of logistics conventionally represents a key consideration in the pricing decisions of firms, which suggests that cost efficiency in logistics can go a long way to improve margins. Warehousing, which is a key part of logistics operations, has the prospect of influencing operational efficiency in logistics management as well as customer value, but this potential has often not been recognized. It has been found that there is a paucity of research that evaluates the efficiency of warehouses. Indeed, limited research has been conducted to examine potential barriers to effective warehousing management. Due to this paucity of research, there is limited knowledge on how to address the obstacles associated with warehousing management. In order for warehousing management to become profitable, there is the need to integrate, balance, and manage the economic inputs and outputs of the entire warehouse operations, something that many firms tend to ignore. Management of warehousing is not solely related to storage functions. Instead, effective warehousing management requires such practices as maximum possible mechanization and automation of operations, optimal use of space and capacity of storage facilities, organization through "continuous flow" of goods, a planned system of storage operations, and safety of goods. For example, there is an important need for space utilization of the warehouse surface as it is a good way to evaluate the storing operation and pick items per hour. In the setting of Mantrac Ghana, not much knowledge regarding the management of the warehouses exists. The researcher has personally observed many gaps in the management of the warehouse facilities in the case organization Mantrac Ghana. It is important, therefore, to assess the warehouse facility management of the case company with the objective of identifying weaknesses for improvement. The study employs an in-depth qualitative research approach using interviews as a mode of data collection. Respondents in the study mainly comprised warehouse facility managers in the studied company. A total of 10 participants were selected for the study using a purposive sampling strategy. Results emanating from the study demonstrate limited warehousing effectiveness in the case company. Findings further reveal that the major barriers to effective warehousing facility management comprise poor layout, poor picking optimization, labour costs, and inaccurate orders; policy implications of the study findings are finally outlined.

Keywords: assessing, warehousing, facility, management

Procedia PDF Downloads 37
40575 Data-Mining Approach to Analyzing Industrial Process Information for Real-Time Monitoring

Authors: Seung-Lock Seo

Abstract:

This work presents a data-mining empirical monitoring scheme for industrial processes with partially unbalanced data. Measurement data of good operations are relatively easy to gather, but in unusual special events or faults it is generally difficult to collect process information or almost impossible to analyze some noisy data of industrial processes. At this time some noise filtering techniques can be used to enhance process monitoring performance in a real-time basis. In addition, pre-processing of raw process data is helpful to eliminate unwanted variation of industrial process data. In this work, the performance of various monitoring schemes was tested and demonstrated for discrete batch process data. It showed that the monitoring performance was improved significantly in terms of monitoring success rate of given process faults.

Keywords: data mining, process data, monitoring, safety, industrial processes

Procedia PDF Downloads 373
40574 The Impact of Interrelationship between Business Intelligence and Knowledge Management on Decision Making Process: An Empirical Investigation of Banking Sector in Jordan

Authors: Issa M. Shehabat, Huda F. Y. Nimri

Abstract:

This paper aims to study the relationship between knowledge management in its processes, including knowledge creation, knowledge sharing, knowledge organization, and knowledge application, and business intelligence tools, including OLAP, data mining, and data warehouse, and their impact on the decision-making process in the banking sector in Jordan. A total of 200 questionnaires were distributed to the sample of the study. The study hypotheses were tested using the statistical package SPSS. Study findings suggest that decision-making processes were positively related to knowledge management processes. Additionally, the components of business intelligence had a positive impact on decision-making. The study recommended conducting studies similar to this study in other sectors such as the industrial, telecommunications, and service sectors to contribute to enhancing understanding of the role of the knowledge management processes and business intelligence tools.

Keywords: business intelligence, knowledge management, decision making, Jordan, banking sector

Procedia PDF Downloads 113
40573 Big Brain: A Single Database System for a Federated Data Warehouse Architecture

Authors: X. Gumara Rigol, I. Martínez de Apellaniz Anzuola, A. Garcia Serrano, A. Franzi Cros, O. Vidal Calbet, A. Al Maruf

Abstract:

Traditional federated architectures for data warehousing work well when corporations have existing regional data warehouses and there is a need to aggregate data at a global level. Schibsted Media Group has been maturing from a decentralised organisation into a more globalised one and needed to build both some of the regional data warehouses for some brands at the same time as the global one. In this paper, we present the architectural alternatives studied and why a custom federated approach was the notable recommendation to go further with the implementation. Although the data warehouses are logically federated, the implementation uses a single database system which presented many advantages like: cost reduction and improved data access to global users allowing consumers of the data to have a common data model for detailed analysis across different geographies and a flexible layer for local specific needs in the same place.

Keywords: data integration, data warehousing, federated architecture, Online Analytical Processing (OLAP)

Procedia PDF Downloads 216
40572 Integration Process and Analytic Interface of different Environmental Open Data Sets with Java/Oracle and R

Authors: Pavel H. Llamocca, Victoria Lopez

Abstract:

The main objective of our work is the comparative analysis of environmental data from Open Data bases, belonging to different governments. This means that you have to integrate data from various different sources. Nowadays, many governments have the intention of publishing thousands of data sets for people and organizations to use them. In this way, the quantity of applications based on Open Data is increasing. However each government has its own procedures to publish its data, and it causes a variety of formats of data sets because there are no international standards to specify the formats of the data sets from Open Data bases. Due to this variety of formats, we must build a data integration process that is able to put together all kind of formats. There are some software tools developed in order to give support to the integration process, e.g. Data Tamer, Data Wrangler. The problem with these tools is that they need data scientist interaction to take part in the integration process as a final step. In our case we don’t want to depend on a data scientist, because environmental data are usually similar and these processes can be automated by programming. The main idea of our tool is to build Hadoop procedures adapted to data sources per each government in order to achieve an automated integration. Our work focus in environment data like temperature, energy consumption, air quality, solar radiation, speeds of wind, etc. Since 2 years, the government of Madrid is publishing its Open Data bases relative to environment indicators in real time. In the same way, other governments have published Open Data sets relative to the environment (like Andalucia or Bilbao). But all of those data sets have different formats and our solution is able to integrate all of them, furthermore it allows the user to make and visualize some analysis over the real-time data. Once the integration task is done, all the data from any government has the same format and the analysis process can be initiated in a computational better way. So the tool presented in this work has two goals: 1. Integration process; and 2. Graphic and analytic interface. As a first approach, the integration process was developed using Java and Oracle and the graphic and analytic interface with Java (jsp). However, in order to open our software tool, as second approach, we also developed an implementation with R language as mature open source technology. R is a really powerful open source programming language that allows us to process and analyze a huge amount of data with high performance. There are also some R libraries for the building of a graphic interface like shiny. A performance comparison between both implementations was made and no significant differences were found. In addition, our work provides with an Official Real-Time Integrated Data Set about Environment Data in Spain to any developer in order that they can build their own applications.

Keywords: open data, R language, data integration, environmental data

Procedia PDF Downloads 288
40571 Effect on Occupational Health Safety and Environment at Work from Metal Handicraft Using Rattanakosin Local Wisdom

Authors: Witthaya Mekhum, Waleerak Sittisom

Abstract:

This research investigated the effect on occupational health safety and environment at work from metal handicraft using Rattanakosin local wisdom focusing on pollution, accidents, and injuries from work. The sample group in this study included 48 metal handicraft workers in 5 communities by using questionnaires and interview to collect data. The evaluation form TISI 18001 was used to analyze job safety analysis (JSA). The results showed that risk at work reduced after applying the developed model. Banbu Community produces alloy bowl rubbed with stone. The high risk process is melting and hitting process. Before the application, the work risk was 82.71%. After the application of the developed model, the work risk was reduced to 50.61%. Banbart Community produces monk’s food bowl. The high risk process is blow pipe welding. Before the application, the work risk was 93.59%. After the application of the developed model, the work risk was reduced to 48.14%. Bannoen Community produces circle gong. The high risk process is milling process. Before the application, the work risk was 85.18%. After the application of the developed model, the work risk was reduced to 46.91%. Teethong Community produces gold leaf. The high risk process is hitting and spreading process. Before the application, the work risk was 86.42%. After the application of the developed model, the work risk was reduced to 64.19%. Ban Changthong Community produces gold ornament. The high risk process is gold melting process. Before the application, the work risk was 67.90%. After the application of the developed model, the work risk was reduced to 37.03%. It can be concluded that with the application of the developed model, the work risk of 5 communities was reduced in the 3 main groups: (1) Work illness reduced by 16.77%; (2) Pollution from work reduced by 10.31%; (3) Accidents and injuries from work reduced by 15.62%.

Keywords: occupational health, safety, local wisdom, Rattanakosin

Procedia PDF Downloads 423
40570 Analysis of Economic Order Quantity, Safety Stock, Maximum Inventory Control, Lot Size and Reorder Point for Engro Polymers and Chemicals

Authors: Ali Akber Jaffri, Asad Naseem, Javeria Khan, Zubair Hamza, Ishtiaq

Abstract:

The purpose of this study is to determine safety stock, maximum inventory level, reordering point, and reordering quantity by rearranging lot sizes for supplier and customer in MRO (maintenance repair operations) warehouse of Engro Polymers & Chemicals. To achieve the aim, physical analysis method and excel commands were carried out to elicit the customer and supplier data provided by the company. Initially, we rearranged the current lot sizes and MOUs (measure of units) in SAP software. Due to change in lot sizes, we have to determine the new quantities for safety stock, maximum inventory, reordering point and reordering quantity as per company's demand. By proposed system, we saved extra cost in terms of reducing time of receiving from vendor and in issuance to customer, ease of material handling in MRO warehouse and also reduce human efforts.

Keywords: maintenance repair operation, maximum inventory, reorder quantity, safety stock

Procedia PDF Downloads 254
40569 Methods for Business Process Simulation Based on Petri Nets

Authors: K. Shoylekova, K. Grigorova

Abstract:

The Petri nets are the first standard for business process modeling. Most probably, it is one of the core reasons why all new standards created afterwards have to be so reformed as to reach the stage of mapping the new standard onto Petri nets. The paper presents a Business process repository based on a universal database. The repository provides the possibility the data about a given process to be stored in three different ways. Business process repository is developed with regard to the reformation of a given model to a Petri net in order to be easily simulated two different techniques for business process simulation based on Petri nets - Yasper and Woflan are discussed. Their advantages and drawbacks are outlined. The way of simulating business process models, stored in the Business process repository is shown.

Keywords: business process repository, petri nets, simulation, Woflan, Yasper

Procedia PDF Downloads 344
40568 Optimization of Headspace Solid Phase Microextraction (SPME) Technique Coupled with GC MS for Identification of Volatile Organic Compounds Released by Trogoderma Variabile

Authors: Thamer Alshuwaili, Yonglin Ren, Bob Du, Manjree Agarwal

Abstract:

The warehouse beetle, Trogoderma variabile Ballion (Coleoptera: Dermestidae), is a major pest of packaged and processed stored products. Warehouse beetle is the common name which was given by Okumura (1972). This pest has been reported to infest 119 different commodities, and it is distributed throughout the tropical and subtropical parts of the world. Also, it is difficult to control because of the insect's ability to stay without food for long times, and it can survive for years under dry conditions and low-moisture food, and it has also developed resistance to many insecticides. The young larvae of these insects can cause damage to seeds, but older larvae prefer to feed on whole grains. The percentage of damage caused by these insects range between 30-70% in the storage. T. variabile is the species most responsible for causing significant damage in grain stores worldwide. Trogoderma spp. is a huge problem for cereal grains, and there are many countries, such as the USA, Australia, China, Kenya, Uganda and Tanzania who have specific quarantine regulations against possible importation. Also, grain stocks can be almost completely destroyed because of the massive populations the insect may develop. However, the purpose of the current research was to optimize conditions to collect volatile organic compound from Trogoderma variabile at different life stages by using headspace solid phase microextraction (SPME) coupled with gas chromatography-mass spectrometry (GC-MS) and flame ionization detection (FID). Using SPME technique to extract volatile from insects is an efficient, straightforward and nondestructive method. Result of the study shows that 15 insects were optimal number for larvae and adults. Selection of the number of insects depend on the height of the peak area and the number of peaks. Sixteen hours were optimized as the best extraction time for larvae and 8 hours was the optimal number of adults.

Keywords: Trogoderma variabile, warehouse beetle , GC-MS, Solid phase microextraction

Procedia PDF Downloads 101
40567 High Performance Computing and Big Data Analytics

Authors: Branci Sarra, Branci Saadia

Abstract:

Because of the multiplied data growth, many computer science tools have been developed to process and analyze these Big Data. High-performance computing architectures have been designed to meet the treatment needs of Big Data (view transaction processing standpoint, strategic, and tactical analytics). The purpose of this article is to provide a historical and global perspective on the recent trend of high-performance computing architectures especially what has a relation with Analytics and Data Mining.

Keywords: high performance computing, HPC, big data, data analysis

Procedia PDF Downloads 489
40566 Comprehensive Assessment of Energy Efficiency within the Production Process

Authors: S. Kreitlein, N. Eder, J. Franke

Abstract:

The importance of energy efficiency within the production process increases steadily. Unfortunately, so far no tools for a comprehensive assessment of energy efficiency within the production process exist. Therefore the Institute for Factory Automation and Production Systems of the Friedrich-Alexander-University Erlangen-Nuremberg has developed two methods with the goal of achieving transparency and a quantitative assessment of energy efficiency: EEV (Energy Efficiency Value) and EPE (Energetic Process Efficiency). This paper describes the basics and state of the art as well as the developed approaches.

Keywords: energy efficiency, energy efficiency value, energetic process efficiency, production

Procedia PDF Downloads 701
40565 Towards a Distributed Computation Platform Tailored for Educational Process Discovery and Analysis

Authors: Awatef Hicheur Cairns, Billel Gueni, Hind Hafdi, Christian Joubert, Nasser Khelifa

Abstract:

Given the ever changing needs of the job markets, education and training centers are increasingly held accountable for student success. Therefore, education and training centers have to focus on ways to streamline their offers and educational processes in order to achieve the highest level of quality in curriculum contents and managerial decisions. Educational process mining is an emerging field in the educational data mining (EDM) discipline, concerned with developing methods to discover, analyze and provide a visual representation of complete educational processes. In this paper, we present our distributed computation platform which allows different education centers and institutions to load their data and access to advanced data mining and process mining services. To achieve this, we present also a comparative study of the different clustering techniques developed in the context of process mining to partition efficiently educational traces. Our goal is to find the best strategy for distributing heavy analysis computations on many processing nodes of our platform.

Keywords: educational process mining, distributed process mining, clustering, distributed platform, educational data mining, ProM

Procedia PDF Downloads 429
40564 Analysis of Inventory Control, Lot Size and Reorder Point for Engro Polymers and Chemicals

Authors: Ali Akber Jaffri, Asad Naseem, Javeria Khan

Abstract:

The purpose of this study is to determine safety stock, maximum inventory level, reordering point, and reordering quantity by rearranging lot sizes for supplier and customer in MRO (maintenance repair operations) warehouse of Engro Polymers & Chemicals. To achieve the aim, physical analysis method and excel commands were carried out to elicit the customer and supplier data provided by the company. Initially, we rearranged the current lot sizes and MOUs (measure of units) in SAP software. Due to change in lot sizes we have to determine the new quantities for safety stock, maximum inventory, reordering point and reordering quantity as per company's demand. By proposed system, we saved extra cost in terms of reducing the time of receiving from vendor and in issuance to customer, ease of material handling in MRO warehouse and also reduce human efforts. The information requirements identified in this study can be utilized in calculating Economic Order Quantity.

Keywords: carrying cost, economic order quantity, fast moving, lead time, lot size, MRO, maximum inventory, ordering cost, physical inspection, reorder point

Procedia PDF Downloads 218
40563 Operating Speed Models on Tangent Sections of Two-Lane Rural Roads

Authors: Dražen Cvitanić, Biljana Maljković

Abstract:

This paper presents models for predicting operating speeds on tangent sections of two-lane rural roads developed on continuous speed data. The data corresponds to 20 drivers of different ages and driving experiences, driving their own cars along an 18 km long section of a state road. The data were first used for determination of maximum operating speeds on tangents and their comparison with speeds in the middle of tangents i.e. speed data used in most of operating speed studies. Analysis of continuous speed data indicated that the spot speed data are not reliable indicators of relevant speeds. After that, operating speed models for tangent sections were developed. There was no significant difference between models developed using speed data in the middle of tangent sections and models developed using maximum operating speeds on tangent sections. All developed models have higher coefficient of determination then models developed on spot speed data. Thus, it can be concluded that the method of measuring has more significant impact on the quality of operating speed model than the location of measurement.

Keywords: operating speed, continuous speed data, tangent sections, spot speed, consistency

Procedia PDF Downloads 426
40562 A Model Architecture Transformation with Approach by Modeling: From UML to Multidimensional Schemas of Data Warehouses

Authors: Ouzayr Rabhi, Ibtissam Arrassen

Abstract:

To provide a complete analysis of the organization and to help decision-making, leaders need to have relevant data; Data Warehouses (DW) are designed to meet such needs. However, designing DW is not trivial and there is no formal method to derive a multidimensional schema from heterogeneous databases. In this article, we present a Model-Driven based approach concerning the design of data warehouses. We describe a multidimensional meta-model and also specify a set of transformations starting from a Unified Modeling Language (UML) metamodel. In this approach, the UML metamodel and the multidimensional one are both considered as a platform-independent model (PIM). The first meta-model is mapped into the second one through transformation rules carried out by the Query View Transformation (QVT) language. This proposal is validated through the application of our approach to generating a multidimensional schema of a Balanced Scorecard (BSC) DW. We are interested in the BSC perspectives, which are highly linked to the vision and the strategies of an organization.

Keywords: data warehouse, meta-model, model-driven architecture, transformation, UML

Procedia PDF Downloads 129
40561 Block Mining: Block Chain Enabled Process Mining Database

Authors: James Newman

Abstract:

Process mining is an emerging technology that looks to serialize enterprise data in time series data. It has been used by many companies and has been the subject of a variety of research papers. However, the majority of current efforts have looked at how to best create process mining from standard relational databases. This paper is the first pass at outlining a database custom-built for the minimal viable product of process mining. We present Block Miner, a blockchain protocol to store process mining data across a distributed network. We demonstrate the feasibility of storing process mining data on the blockchain. We present a proof of concept and show how the intersection of these two technologies helps to solve a variety of issues, including but not limited to ransomware attacks, tax documentation, and conflict resolution.

Keywords: blockchain, process mining, memory optimization, protocol

Procedia PDF Downloads 67
40560 Traditional Drawing, BIM and Erudite Design Process

Authors: Maryam Kalkatechi

Abstract:

Nowadays, parametric design, scientific analysis, and digital fabrication are dominant. Many architectural practices are increasingly seeking to incorporate advanced digital software and fabrication in their projects. Proposing an erudite design process that combines digital and practical aspects in a strong frame within the method was resulted from the dissertation research. The digital aspects are the progressive advancements in algorithm design and simulation software. These aspects have assisted the firms to develop more holistic concepts at the early stage and maintain collaboration among disciplines during the design process. The erudite design process enhances the current design processes by encouraging the designer to implement the construction and architecture knowledge within the algorithm to make successful design processes. The erudite design process also involves the ongoing improvements of applying the new method of 3D printing in construction. This is achieved through the ‘data-sketches’. The term ‘data-sketch’ was developed by the author in the dissertation that was recently completed. It accommodates the decisions of the architect on the algorithm. This paper introduces the erudite design process and its components. It will summarize the application of this process in development of the ‘3D printed construction unit’. This paper contributes to overlaying the academic and practice with advanced technology by presenting a design process that transfers the dominance of tool to the learned architect and encourages innovation in design processes.

Keywords: erudite, data-sketch, algorithm design in architecture, design process

Procedia PDF Downloads 246
40559 Industrial Process Mining Based on Data Pattern Modeling and Nonlinear Analysis

Authors: Hyun-Woo Cho

Abstract:

Unexpected events may occur with serious impacts on industrial process. This work utilizes a data representation technique to model and to analyze process data pattern for the purpose of diagnosis. In this work, the use of triangular representation of process data is evaluated using simulation process. Furthermore, the effect of using different pre-treatment techniques based on such as linear or nonlinear reduced spaces was compared. This work extracted the fault pattern in the reduced space, not in the original data space. The results have shown that the non-linear technique based diagnosis method produced more reliable results and outperforms linear method.

Keywords: process monitoring, data analysis, pattern modeling, fault, nonlinear techniques

Procedia PDF Downloads 362
40558 A Machine Learning Decision Support Framework for Industrial Engineering Purposes

Authors: Anli Du Preez, James Bekker

Abstract:

Data is currently one of the most critical and influential emerging technologies. However, the true potential of data is yet to be exploited since, currently, about 1% of generated data are ever actually analyzed for value creation. There is a data gap where data is not explored due to the lack of data analytics infrastructure and the required data analytics skills. This study developed a decision support framework for data analytics by following Jabareen’s framework development methodology. The study focused on machine learning algorithms, which is a subset of data analytics. The developed framework is designed to assist data analysts with little experience, in choosing the appropriate machine learning algorithm given the purpose of their application.

Keywords: Data analytics, Industrial engineering, Machine learning, Value creation

Procedia PDF Downloads 144
40557 Intelligent Process Data Mining for Monitoring for Fault-Free Operation of Industrial Processes

Authors: Hyun-Woo Cho

Abstract:

The real-time fault monitoring and diagnosis of large scale production processes is helpful and necessary in order to operate industrial process safely and efficiently producing good final product quality. Unusual and abnormal events of the process may have a serious impact on the process such as malfunctions or breakdowns. This work try to utilize process measurement data obtained in an on-line basis for the safe and some fault-free operation of industrial processes. To this end, this work evaluated the proposed intelligent process data monitoring framework based on a simulation process. The monitoring scheme extracts the fault pattern in the reduced space for the reliable data representation. Moreover, this work shows the results of using linear and nonlinear techniques for the monitoring purpose. It has shown that the nonlinear technique produced more reliable monitoring results and outperforms linear methods. The adoption of the qualitative monitoring model helps to reduce the sensitivity of the fault pattern to noise.

Keywords: process data, data mining, process operation, real-time monitoring

Procedia PDF Downloads 610
40556 Parametric Study of Vertical Diffusion Stills for Water Desalination

Authors: A. Seleem, M. Mortada, M. El-Morsi, M. Younan

Abstract:

Diffusion stills have been effective in water desalination. The present work represents a model of the distillation process by using vertical single-effect diffusion stills. A semi-analytical model has been developed to model the process. A software computer code using Engineering Equation Solver EES software has been developed to solve the equations of the developed model. An experimental setup has been constructed, and used for the validation of the model. The model is also validated against former literature results. The results obtained from the present experimental test rig, and the data from the literature, have been compared with the results of the code to find its best range of validity. In addition, a parametric analysis of the system has been developed using the model to determine the effect of operating conditions on the system's performance. The dominant parameters that affect the productivity of the still are the hot plate temperature that ranges from (55-90 °C) and feed flow rate in range of (0.00694-0.0211 kg/m2-s).

Keywords: analytical model, solar distillation, sustainable water systems, vertical diffusion still

Procedia PDF Downloads 381
40555 Applying (1, T) Ordering Policy in a Multi-Vendor-Single-Buyer Inventory System with Lost Sales and Poisson Demand

Authors: Adel Nikfarjam, Hamed Tayebi, Sadoullah Ebrahimnejad

Abstract:

This paper considers a two-echelon inventory system with a number of warehouses and a single retailer. The retailer replenishes its required items from warehouses, and assembles them into a single final product. We assume that each warehouse supplies only one kind of the raw material for the retailer. The demand process of the final product is assumed to be Poissson, and unsatisfied demand of the final product will be lost. The retailer applies one-for-one-period ordering policy which is also known as (1, T) ordering policy. In this policy the retailer orders to each warehouse a fixed quantity of each item at fixed time intervals, which the fixed quantity is equal to the utilization of the item in the final product. Since, this policy eliminates all demand uncertainties at the upstream echelon, the standard lot sizing model can be applied at all warehouses. In this paper, we calculate the total cost function of the inventory system. Then, based on this function, we present a procedure to obtain the optimal time interval between two consecutive order placements from retailer to the warehouses, and the optimal order quantities of warehouses (assuming that there are positive ordering costs at warehouses). Finally, we present some numerical examples, and conduct numerical sensitivity analysis for cost parameters.

Keywords: two-echelon supply chain, multi-vendor-single-buyer inventory system, lost sales, Poisson demand, one-for-one-period policy, lot sizing model

Procedia PDF Downloads 286
40554 Inventory Management to Minimize Storage Costs and Improve Delivery Time in a Pharmaceutical Industry

Authors: Israel Becerril Rosales, Manuel González De La Rosa, Gerardo Villa Sánchez

Abstract:

In this work, the effects that produce not having a good inventory management is analyzed, in addition of the way that how it affects the storage costs. The research began conducting the historical analysis about stored products, its storage capacity, and distribution. The results were not optimal, since in all its raw materials (RM) have overstocking, the warehouse capacity is only used by 61%, does not have a specific place for each of its RM, causing that the delivery times increases and makes difficult a cyclical inventory. These shortcomings allowed to view and select as design alternatives the inventory ABC, so that depending on the consumption of each RM would be redistributed by using economic amount requested. Also, the Delphi method to ensure the practical applicability of the proposed tool was used, taking in account comments and suggestions of the involved experts, as well as the compliance of NOM-059-SSA1-2015 good manufacturing practices of drug. With the actions implemented, the utilization rate drops of 61% to 32% capacity, it shows that the warehouse was not designed properly due to there is not an industrial engineering area.

Keywords: lead time, improve delivery, storage costs, inventory management

Procedia PDF Downloads 205
40553 Designing of Tooling Solution for Material Handling in Highly Automated Manufacturing System

Authors: Muhammad Umair, Yuri Nikolaev, Denis Artemov, Ighor Uzhinsky

Abstract:

A flexible manufacturing system is an integral part of a smart factory of industry 4.0 in which every machine is interconnected and works autonomously. Robots are in the process of replacing humans in every industrial sector. As the cyber-physical-system (CPS) and artificial intelligence (AI) are advancing, the manufacturing industry is getting more dependent on computers than human brains. This modernization has boosted the production with high quality and accuracy and shifted from classic production to smart manufacturing systems. However, material handling for such automated productions is a challenge and needs to be addressed with the best possible solution. Conventional clamping systems are designed for manual work and not suitable for highly automated production systems. Researchers and engineers are trying to find the most economical solution for loading/unloading and transportation workpieces from a warehouse to a machine shop for machining operations and back to the warehouse without human involvement. This work aims to propose an advanced multi-shape tooling solution for highly automated manufacturing systems. The currently obtained result shows that it could function well with automated guided vehicles (AGVs) and modern conveyor belts. The proposed solution is following requirements to be automation-friendly, universal for different part geometry and production operations. We used a bottom-up approach in this work, starting with studying different case scenarios and their limitations and finishing with the general solution.

Keywords: artificial intelligence, cyber physics system, Industry 4.0, material handling, smart factory, flexible manufacturing system

Procedia PDF Downloads 112
40552 Process Data-Driven Representation of Abnormalities for Efficient Process Control

Authors: Hyun-Woo Cho

Abstract:

Unexpected operational events or abnormalities of industrial processes have a serious impact on the quality of final product of interest. In terms of statistical process control, fault detection and diagnosis of processes is one of the essential tasks needed to run the process safely. In this work, nonlinear representation of process measurement data is presented and evaluated using a simulation process. The effect of using different representation methods on the diagnosis performance is tested in terms of computational efficiency and data handling. The results have shown that the nonlinear representation technique produced more reliable diagnosis results and outperforms linear methods. The use of data filtering step improved computational speed and diagnosis performance for test data sets. The presented scheme is different from existing ones in that it attempts to extract the fault pattern in the reduced space, not in the original process variable space. Thus this scheme helps to reduce the sensitivity of empirical models to noise.

Keywords: fault diagnosis, nonlinear technique, process data, reduced spaces

Procedia PDF Downloads 225
40551 The Technological Problem of Simulation of the Logistics Center

Authors: Juraj Camaj, Anna Dolinayova, Jana Lalinska, Miroslav Bariak

Abstract:

Planning of infrastructure and processes in logistic center within the frame of various kinds of logistic hubs and technological activities in them represent quite complex problem. The main goal is to design appropriate layout, which enables to realize expected operation on the desired levels. The simulation software represents progressive contemporary experimental technique, which can support complex processes of infrastructure planning and all of activities on it. It means that simulation experiments, reflecting various planned infrastructure variants, investigate and verify their eligibilities in relation with corresponding expected operation. The inducted approach enables to make qualified decisions about infrastructure investments or measures, which derive benefit from simulation-based verifications. The paper represents simulation software for simulation infrastructural layout and technological activities in marshalling yard, intermodal terminal, warehouse and combination between them as the parts of logistic center.

Keywords: marshalling yard, intermodal terminal, warehouse, transport technology, simulation

Procedia PDF Downloads 491