Search results for: robust process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15883

Search results for: robust process

15433 Aircraft Landing Process Simulation Using Multi-Body Multi-Dynamics Software

Authors: Ahmad Kavousi, Ali Delaviz

Abstract:

In this project, the landing process is simulated by using of multi-body dynamics commercial software. Various factors, including landing situations, aircraft structures and climate are used in this simulation. The purpose of this project is to determine the forces exerted on the aircraft landing gears in landing process in various landing conditions. For this purpose, the ADAMS multi-body dynamics software is used. Different scenarios based on FAR-25, including level landing, tail-down landing, crab landing are simulated. Results of dynamic simulation software with landing load factor obtained from the analytical solution are compared. The effect of fuselage elasticity on the landing load is studied. For this purpose, both of elastic and rigid body assumptions are used in the simulation process, and the results are compared and some conclusions are made.

Keywords: landing gear, landing process, aircraft, multi-body dynamics

Procedia PDF Downloads 478
15432 Process Simulation of 1-Butene Separation from C4 Mixture by Extractive Distillation

Authors: Muhammad Naeem, Abdulrahman A. Al-Rabiah, Wasif Mughees

Abstract:

Technical mixture of C4 containing 1-butene and n-butane are very close to each other with regard to their boiling points i.e. -6.3°C for 1-butene and -1°C for n-butane. Extractive distillation process is used for the separation of 1-butene from the existing mixture of C4. The solvent is the essential of extractive distillation, and an appropriate solvent plays an important role in the process economy of extractive distillation. Aspen Plus has been applied for the separation of these hydrocarbons as a simulator. Moreover, NRTL activity coefficient model was used in the simulation. This model indicated that the material balances in this separation process were accurate for several solvent flow rates. Mixture of acetonitrile and water used as a solvent and 99% pure 1-butene was separated. This simulation proposed the ratio of the feed to solvent as 1: 7.9 and 15 plates for the solvent recovery column. Previously feed to solvent ratio was more than this and the number of proposed plates were 30, which shows that the separation process can be economized.

Keywords: extractive distillation, 1-butene, aspen plus, ACN solvent

Procedia PDF Downloads 502
15431 Identification and Selection of a Supply Chain Target Process for Re-Design

Authors: Jaime A. Palma-Mendoza

Abstract:

A supply chain consists of different processes and when conducting supply chain re-design is necessary to identify the relevant processes and select a target for re-design. A solution was developed which consists to identify first the relevant processes using the Supply Chain Operations Reference (SCOR) model, then to use Analytical Hierarchy Process (AHP) for target process selection. An application was conducted in an Airline MRO supply chain re-design project which shows this combination can clearly aid the identification of relevant supply chain processes and the selection of a target process for re-design.

Keywords: decision support systems, multiple criteria analysis, supply chain management

Procedia PDF Downloads 464
15430 Image Segmentation Techniques: Review

Authors: Lindani Mbatha, Suvendi Rimer, Mpho Gololo

Abstract:

Image segmentation is the process of dividing an image into several sections, such as the object's background and the foreground. It is a critical technique in both image-processing tasks and computer vision. Most of the image segmentation algorithms have been developed for gray-scale images and little research and algorithms have been developed for the color images. Most image segmentation algorithms or techniques vary based on the input data and the application. Nearly all of the techniques are not suitable for noisy environments. Most of the work that has been done uses the Markov Random Field (MRF), which involves the computations and is said to be robust to noise. In the past recent years' image segmentation has been brought to tackle problems such as easy processing of an image, interpretation of the contents of an image, and easy analysing of an image. This article reviews and summarizes some of the image segmentation techniques and algorithms that have been developed in the past years. The techniques include neural networks (CNN), edge-based techniques, region growing, clustering, and thresholding techniques and so on. The advantages and disadvantages of medical ultrasound image segmentation techniques are also discussed. The article also addresses the applications and potential future developments that can be done around image segmentation. This review article concludes with the fact that no technique is perfectly suitable for the segmentation of all different types of images, but the use of hybrid techniques yields more accurate and efficient results.

Keywords: clustering-based, convolution-network, edge-based, region-growing

Procedia PDF Downloads 59
15429 Robust Method for Evaluation of Catchment Response to Rainfall Variations Using Vegetation Indices and Surface Temperature

Authors: Revalin Herdianto

Abstract:

Recent climate changes increase uncertainties in vegetation conditions such as health and biomass globally and locally. The detection is, however, difficult due to the spatial and temporal scale of vegetation coverage. Due to unique vegetation response to its environmental conditions such as water availability, the interplay between vegetation dynamics and hydrologic conditions leave a signature in their feedback relationship. Vegetation indices (VI) depict vegetation biomass and photosynthetic capacity that indicate vegetation dynamics as a response to variables including hydrologic conditions and microclimate factors such as rainfall characteristics and land surface temperature (LST). It is hypothesized that the signature may be depicted by VI in its relationship with other variables. To study this signature, several catchments in Asia, Australia, and Indonesia were analysed to assess the variations in hydrologic characteristics with vegetation types. Methods used in this study includes geographic identification and pixel marking for studied catchments, analysing time series of VI and LST of the marked pixels, smoothing technique using Savitzky-Golay filter, which is effective for large area and extensive data. Time series of VI, LST, and rainfall from satellite and ground stations coupled with digital elevation models were analysed and presented. This study found that the hydrologic response of vegetation to rainfall variations may be shown in one hydrologic year, in which a drought event can be detected a year later as a suppressed growth. However, an annual rainfall of above average do not promote growth above average as shown by VI. This technique is found to be a robust and tractable approach for assessing catchment dynamics in changing climates.

Keywords: vegetation indices, land surface temperature, vegetation dynamics, catchment

Procedia PDF Downloads 263
15428 Optimal Bayesian Chart for Controlling Expected Number of Defects in Production Processes

Authors: V. Makis, L. Jafari

Abstract:

In this paper, we develop an optimal Bayesian chart to control the expected number of defects per inspection unit in production processes with long production runs. We formulate this control problem in the optimal stopping framework. The objective is to determine the optimal stopping rule minimizing the long-run expected average cost per unit time considering partial information obtained from the process sampling at regular epochs. We prove the optimality of the control limit policy, i.e., the process is stopped and the search for assignable causes is initiated when the posterior probability that the process is out of control exceeds a control limit. An algorithm in the semi-Markov decision process framework is developed to calculate the optimal control limit and the corresponding average cost. Numerical examples are presented to illustrate the developed optimal control chart and to compare it with the traditional u-chart.

Keywords: Bayesian u-chart, economic design, optimal stopping, semi-Markov decision process, statistical process control

Procedia PDF Downloads 542
15427 Intelligent Production Machine

Authors: A. Şahinoğlu, R. Gürbüz, A. Güllü, M. Karhan

Abstract:

This study in production machines, it is aimed that machine will automatically perceive cutting data and alter cutting parameters. The two most important parameters have to be checked in machine control unit are progress feed rate and speeds. These parameters are aimed to be controlled by sounds of machine. Optimum sound’s features introduced to computer. During process, real time data is received and converted by Matlab software. Data is converted into numerical values. According to them progress and speeds decreases/increases at a certain rate and thus optimum sound is acquired. Cutting process is made in respect of optimum cutting parameters. During chip remove progress, features of cutting tools, kind of cut material, cutting parameters and used machine; affects on various parameters. Instead of required parameters need to be measured such as temperature, vibration, and tool wear that emerged during cutting process; detailed analysis of the sound emerged during cutting process will provide detection of various data that included in the cutting process by the much more easy and economic way. The relation between cutting parameters and sound is being identified.

Keywords: cutting process, sound processing, intelligent late, sound analysis

Procedia PDF Downloads 306
15426 Wait-Optimized Scheduler Algorithm for Efficient Process Scheduling in Computer Systems

Authors: Md Habibur Rahman, Jaeho Kim

Abstract:

Efficient process scheduling is a crucial factor in ensuring optimal system performance and resource utilization in computer systems. While various algorithms have been proposed over the years, there are still limitations to their effectiveness. This paper introduces a new Wait-Optimized Scheduler (WOS) algorithm that aims to minimize process waiting time by dividing them into two layers and considering both process time and waiting time. The WOS algorithm is non-preemptive and prioritizes processes with the shortest WOS. In the first layer, each process runs for a predetermined duration, and any unfinished process is subsequently moved to the second layer, resulting in a decrease in response time. Whenever the first layer is free or the number of processes in the second layer is twice that of the first layer, the algorithm sorts all the processes in the second layer based on their remaining time minus waiting time and sends one process to the first layer to run. This ensures that all processes eventually run, optimizing waiting time. To evaluate the performance of the WOS algorithm, we conducted experiments comparing its performance with traditional scheduling algorithms such as First-Come-First-Serve (FCFS) and Shortest-Job-First (SJF). The results showed that the WOS algorithm outperformed the traditional algorithms in reducing the waiting time of processes, particularly in scenarios with a large number of short tasks with long wait times. Our study highlights the effectiveness of the WOS algorithm in improving process scheduling efficiency in computer systems. By reducing process waiting time, the WOS algorithm can improve system performance and resource utilization. The findings of this study provide valuable insights for researchers and practitioners in developing and implementing efficient process scheduling algorithms.

Keywords: process scheduling, wait-optimized scheduler, response time, non-preemptive, waiting time, traditional scheduling algorithms, first-come-first-serve, shortest-job-first, system performance, resource utilization

Procedia PDF Downloads 67
15425 The Impact of the Business Process Reengineering on the Practices of the Human Resources Management in the Franco Tunisian Company-Network

Authors: Nesrine Bougarech, Habib Affes

Abstract:

This research lays the emphasis on the business process reengineering (BPR) which consists in radically altering the organizational processes through the optimal use of information technology (IT) to attain major enhancements in terms of quality, performance and productivity. A survey of the business process reengineering (BPR) was carried out in three French groups and their subsidiaries in Tunisia. The data collected were qualitatively analyzed in an attempt to test the main indicators of the success of a business process reengineering project (BPR) and to compare the importance of these indicators in the context of France versus Tunisia. The study corroborates that the respect of the inherent principles of the business process reengineering (BPR) and the diversity of the human resources involved in the project can lead to better productivity, higher quality of the goods or services and lower cost. Additionally, our results mirror the extent to which the respect of the principles and the diversity of resources are more important in the French companies than in their Tunisian subsidiaries.

Keywords: business process reengineering (BPR), human resources management (HRM), information technology (IT), management

Procedia PDF Downloads 383
15424 A Systematic Review of Process Research in Software Engineering

Authors: Tulasi Rayasa, Phani Kumar Pullela

Abstract:

A systematic review is a research method that involves collecting and evaluating the information on a specific topic in order to provide a comprehensive and unbiased review. This type of review aims to improve the software development process by ensuring that the research is thorough and accurate. To ensure objectivity, it is important to follow systematic guidelines and consider multiple sources, such as literature reviews, interviews, and surveys. The evaluation process should also be streamlined by incorporating research from journals and other sources, such as grey literature. The main goal of a systematic review is to identify the consistency of current models in the field of computer application and software engineering.

Keywords: computer application, software engineering, process research, data science

Procedia PDF Downloads 70
15423 Seismic Performance Evaluation of Existing Building Using Structural Information Modeling

Authors: Byungmin Cho, Dongchul Lee, Taejin Kim, Minhee Lee

Abstract:

The procedure for the seismic retrofit of existing buildings includes the seismic evaluation. In the evaluation step, it is assessed whether the buildings have satisfactory performance against seismic load. Based on the results of that, the buildings are upgraded. To evaluate seismic performance of the buildings, it usually goes through the model transformation from elastic analysis to inelastic analysis. However, when the data is not delivered through the interwork, engineers should manually input the data. In this process, since it leads to inaccuracy and loss of information, the results of the analysis become less accurate. Therefore, in this study, the process for the seismic evaluation of existing buildings using structural information modeling is suggested. This structural information modeling makes the work economic and accurate. To this end, it is determined which part of the process could be computerized through the investigation of the process for the seismic evaluation based on ASCE 41. The structural information modeling process is developed to apply to the seismic evaluation using Perform 3D program usually used for the nonlinear response history analysis. To validate this process, the seismic performance of an existing building is investigated.

Keywords: existing building, nonlinear analysis, seismic performance, structural information modeling

Procedia PDF Downloads 352
15422 Understanding Tacit Knowledge and DIKW

Authors: Bahadir Aydin

Abstract:

Today it is difficult to reach accurate knowledge because of mass data. This huge data makes the environment more and more caotic. Data is a main piller of intelligence. There is a close tie between knowledge and intelligence. Information gathered from different sources can be modified, interpreted and classified by using knowledge development process. This process is applied in order to attain intelligence. Within this process the effect of knowledge is crucial. Knowledge is classified as explicit and tacit knowledge. Tacit knowledge can be seen as "only the tip of the iceberg”. This tacit knowledge accounts for much more than we guess in all intelligence cycle. If the concept of intelligence scrutinized, it can be seen that it contains risks, threats as well as success. The main purpose for all organization is to be succesful by eliminating risks and threats. Therefore, there is a need to connect or fuse existing information and the processes which can be used to develop it. By the help of process the decision-maker can be presented with a clear holistic understanding, as early as possible in the decision making process. Planning, execution and assessments are the key functions that connects to information to knowledge. Altering from the current traditional reactive approach to a proactive knowledge development approach would reduce extensive duplication of work in the organization. By new approach to this process, knowledge can be used more effectively.

Keywords: knowledge, intelligence cycle, tacit knowledge, KIDW

Procedia PDF Downloads 496
15421 The Extended Skew Gaussian Process for Regression

Authors: M. T. Alodat

Abstract:

In this paper, we propose a generalization to the Gaussian process regression(GPR) model called the extended skew Gaussian process for regression(ESGPr) model. The ESGPR model works better than the GPR model when the errors are skewed. We derive the predictive distribution for the ESGPR model at a new input. Also we apply the ESGPR model to FOREX data and we find that it fits the Forex data better than the GPR model.

Keywords: extended skew normal distribution, Gaussian process for regression, predictive distribution, ESGPr model

Procedia PDF Downloads 525
15420 Multi-Tooled Robotic Hand for Tele-Operation of Explosive Devices

Authors: Faik Derya Ince, Ugur Topgul, Alp Gunay, Can Bayoglu, Dante J. Dorantes-Gonzalez

Abstract:

Explosive attacks are arguably the most lethal threat that may occur in terrorist attacks. In order to counteract this issue, explosive ordnance disposal operators put their lives on the line to dispose of a possible improvised explosive device. Robots can make the disposal process more accurately and saving human lives. For this purpose, there is a demand for more accurate and dexterous manipulating robotic hands that can be teleoperated from a distance. The aim of this project is to design a robotic hand that contains two active and two passive DOF for each finger, as well as a minimum set of tools for mechanical cutting and screw driving within the same robotic hand. Both hand and toolset, are teleoperated from a distance from a haptic robotic glove in order to manipulate dangerous objects such as improvised explosive devices. SolidWorks® Computer-Aided Design, computerized dynamic simulation, and MATLAB® kinematic and static analysis were used for the robotic hand and toolset design. Novel, dexterous and robust solutions for the fingers were obtained, and six servo motors are used in total to remotely control the multi-tooled robotic hand. This project is still undergoing and presents currents results. Future research steps are also presented.

Keywords: Explosive Manipulation, Robotic Hand, Tele-Operation, Tool Integration

Procedia PDF Downloads 115
15419 Simulation of the Reactive Rotational Molding Using Smoothed Particle Hydrodynamics

Authors: A. Hamidi, S. Khelladi, L. Illoul, A. Tcharkhtchi

Abstract:

Reactive rotational molding (RRM) is a process to manufacture hollow plastic parts with reactive material has several advantages compared to conventional roto molding of thermoplastic powders: process cycle time is shorter; raw material is less expensive because polymerization occurs during processing and high-performance polymers may be used such as thermosets, thermoplastics or blends. However, several phenomena occur during this process which makes the optimization of the process quite complex. In this study, we have used a mixture of isocyanate and polyol as a reactive system. The chemical transformation of this system to polyurethane has been studied by thermal analysis and rheology tests. Thanks to these results of the curing process and rheological measurements, the kinetic and rheokinetik of polyurethane was identified. Smoothed Particle Hydrodynamics, a Lagrangian meshless method, was chosen to simulate reactive fluid flow in 2 and 3D configurations of the polyurethane during the process taking into account the chemical, and chemiorehological results obtained experimentally in this study.

Keywords: reactive rotational molding, simulation, smoothed particle hydrodynamics, surface tension, rheology, free surface flows, viscoelastic, interpolation

Procedia PDF Downloads 262
15418 Buoyancy Effects in Pressure Retarded Osmosis with Extremely High Draw Solution Concentration

Authors: Ivonne Tshuma, Ralf Cord-Ruwisch, Wendell Ela

Abstract:

Water crisis is a world-wide problem because of population growth and climate change. Hence, desalination is a solution to water scarcity, which threatens the world. Reverse osmosis (RO) is the most used technique for desalination; unfortunately, this process, usually requires high-pressure requirement hence requires a lot of energy about 3 – 5.5 KWhr/m³ of electrical energy. The pressure requirements of RO can be alleviated by the use of PRO (pressure retarded osmosis) to drive the RO process. This paper proposes a process of utilizing the energy directly from PRO to drive an RO process. The paper mostly analyses the PRO process parameters such as cross-flow velocity, density, and buoyancy and how these have an effect on PRO hence ultimately the RO process. The experimental study of the PRO with various feed solution concentrations and cross-flow velocities at fixed applied pressure with different orientations of the PRO cell was performed. The study revealed that without cross-flow velocity, buoyancy effects were observed but not with cross-flow velocity.

Keywords: cross-flow velocity, pressure retarded osmosis, density, buoyancy

Procedia PDF Downloads 117
15417 Fault-Tolerant Control Study and Classification: Case Study of a Hydraulic-Press Model Simulated in Real-Time

Authors: Jorge Rodriguez-Guerra, Carlos Calleja, Aron Pujana, Iker Elorza, Ana Maria Macarulla

Abstract:

Society demands more reliable manufacturing processes capable of producing high quality products in shorter production cycles. New control algorithms have been studied to satisfy this paradigm, in which Fault-Tolerant Control (FTC) plays a significant role. It is suitable to detect, isolate and adapt a system when a harmful or faulty situation appears. In this paper, a general overview about FTC characteristics are exposed; highlighting the properties a system must ensure to be considered faultless. In addition, a research to identify which are the main FTC techniques and a classification based on their characteristics is presented in two main groups: Active Fault-Tolerant Controllers (AFTCs) and Passive Fault-Tolerant Controllers (PFTCs). AFTC encompasses the techniques capable of re-configuring the process control algorithm after the fault has been detected, while PFTC comprehends the algorithms robust enough to bypass the fault without further modifications. The mentioned re-configuration requires two stages, one focused on detection, isolation and identification of the fault source and the other one in charge of re-designing the control algorithm by two approaches: fault accommodation and control re-design. From the algorithms studied, one has been selected and applied to a case study based on an industrial hydraulic-press. The developed model has been embedded under a real-time validation platform, which allows testing the FTC algorithms and analyse how the system will respond when a fault arises in similar conditions as a machine will have on factory. One AFTC approach has been picked up as the methodology the system will follow in the fault recovery process. In a first instance, the fault will be detected, isolated and identified by means of a neural network. In a second instance, the control algorithm will be re-configured to overcome the fault and continue working without human interaction.

Keywords: fault-tolerant control, electro-hydraulic actuator, fault detection and isolation, control re-design, real-time

Procedia PDF Downloads 150
15416 Medical Diagnosis of Retinal Diseases Using Artificial Intelligence Deep Learning Models

Authors: Ethan James

Abstract:

Over one billion people worldwide suffer from some level of vision loss or blindness as a result of progressive retinal diseases. Many patients, particularly in developing areas, are incorrectly diagnosed or undiagnosed whatsoever due to unconventional diagnostic tools and screening methods. Artificial intelligence (AI) based on deep learning (DL) convolutional neural networks (CNN) have recently gained a high interest in ophthalmology for its computer-imaging diagnosis, disease prognosis, and risk assessment. Optical coherence tomography (OCT) is a popular imaging technique used to capture high-resolution cross-sections of retinas. In ophthalmology, DL has been applied to fundus photographs, optical coherence tomography, and visual fields, achieving robust classification performance in the detection of various retinal diseases including macular degeneration, diabetic retinopathy, and retinitis pigmentosa. However, there is no complete diagnostic model to analyze these retinal images that provide a diagnostic accuracy above 90%. Thus, the purpose of this project was to develop an AI model that utilizes machine learning techniques to automatically diagnose specific retinal diseases from OCT scans. The algorithm consists of neural network architecture that was trained from a dataset of over 20,000 real-world OCT images to train the robust model to utilize residual neural networks with cyclic pooling. This DL model can ultimately aid ophthalmologists in diagnosing patients with these retinal diseases more quickly and more accurately, therefore facilitating earlier treatment, which results in improved post-treatment outcomes.

Keywords: artificial intelligence, deep learning, imaging, medical devices, ophthalmic devices, ophthalmology, retina

Procedia PDF Downloads 152
15415 Stereo Camera Based Speed-Hump Detection Process for Real Time Driving Assistance System in the Daytime

Authors: Hyun-Koo Kim, Yong-Hun Kim, Soo-Young Suk, Ju H. Park, Ho-Youl Jung

Abstract:

This paper presents an effective speed hump detection process at the day-time. we focus only on round types of speed humps in the day-time dynamic road environment. The proposed speed hump detection scheme consists mainly of two process as stereo matching and speed hump detection process. Our proposed process focuses to speed hump detection process. Speed hump detection process consist of noise reduction step, data fusion step, and speed hemp detection step. The proposed system is tested on Intel Core CPU with 2.80 GHz and 4 GB RAM tested in the urban road environments. The frame rate of test videos is 30 frames per second and the size of each frame of grabbed image sequences is 1280 pixels by 670 pixels. Using object-marked sequences acquired with an on-vehicle camera, we recorded speed humps and non-speed humps samples. Result of the tests, our proposed method can be applied in real-time systems by computation time is 13 ms. For instance; our proposed method reaches 96.1 %.

Keywords: data fusion, round types speed hump, speed hump detection, surface filter

Procedia PDF Downloads 491
15414 Optimizing the Passenger Throughput at an Airport Security Checkpoint

Authors: Kun Li, Yuzheng Liu, Xiuqi Fan

Abstract:

High-security standard and high efficiency of screening seem to be contradictory to each other in the airport security check process. Improving the efficiency as far as possible while maintaining the same security standard is significantly meaningful. This paper utilizes the knowledge of Operation Research and Stochastic Process to establish mathematical models to explore this problem. We analyze the current process of airport security check and use the M/G/1 and M/G/k models in queuing theory to describe the process. Then we find the least efficient part is the pre-check lane, the bottleneck of the queuing system. To improve passenger throughput and reduce the variance of passengers’ waiting time, we adjust our models and use Monte Carlo method, then put forward three modifications: adjust the ratio of Pre-Check lane to regular lane flexibly, determine the optimal number of security check screening lines based on cost analysis and adjust the distribution of arrival and service time based on Monte Carlo simulation results. We also analyze the impact of cultural differences as the sensitivity analysis. Finally, we give the recommendations for the current process of airport security check process.

Keywords: queue theory, security check, stochatic process, Monte Carlo simulation

Procedia PDF Downloads 171
15413 Automated CNC Part Programming and Process Planning for Turned Components

Authors: Radhey Sham Rajoria

Abstract:

Pressure to increase the competitiveness in the manufacturing sector and for the survival in the market has led to the development of machining centres, which enhance productivity, improve quality, shorten the lead time, and reduce the manufacturing cost. With the innovation of machining centres in the manufacturing sector the production lines have been replaced by these machining centers, having the ability to machine various processes and multiple tooling with automatic tool changer (ATC) for the same part. Also the process plans can be easily generated for complex components. Some means are required to utilize the machining center at its best. The present work is concentrated on the automated part program generation, and in turn automated process plan generation for the turned components on Denford “MIRAC” 8 stations ATC lathe machining centre. A package in C++ on DOS platform is developed which generates the complete CNC part program, process plan and process sequence for the turned components. The input to this system is in the form of a blueprint in graphical format with machining parameters and variables, and the output is the CNC part program which is stored in a .mir file, ready for execution on the machining centre.

Keywords: CNC, MIRAC, ATC, process planning

Procedia PDF Downloads 240
15412 Application of Lean Six Sigma Tools to Minimize Time and Cost in Furniture Packaging

Authors: Suleiman Obeidat, Nabeel Mandahawi

Abstract:

In this work, the packaging process for a move is improved. The customers of this move need their household stuff to be moved from their current house to the new one with minimum damage, in an organized manner, on time and with the minimum cost. Our goal was to improve the process between 10% and 20% time efficiency, 90% reduction in damaged parts and an acceptable improvement in the cost of the total move process. The expected ROI was 833%. Many improvement techniques have been used in terms of the way the boxes are prepared, their preparation cost, packing the goods, labeling them and moving them to a place for moving out. DMAIC technique is used in this work: SIPOC diagram, value stream map of “As Is” process, Root Cause Analysis, Maps of “Future State” and “Ideal State” and an Improvement Plan. A value of ROI=624% is obtained which is lower than the expected value of 833%. The work explains the techniques of improvement and the deficiencies in the old process.

Keywords: packaging, lean tools, six sigma, DMAIC methodology, SIPOC

Procedia PDF Downloads 397
15411 Facile Synthesis and Structure Characterization of Europium (III) Tungstate Nanoparticles

Authors: Mehdi Rahimi-Nasrabadi, Seied Mahdi Pourmortazavi

Abstract:

Taguchi robust design as a statistical method was applied for optimization of the process parameters in order to tunable, simple and fast synthesis of europium (III) tungstate nanoparticles. Europium (III) tungstate nanoparticles were synthesized by a chemical precipitation reaction involving direct addition of europium ion aqueous solution to the tungstate reagent solved in aqueous media. Effects of some synthesis procedure variables i.e., europium and tungstate concentrations, flow rate of cation reagent addition, and temperature of reaction reactor on the particle size of europium (III) tungstate nanoparticles were studied experimentally in order to tune particle size of europium (III) tungstate. Analysis of variance shows the importance of controlling tungstate concentration, cation feeding flow rate and temperature for preparation of europium (III) tungstate nanoparticles by the proposed chemical precipitation reaction. Finally, europium (III) tungstate nanoparticles were synthesized at the optimum conditions of the proposed method and the morphology and chemical composition of the prepared nano-material were characterized by means of X-Ray diffraction, scanning electron microscopy, transmission electron microscopy, FT-IR spectroscopy, and fluorescence.

Keywords: europium (III) tungstate, nano-material, particle size control, procedure optimization

Procedia PDF Downloads 362
15410 Accelerated Aging of Photopolymeric Material Used in Flexography

Authors: S. Mahovic Poljacek, T. Tomasegovic, T. Cigula, D. Donevski, R. Szentgyörgyvölgyi, S. Jakovljevic

Abstract:

In this paper, a degradation of the photopolymeric material (PhPM), used as printing plate in the flexography reproduction technique, caused by accelerated aging has been observed. Since the basis process for production of printing plates from the PhPM is a radical cross-linking process caused by exposing to UV wavelengths, the assumption was that improper storage or irregular handling of the PhPM plate can change the surface and structure characteristics of the plates. Results have shown that the aging process causes degradation in the structure and changes in the surface of the PhPM printing plate.

Keywords: aging process, artificial treatment, flexography, photopolymeric material (PhPM)

Procedia PDF Downloads 328
15409 A Detailed Experimental Study and Evaluation of Springback under Stretch Bending Process

Authors: A. Soualem

Abstract:

The design of multi stage deep drawing processes requires the evaluation of many process parameters such as the intermediate die geometry, the blank shape, the sheet thickness, the blank holder force, friction, lubrication etc..These process parameters have to be determined for the optimum forming conditions before the process design. In general sheet metal forming may involve stretching drawing or various combinations of these basic modes of deformation. It is important to determine the influence of the process variables in the design of sheet metal working process. Especially, the punch and die corner for deep drawing will affect the formability. At the same time the prediction of sheet metals springback after deep drawing is an important issue to solve for the control of manufacturing processes. Nowadays, the importance of this problem increases because of the use of steel sheeting with high stress and also aluminum alloys. The aim of this paper is to give a better understanding of the springback and its effect in various sheet metals forming process such as expansion and restraint deep drawing in the cup drawing process, by varying radius die, lubricant for two commercially available materials e.g. galvanized steel and Aluminum sheet. To achieve these goals experiments were carried out and compared with other results. The original of our purpose consist on tests which are ensured by adapting a U-type stretching-bending device on a tensile testing machine, where we studied and quantified the variation of the springback.

Keywords: springback, deep drawing, expansion, restricted deep drawing

Procedia PDF Downloads 427
15408 Islamic Banking Recovery Process and Its Parameters: A Practitioner’s Viewpoints in the Light of Humanising Financial Services

Authors: Muhammad Izzam Bin Mohd Khazar, Nur Adibah Binti Zainudin

Abstract:

Islamic banking as one of the financial institutions is highly required to maintain a prudent approach to ensure that any financing given is able to generate income to their respective shareholders. As the default payment of customers is probably occurred in the financing, having a prudent approach in the recovery process is a must to ensure that financing losses are within acceptable limits. The objective of this research is to provide the best practice of recovery which is anticipated to benefit both bank and customers. This study will address arising issue on the current practice of recovery process and followed by providing humanising recovery solutions in the light of the Maqasid Shariah. The study identified main issues pertaining to Islamic recovery process which can be categorized into knowledge crisis, process issues, specific treatment cases and system issues. Knowledge crisis is related to direct parties including judges, solicitors and salesperson, while the recovery process issues include the process of issuance of reminder, foreclosure and repossession of asset. Furthermore, special treatment for particular cases also should be observed since different contracts in Islamic banking products will need different treatment. Finally, issues in the system used in the recovery process are still unresolved since the existing technology is still young in this area to embraced Islamic finance requirements and nature of calculation. In order to humanize the financial services in Islamic banking recovery process, we have highlighted four main recommendation to be implemented by Islamic Financial Institutions namely; 1) early deterrent by improving the awareness, 2) improvement of the internal process, 3) reward mechanism, and 4) creative penalty to provide awareness to all stakeholders.

Keywords: humanizing financial services, Islamic Finance, Maqasid Syariah, recovery process

Procedia PDF Downloads 163
15407 A Comparative Study on ANN, ANFIS and SVM Methods for Computing Resonant Frequency of A-Shaped Compact Microstrip Antennas

Authors: Ahmet Kayabasi, Ali Akdagli

Abstract:

In this study, three robust predicting methods, namely artificial neural network (ANN), adaptive neuro fuzzy inference system (ANFIS) and support vector machine (SVM) were used for computing the resonant frequency of A-shaped compact microstrip antennas (ACMAs) operating at UHF band. Firstly, the resonant frequencies of 144 ACMAs with various dimensions and electrical parameters were simulated with the help of IE3D™ based on method of moment (MoM). The ANN, ANFIS and SVM models for computing the resonant frequency were then built by considering the simulation data. 124 simulated ACMAs were utilized for training and the remaining 20 ACMAs were used for testing the ANN, ANFIS and SVM models. The performance of the ANN, ANFIS and SVM models are compared in the training and test process. The average percentage errors (APE) regarding the computed resonant frequencies for training of the ANN, ANFIS and SVM were obtained as 0.457%, 0.399% and 0.600%, respectively. The constructed models were then tested and APE values as 0.601% for ANN, 0.744% for ANFIS and 0.623% for SVM were achieved. The results obtained here show that ANN, ANFIS and SVM methods can be successfully applied to compute the resonant frequency of ACMAs, since they are useful and versatile methods that yield accurate results.

Keywords: a-shaped compact microstrip antenna, artificial neural network (ANN), adaptive neuro-fuzzy inference system (ANFIS), support vector machine (SVM)

Procedia PDF Downloads 415
15406 Exploring Data Leakage in EEG Based Brain-Computer Interfaces: Overfitting Challenges

Authors: Khalida Douibi, Rodrigo Balp, Solène Le Bars

Abstract:

In the medical field, applications related to human experiments are frequently linked to reduced samples size, which makes the training of machine learning models quite sensitive and therefore not very robust nor generalizable. This is notably the case in Brain-Computer Interface (BCI) studies, where the sample size rarely exceeds 20 subjects or a few number of trials. To address this problem, several resampling approaches are often used during the data preparation phase, which is an overly critical step in a data science analysis process. One of the naive approaches that is usually applied by data scientists consists in the transformation of the entire database before the resampling phase. However, this can cause model’ s performance to be incorrectly estimated when making predictions on unseen data. In this paper, we explored the effect of data leakage observed during our BCI experiments for device control through the real-time classification of SSVEPs (Steady State Visually Evoked Potentials). We also studied potential ways to ensure optimal validation of the classifiers during the calibration phase to avoid overfitting. The results show that the scaling step is crucial for some algorithms, and it should be applied after the resampling phase to avoid data leackage and improve results.

Keywords: data leackage, data science, machine learning, SSVEP, BCI, overfitting

Procedia PDF Downloads 125
15405 A Comparison between Fuzzy Analytic Hierarchy Process and Fuzzy Analytic Network Process for Rationality Evaluation of Land Use Planning Locations in Vietnam

Authors: X. L. Nguyen, T. Y. Chou, F. Y. Min, F. C. Lin, T. V. Hoang, Y. M. Huang

Abstract:

In Vietnam, land use planning is utilized as an efficient tool for the local government to adjust land use. However, planned locations are facing disapproval from people who live near these planned sites because of environmental problems. The selection of these locations is normally based on the subjective opinion of decision-makers and is not supported by any scientific methods. Many researchers have applied Multi-Criteria Analysis (MCA) methods in which Analytic Hierarchy Process (AHP) is the most popular techniques in combination with Fuzzy set theory for the subject of rationality assessment of land use planning locations. In this research, the Fuzzy set theory and Analytic Network Process (ANP) multi-criteria-based technique were used for the assessment process. The Fuzzy Analytic Hierarchy Process was also utilized, and the output results from two methods were compared to extract the differences. The 20 planned landfills in Hung Ha district, Thai Binh province, Vietnam was selected as a case study. The comparison results indicate that there are different between weights computed by AHP and ANP methods and the assessment outputs produced from these two methods also slight differences. After evaluation of existing planned sites, some potential locations were suggested to the local government for possibility of land use planning adjusts.

Keywords: Analytic Hierarchy Process, Analytic Network Process, Fuzzy set theory, land use planning

Procedia PDF Downloads 386
15404 Detection of Clipped Fragments in Speech Signals

Authors: Sergei Aleinik, Yuri Matveev

Abstract:

In this paper a novel method for the detection of clipping in speech signals is described. It is shown that the new method has better performance than known clipping detection methods, is easy to implement, and is robust to changes in signal amplitude, size of data, etc. Statistical simulation results are presented.

Keywords: clipping, clipped signal, speech signal processing, digital signal processing

Procedia PDF Downloads 368