Search results for: optimization process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17266

Search results for: optimization process

16516 Restoration and Conservation of Historical Textiles Using Covalently Immobilized Enzymes on Nanoparticles

Authors: Mohamed Elbehery

Abstract:

Historical textiles in the burial environment or in museums are exposed to many types of stains and dirt that are associated with historical textiles by multiple chemical bonds that cause damage to historical textiles. The cleaning process must be carried out with great care, with no irreversible damage, and sediments removed without affecting the original material of the surface being cleaned. Science and technology continue to provide innovative systems in the bio-cleaning process (using pure enzymes) of historical textiles and artistic surfaces. Lipase and α-amylase were immobilized on nanoparticles of alginate/κ-carrageenan nanoparticle complex and used in historical textiles cleaning. Preparation of nanoparticles, activation, and enzymes immobilization were characterized. Optimization of loading time and units of the two enzymes were done. It was found that, the optimum time and units of amylase were 4 hrs and 25U, respectively. While, the optimum time and units of lipase were 3 hrs and 15U, respectively. The methods used to examine the fibers using a scanning electron microscope equipped with an X-ray energy dispersal unit: SEM with EDX unit.

Keywords: nanoparticles, enzymes, immobilization, textiles

Procedia PDF Downloads 93
16515 Development of Algorithms for Solving and Analyzing Special Problems Transports Type

Authors: Dmitri Terzi

Abstract:

The article presents the results of an algorithmic study of a special optimization problem of the transport type (traveling salesman problem): 1) To solve the problem, a new natural algorithm has been developed based on the decomposition of the initial data into convex hulls, which has a number of advantages; it is applicable for a fairly large dimension, does not require a large amount of memory, and has fairly good performance. The relevance of the algorithm lies in the fact that, in practice, programs for problems with the number of traversal points of no more than twenty are widely used. For large-scale problems, the availability of algorithms and programs of this kind is difficult. The proposed algorithm is natural because the optimal solution found by the exact algorithm is not always feasible due to the presence of many other factors that may require some additional restrictions. 2) Another inverse problem solved here is to describe a class of traveling salesman problems that have a predetermined optimal solution. The constructed algorithm 2 allows us to characterize the structure of traveling salesman problems, as well as construct test problems to evaluate the effectiveness of algorithms and other purposes. 3) The appendix presents a software implementation of Algorithm 1 (in MATLAB), which can be used to solve practical problems, as well as in the educational process on operations research and optimization methods.

Keywords: traveling salesman problem, solution construction algorithm, convex hulls, optimality verification

Procedia PDF Downloads 60
16514 The Possibility to Assess the Industrial Enterprise Sustainability

Authors: G. Khasaev, S. Ashmarina , A. Zotova

Abstract:

The priority of Russian enterprises development has been given to the optimization process of industrial enterprise activity for their sustainable development in a long-term period. The assessment of sustainable development level as one of the most efficient instruments of sustainable development management at the industrial enterprise gives a complex view of its state. In order to perform accurate analysis of the current state of the industrial enterprise, it is necessary to perform the assessment of its sustainable development and using its results to elaborate the further tactic of enterprise functioning. The assessment of sustainable development level of the enterprise may help the effective management of strategy development only if the corresponding indicators system is created. The elaboration and usage the sustainable development indicators allows the enterprise to implement analysis of its activity results and monitoring of sustainable enterprise functioning. The authors’ methods are based on general aspects of the industrial enterprise functioning such as finance, customers, inner economic process, and staff system.

Keywords: assessment methods, indicators system, industrial enterprise, sustainable development

Procedia PDF Downloads 359
16513 Statistical Analysis and Optimization of a Process for CO2 Capture

Authors: Muftah H. El-Naas, Ameera F. Mohammad, Mabruk I. Suleiman, Mohamed Al Musharfy, Ali H. Al-Marzouqi

Abstract:

CO2 capture and storage technologies play a significant role in contributing to the control of climate change through the reduction of carbon dioxide emissions into the atmosphere. The present study evaluates and optimizes CO2 capture through a process, where carbon dioxide is passed into pH adjusted high salinity water and reacted with sodium chloride to form a precipitate of sodium bicarbonate. This process is based on a modified Solvay process with higher CO2 capture efficiency, higher sodium removal, and higher pH level without the use of ammonia. The process was tested in a bubble column semi-batch reactor and was optimized using response surface methodology (RSM). CO2 capture efficiency and sodium removal were optimized in terms of major operating parameters based on four levels and variables in Central Composite Design (CCD). The operating parameters were gas flow rate (0.5–1.5 L/min), reactor temperature (10 to 50 oC), buffer concentration (0.2-2.6%) and water salinity (25-197 g NaCl/L). The experimental data were fitted to a second-order polynomial using multiple regression and analyzed using analysis of variance (ANOVA). The optimum values of the selected variables were obtained using response optimizer. The optimum conditions were tested experimentally using desalination reject brine with salinity ranging from 65,000 to 75,000 mg/L. The CO2 capture efficiency in 180 min was 99% and the maximum sodium removal was 35%. The experimental and predicted values were within 95% confidence interval, which demonstrates that the developed model can successfully predict the capture efficiency and sodium removal using the modified Solvay method.

Keywords: CO2 capture, water desalination, Response Surface Methodology, bubble column reactor

Procedia PDF Downloads 282
16512 Optimizing Boiler Combustion System in a Petrochemical Plant Using Neuro-Fuzzy Inference System and Genetic Algorithm

Authors: Yul Y. Nazaruddin, Anas Y. Widiaribowo, Satriyo Nugroho

Abstract:

Boiler is one of the critical unit in a petrochemical plant. Steam produced by the boiler is used for various processes in the plant such as urea and ammonia plant. An alternative method to optimize the boiler combustion system is presented in this paper. Adaptive Neuro-Fuzzy Inference System (ANFIS) approach is applied to model the boiler using real-time operational data collected from a boiler unit of the petrochemical plant. Nonlinear equation obtained is then used to optimize the air to fuel ratio using Genetic Algorithm, resulting an optimal ratio of 15.85. This optimal ratio is then maintained constant by ratio controller designed using inverse dynamics based on ANFIS. As a result, constant value of oxygen content in the flue gas is obtained which indicates more efficient combustion process.

Keywords: ANFIS, boiler, combustion process, genetic algorithm, optimization.

Procedia PDF Downloads 246
16511 Application of Medium High Hydrostatic Pressure in Preserving Textural Quality and Safety of Pineapple Compote

Authors: Nazim Uddin, Yohiko Nakaura, Kazutaka Yamamoto

Abstract:

Compote (fruit in syrup) of pineapple (Ananas comosus L. Merrill) is expected to have a high market potential as one of convenient ready-to-eat (RTE) foods worldwide. High hydrostatic pressure (HHP) in combination with low temperature (LT) was applied to the processing of pineapple compote as well as medium HHP (MHHP) in combination with medium-high temperature (MHT) since both processes can enhance liquid impregnation and inactivate microbes. MHHP+MHT (55 or 65 °C) process, as well as the HHP+LT process, has successfully inactivated the microbes in the compote to a non-detectable level. Although the compotes processed by MHHP+MHT or HHP+LT have lost the fresh texture as in a similar manner as those processed solely by heat, it was indicated that the texture degradations by heat were suppressed under MHHP. Degassing process reduced the hardness, while calcium (Ca) contributed to be retained hardness in MHT and MHHP+MHT processes. Electrical impedance measurement supported the damage due to degassing and heat. The color, Brix, and appearance were not affected by the processing methods significantly. MHHP+MHT and HHP+LT processes may be applicable to produce high-quality, safe RTE pineapple compotes. Further studies on the optimization of packaging and storage condition will be indispensable for commercialization.

Keywords: compote of pineapple, RTE, medium high hydrostatic pressure, postharvest loss, texture

Procedia PDF Downloads 130
16510 Design an Intelligent Fire Detection System Based on Neural Network and Particle Swarm Optimization

Authors: Majid Arvan, Peyman Beygi, Sina Rokhsati

Abstract:

In-time detection of fire in buildings is of great importance. Employing intelligent methods in data processing in fire detection systems leads to a significant reduction of fire damage at lowest cost. In this paper, the raw data obtained from the fire detection sensor networks in buildings is processed by using intelligent methods based on neural networks and the likelihood of fire happening is predicted. In order to enhance the quality of system, the noise in the sensor data is reduced by analyzing wavelets and applying SVD technique. Meanwhile, the proposed neural network is trained using particle swarm optimization (PSO). In the simulation work, the data is collected from sensor network inside the room and applied to the proposed network. Then the outputs are compared with conventional MLP network. The simulation results represent the superiority of the proposed method over the conventional one.

Keywords: intelligent fire detection, neural network, particle swarm optimization, fire sensor network

Procedia PDF Downloads 376
16509 Design and Characterization of a CMOS Process Sensor Utilizing Vth Extractor Circuit

Authors: Rohana Musa, Yuzman Yusoff, Chia Chieu Yin, Hanif Che Lah

Abstract:

This paper presents the design and characterization of a low power Complementary Metal Oxide Semiconductor (CMOS) process sensor. The design is targeted for implementation using Silterra’s 180 nm CMOS process technology. The proposed process sensor employs a voltage threshold (Vth) extractor architecture for detection of variations in the fabrication process. The process sensor generates output voltages in the range of 401 mV (fast-fast corner) to 443 mV (slow-slow corner) at nominal condition. The power dissipation for this process sensor is 6.3 µW with a supply voltage of 1.8V with a silicon area of 190 µm X 60 µm. The preliminary result of this process sensor that was fabricated indicates a close resemblance between test and simulated results.

Keywords: CMOS process sensor, PVT sensor, threshold extractor circuit, Vth extractor circuit

Procedia PDF Downloads 169
16508 Effect of Ultrasound on Carotenoids Extraction from Pepper and Process Optimization Using Response Surface Methodology (RSM)

Authors: Elham Mahdian, Reza Karazhian, Rahele Dehghan Tanha

Abstract:

Pepper (Capsicum annum L.) which belong to the family Solananceae, are known for their versatility as a vegetable crop and are consumed both as fresh vegetables or dehydrated for spices. Pepper is considered an excellent source of bioactive nutrients. Ascorbic acid, carotenoids and phenolic compounds are its main antioxidant constituents. Ultrasound assisted extraction is an inexpensive, simple and efficient alternative to conventional extraction techniques. The mechanism of action for ultrasound-assisted extraction are attributed to cavitations, mechanical forces and thermal impact, which result in disruption of cells walls, reduce particle size, and enhance mass transfer across cell membranes. In this study, response surface methodology was used to optimize experimental conditions for ultrasonic assisted extraction of carotenoid compounds from Chili peppers. Variables were included extraction temperatures at 3 levels (30, 40 and 50 °C), extraction times at 3 levels (10, 25 and 40 minutes) and power at 3 levels (30, 60 and 90 %). It was observed that ultrasound waves applied at temperature of 49°C, time of 10 minutes and power 89 % resulted to the highest carotenoids contents (lycopene and β-carotene), while the lowest value was recorded in the control. Thus, results showed that ultrasound waves have strong impact on extraction of carotenoids from pepper.

Keywords: carotenoids, optimization, pepper, response surface methodology

Procedia PDF Downloads 467
16507 Business Process Mashup

Authors: Fethia Zenak, Salima Benbernou, Linda Zaoui

Abstract:

Recently, many companies are based on process development from scratch to achieve their business goals. The process development is not trivial and the main objective of enterprise managing processes is to decrease the software development time. Several concepts have been proposed in the field of business process-based reused development, known as BP Mashup. This concept consists of reusing existing business processes which have been modeled in order to respond to a particular goal. To meet user process requirements, our contribution is to mix parts of processes as 'processes fragments' components to build a new process (i.e. process mashup). The main idea of our paper is to offer graphical framework tool for both creating and running processes mashup. Allow users to perform a mixture of fragments, using a simple interface with set of graphical mixture operators based on a proposed formal model. A process mashup and mixture behavior are described within a new specification of a high-level language, language for process mashup (BPML).

Keywords: business process, mashup, fragments, bp mashup

Procedia PDF Downloads 627
16506 Improvement in Drying Characteristics of Raisin by Carbonic Maceration– Process Optimization

Authors: Nursac Akyol, Merve S. Turan, Mustafa Ozcelik, Erdogan Kucukoner, Erkan Karacabey

Abstract:

Traditional raisin production is a long time drying process under sunlight. During this procedure, grapes are open to some environmental effects besides the adverse effects of the long drying period. Thus, there is a need to develop an alternative method being applicable instead of traditional one. To this extent, a combination of a potential pretreatment (carbonic maceration, CM) with convectional oven drying was examined. CM application was used in raisin production (grape drying) as a pretreatment process before oven drying. Pressure, temperature and time were examined as application parameters of CM. In conventional oven drying, the temperature is a process variable. The aim is to find out how CM and convectional drying processes affect the drying characteristics of grapes as well as their physical and chemical properties. For this purpose, the response surface method was used to determine both the effects of the variables and the optimum pretreatment and drying conditions. The optimum conditions of CM for raisin production were 0.3 MPa of pressure value, 4°C of application temperature and 8 hours of application time. The optimized drying temperature was 77°C. The results showed that the application of CM before the drying process improved the drying characteristics. Drying took only 389 minutes for grapes pretreated by CM under optimum conditions and 495 minutes for the control group dried only by the conventional drying process. According to these results, a decrease of 21% was achieved in the time requirement for raisin production. Also, it was observed that the samples dried under optimum conditions had similar physical properties as those the control group had. It was seen that raisin, which was dried under optimum conditions were in better condition in terms of some of the bioactive contents compared to control groups. In light of all results, it is seen that CM has an important potential in the industrial drying of grape samples. The current study was financially supported by TUBITAK, Turkey (Project no: 116R038).

Keywords: drying time, pretreatment, response surface methodlogy, total phenolic

Procedia PDF Downloads 128
16505 Robust Batch Process Scheduling in Pharmaceutical Industries: A Case Study

Authors: Tommaso Adamo, Gianpaolo Ghiani, Antonio Domenico Grieco, Emanuela Guerriero

Abstract:

Batch production plants provide a wide range of scheduling problems. In pharmaceutical industries a batch process is usually described by a recipe, consisting of an ordering of tasks to produce the desired product. In this research work we focused on pharmaceutical production processes requiring the culture of a microorganism population (i.e. bacteria, yeasts or antibiotics). Several sources of uncertainty may influence the yield of the culture processes, including (i) low performance and quality of the cultured microorganism population or (ii) microbial contamination. For these reasons, robustness is a valuable property for the considered application context. In particular, a robust schedule will not collapse immediately when a cell of microorganisms has to be thrown away due to a microbial contamination. Indeed, a robust schedule should change locally in small proportions and the overall performance measure (i.e. makespan, lateness) should change a little if at all. In this research work we formulated a constraint programming optimization (COP) model for the robust planning of antibiotics production. We developed a discrete-time model with a multi-criteria objective, ordering the different criteria and performing a lexicographic optimization. A feasible solution of the proposed COP model is a schedule of a given set of tasks onto available resources. The schedule has to satisfy tasks precedence constraints, resource capacity constraints and time constraints. In particular time constraints model tasks duedates and resource availability time windows constraints. To improve the schedule robustness, we modeled the concept of (a, b) super-solutions, where (a, b) are input parameters of the COP model. An (a, b) super-solution is one in which if a variables (i.e. the completion times of a culture tasks) lose their values (i.e. cultures are contaminated), the solution can be repaired by assigning these variables values with a new values (i.e. the completion times of a backup culture tasks) and at most b other variables (i.e. delaying the completion of at most b other tasks). The efficiency and applicability of the proposed model is demonstrated by solving instances taken from Sanofi Aventis, a French pharmaceutical company. Computational results showed that the determined super-solutions are near-optimal.

Keywords: constraint programming, super-solutions, robust scheduling, batch process, pharmaceutical industries

Procedia PDF Downloads 612
16504 Sparsity-Based Unsupervised Unmixing of Hyperspectral Imaging Data Using Basis Pursuit

Authors: Ahmed Elrewainy

Abstract:

Mixing in the hyperspectral imaging occurs due to the low spatial resolutions of the used cameras. The existing pure materials “endmembers” in the scene share the spectra pixels with different amounts called “abundances”. Unmixing of the data cube is an important task to know the present endmembers in the cube for the analysis of these images. Unsupervised unmixing is done with no information about the given data cube. Sparsity is one of the recent approaches used in the source recovery or unmixing techniques. The l1-norm optimization problem “basis pursuit” could be used as a sparsity-based approach to solve this unmixing problem where the endmembers is assumed to be sparse in an appropriate domain known as dictionary. This optimization problem is solved using proximal method “iterative thresholding”. The l1-norm basis pursuit optimization problem as a sparsity-based unmixing technique was used to unmix real and synthetic hyperspectral data cubes.

Keywords: basis pursuit, blind source separation, hyperspectral imaging, spectral unmixing, wavelets

Procedia PDF Downloads 191
16503 The Effect of Initial Sample Size and Increment in Simulation Samples on a Sequential Selection Approach

Authors: Mohammad H. Almomani

Abstract:

In this paper, we argue the effect of the initial sample size, and the increment in simulation samples on the performance of a sequential approach that used in selecting the top m designs when the number of alternative designs is very large. The sequential approach consists of two stages. In the first stage the ordinal optimization is used to select a subset that overlaps with the set of actual best k% designs with high probability. Then in the second stage the optimal computing budget is used to select the top m designs from the selected subset. We apply the selection approach on a generic example under some parameter settings, with a different choice of initial sample size and the increment in simulation samples, to explore the impacts on the performance of this approach. The results show that the choice of initial sample size and the increment in simulation samples does affect the performance of a selection approach.

Keywords: Large Scale Problems, Optimal Computing Budget Allocation, ordinal optimization, simulation optimization

Procedia PDF Downloads 349
16502 Optimization of Parameters for Electrospinning of Pan Nanofibers by Taguchi Method

Authors: Gamze Karanfil Celep, Kevser Dincer

Abstract:

The effects of polymer concentration and electrospinning process parameters on the average diameters of electrospun polyacrylonitrile (PAN) nanofibers were experimentally investigated. Besides, mechanical and thermal properties of PAN nanofibers were examined by tensile test and thermogravimetric analysis (TGA), respectively. For this purpose, the polymer concentration, solution feed rate, supply voltage and tip-to-collector distance were determined as the control factors. To succeed these aims, Taguchi’s L16 orthogonal design (4 parameters, 4 level) was employed for the experimental design. Optimal electrospinning conditions were defined using the signal-to-noise (S/N) ratio that was calculated from diameters of the electrospun PAN nanofibers according to "the-smaller-the-better" approachment. In addition, analysis of variance (ANOVA) was evaluated to conclude the statistical significance of the process parameters. The smallest diameter of PAN nanofibers was observed. According to the S/N ratio response results, the most effective parameter on finding out of nanofiber diameter was determined. Finally, the Taguchi design of experiments method has been found to be an effective method to statistically optimize the critical electrospinning parameters used in nanofiber production. After determining the optimum process parameters of nanofiber production, electrical conductivity and fuel cell performance of electrospun PAN nanofibers on the carbon papers will be evaluated.

Keywords: nanofiber, electrospinning, polyacrylonitrile, Taguchi method

Procedia PDF Downloads 197
16501 Optimization of Fermentation Parameters for Bioethanol Production from Waste Glycerol by Microwave Induced Mutant Escherichia coli EC-MW (ATCC 11105)

Authors: Refal Hussain, Saifuddin M. Nomanbhay

Abstract:

Glycerol is a valuable raw material for the production of industrially useful metabolites. Among many promising applications for the use of glycerol is its bioconversion to high value-added compounds, such as bioethanol through microbial fermentation. Bioethanol is an important industrial chemical with emerging potential as a biofuel to replace vanishing fossil fuels. The yield of liquid fuel in this process was greatly influenced by various parameters viz, temperature, pH, glycerol concentration, organic concentration, and agitation speed were considered. The present study was undertaken to investigate optimum parameters for bioethanol production from raw glycerol by immobilized mutant Escherichia coli (E.coli) (ATCC11505) strain on chitosan cross linked glutaraldehyde optimized by Taguchi statistical method in shake flasks. The initial parameters were set each at four levels and the orthogonal array layout of L16 (45) conducted. The important controlling parameters for optimized the operational fermentation was temperature 38 °C, medium pH 6.5, initial glycerol concentration (250 g/l), and organic source concentration (5 g/l). Fermentation with optimized parameters was carried out in a custom fabricated shake flask. The predicted value of bioethanol production under optimized conditions was (118.13 g/l). Immobilized cells are mainly used for economic benefits of continuous production or repeated use in continuous as well as in batch mode.

Keywords: bioethanol, Escherichia coli, immobilization, optimization

Procedia PDF Downloads 647
16500 Sintering of YNbO3:Eu3+ Compound: Correlation between Luminescence and Spark Plasma Sintering Effect

Authors: Veronique Jubera, Ka-Young Kim, U-Chan Chung, Amelie Veillere, Jean-Marc Heintz

Abstract:

Emitting materials and all solid state lasers are widely used in the field of optical applications and materials science as a source of excitement, instrumental measurements, medical applications, metal shaping etc. Recently promising optical efficiencies were recorded on ceramics which result from a cheaper and faster ways to obtain crystallized materials. The choice and optimization of the sintering process is the key point to fabricate transparent ceramics. It includes a high control on the preparation of the powder with the choice of an adequate synthesis, a pre-heat-treatment, the reproducibility of the sintering cycle, the polishing and post-annealing of the ceramic. The densification is the main factor needed to reach a satisfying transparency, and many technologies are now available. The symmetry of the unit cell plays a crucial role in the diffusion rate of the material. Therefore, the cubic symmetry compounds having an isotropic refractive index is preferred. The cubic Y3NbO7 matrix is an interesting host which can accept a high concentration of rare earth doping element and it has been demonstrated that SPS is an efficient way to sinter this material. The optimization of diffusion losses requires a microstructure of fine ceramics, generally less than one hundred nanometers. In this case, grain growth is not an obstacle to transparency. The ceramics properties are then isotropic thereby to free-shaping step by orienting the ceramics as this is the case for the compounds of lower symmetry. After optimization of the synthesis route, several SPS parameters as heating rate, holding, dwell time and pressure were adjusted in order to increase the densification of the Eu3+ doped Y3NbO7 pellets. The luminescence data coupled with X-Ray diffraction analysis and electronic diffraction microscopy highlight the existence of several distorted environments of the doping element in the studied defective fluorite-type host lattice. Indeed, the fast and high crystallization rate obtained to put in evidence a lack of miscibility in the phase diagram, being the final composition of the pellet driven by the ratio between niobium and yttrium elements. By following the luminescence properties, we demonstrate a direct impact on the SPS process on this material.

Keywords: emission, niobate of rare earth, Spark plasma sintering, lack of miscibility

Procedia PDF Downloads 261
16499 Optimization of Pressure in Deep Drawing Process

Authors: Ajay Kumar Choubey, Geeta Agnihotri, C. Sasikumar, Rashmi Dwivedi

Abstract:

Deep-drawing operations are performed widely in industrial applications. It is very important for efficiency to achieve parts with no or minimum defects. Deep drawn parts are used in high performance, high strength and high reliability applications where tension, stress, load and human safety are critical considerations. Wrinkling is a kind of defect caused by stresses in the flange part of the blank during metal forming operations. To avoid wrinkling appropriate blank-holder pressure/force or drawbead can be applied. Now-a-day computer simulation plays a vital role in the field of manufacturing process. So computer simulation of manufacturing has much advantage over previous conventional process i.e. mass production, good quality of product, fast working etc. In this study, a two dimensional elasto-plastic Finite Element (F.E.) model for Mild Steel material blank has been developed to study the behavior of the flange wrinkling and deep drawing parameters under different Blank-Holder Pressure (B.H.P.). For this, commercially available Finite Element software ANSYS 14 has been used in this study. Simulation results are critically studied and salient conclusions have been drawn.

Keywords: ANSYS, deep drawing, BHP, finite element simulation, wrinkling

Procedia PDF Downloads 446
16498 Modelling of Moisture Loss and Oil Uptake during Deep-Fat Frying of Plantain

Authors: James A. Adeyanju, John O. Olajide, Akinbode A. Adedeji

Abstract:

A predictive mathematical model based on the fundamental principles of mass transfer was developed to simulate the moisture content and oil content during Deep-Fat Frying (DFF) process of dodo. The resulting governing equation, that is, partial differential equation that describes rate of moisture loss and oil uptake was solved numerically using explicit Finite Difference Technique (FDT). Computer codes were written in MATLAB environment for the implementation of FDT at different frying conditions and moisture loss as well as oil uptake simulation during DFF of dodo. Plantain samples were sliced into 5 mm thickness and fried at different frying oil temperatures (150, 160 and 170 ⁰C) for periods varying from 2 to 4 min. The comparison between the predicted results and experimental data for the validation of the model showed reasonable agreement. The correlation coefficients between the predicted and experimental values of moisture and oil transfer models ranging from 0.912 to 0.947 and 0.895 to 0.957, respectively. The predicted results could be further used for the design, control and optimization of deep-fat frying process.

Keywords: frying, moisture loss, modelling, oil uptake

Procedia PDF Downloads 438
16497 Multi-Objective Optimization of a Solar-Powered Triple-Effect Absorption Chiller for Air-Conditioning Applications

Authors: Ali Shirazi, Robert A. Taylor, Stephen D. White, Graham L. Morrison

Abstract:

In this paper, a detailed simulation model of a solar-powered triple-effect LiBr–H2O absorption chiller is developed to supply both cooling and heating demand of a large-scale building, aiming to reduce the fossil fuel consumption and greenhouse gas emissions in building sector. TRNSYS 17 is used to simulate the performance of the system over a typical year. A combined energetic-economic-environmental analysis is conducted to determine the system annual primary energy consumption and the total cost, which are considered as two conflicting objectives. A multi-objective optimization of the system is performed using a genetic algorithm to minimize these objectives simultaneously. The optimization results show that the final optimal design of the proposed plant has a solar fraction of 72% and leads to an annual primary energy saving of 0.69 GWh and annual CO2 emissions reduction of ~166 tonnes, as compared to a conventional HVAC system. The economics of this design, however, is not appealing without public funding, which is often the case for many renewable energy systems. The results show that a good funding policy is required in order for these technologies to achieve satisfactory payback periods within the lifetime of the plant.

Keywords: economic, environmental, multi-objective optimization, solar air-conditioning, triple-effect absorption chiller

Procedia PDF Downloads 234
16496 Optimization of MAG Welding Process Parameters Using Taguchi Design Method on Dead Mild Steel

Authors: Tadele Tesfaw, Ajit Pal Singh, Abebaw Mekonnen Gezahegn

Abstract:

Welding is a basic manufacturing process for making components or assemblies. Recent welding economics research has focused on developing the reliable machinery database to ensure optimum production. Research on welding of materials like steel is still critical and ongoing. Welding input parameters play a very significant role in determining the quality of a weld joint. The metal active gas (MAG) welding parameters are the most important factors affecting the quality, productivity and cost of welding in many industrial operations. The aim of this study is to investigate the optimization process parameters for metal active gas welding for 60x60x5mm dead mild steel plate work-piece using Taguchi method to formulate the statistical experimental design using semi-automatic welding machine. An experimental study was conducted at Bishoftu Automotive Industry, Bishoftu, Ethiopia. This study presents the influence of four welding parameters (control factors) like welding voltage (volt), welding current (ampere), wire speed (m/min.), and gas (CO2) flow rate (lit./min.) with three different levels for variability in the welding hardness. The objective functions have been chosen in relation to parameters of MAG welding i.e., welding hardness in final products. Nine experimental runs based on an L9 orthogonal array Taguchi method were performed. An orthogonal array, signal-to-noise (S/N) ratio and analysis of variance (ANOVA) are employed to investigate the welding characteristics of dead mild steel plate and used in order to obtain optimum levels for every input parameter at 95% confidence level. The optimal parameters setting was found is welding voltage at 22 volts, welding current at 125 ampere, wire speed at 2.15 m/min and gas flow rate at 19 l/min by using the Taguchi experimental design method within the constraints of the production process. Finally, six conformations welding have been carried out to compare the existing values; the predicated values with the experimental values confirm its effectiveness in the analysis of welding hardness (quality) in final products. It is found that welding current has a major influence on the quality of welded joints. Experimental result for optimum setting gave a better hardness of welding condition than initial setting. This study is valuable for different material and thickness variation of welding plate for Ethiopian industries.

Keywords: Weld quality, metal active gas welding, dead mild steel plate, orthogonal array, analysis of variance, Taguchi method

Procedia PDF Downloads 475
16495 A Resource Optimization Strategy for CPU (Central Processing Unit) Intensive Applications

Authors: Junjie Peng, Jinbao Chen, Shuai Kong, Danxu Liu

Abstract:

On the basis of traditional resource allocation strategies, the usage of resources on physical servers in cloud data center is great uncertain. It will cause waste of resources if the assignment of tasks is not enough. On the contrary, it will cause overload if the assignment of tasks is too much. This is especially obvious when the applications are the same type because of its resource preferences. Considering CPU intensive application is one of the most common types of application in the cloud, we studied the optimization strategy for CPU intensive applications on the same server. We used resource preferences to analyze the case that multiple CPU intensive applications run simultaneously, and put forward a model which can predict the execution time for CPU intensive applications which run simultaneously. Based on the prediction model, we proposed the method to select the appropriate number of applications for a machine. Experiments show that the model can predict the execution time accurately for CPU intensive applications. To improve the execution efficiency of applications, we propose a scheduling model based on priority for CPU intensive applications. Extensive experiments verify the validity of the scheduling model.

Keywords: cloud computing, CPU intensive applications, resource optimization, strategy

Procedia PDF Downloads 274
16494 Sensitivity Analysis of Prestressed Post-Tensioned I-Girder and Deck System

Authors: Tahsin A. H. Nishat, Raquib Ahsan

Abstract:

Sensitivity analysis of design parameters of the optimization procedure can become a significant factor while designing any structural system. The objectives of the study are to analyze the sensitivity of deck slab thickness parameter obtained from both the conventional and optimum design methodology of pre-stressed post-tensioned I-girder and deck system and to compare the relative significance of slab thickness. For analysis on conventional method, the values of 14 design parameters obtained by the conventional iterative method of design of a real-life I-girder bridge project have been considered. On the other side for analysis on optimization method, cost optimization of this system has been done using global optimization methodology 'Evolutionary Operation (EVOP)'. The problem, by which optimum values of 14 design parameters have been obtained, contains 14 explicit constraints and 46 implicit constraints. For both types of design parameters, sensitivity analysis has been conducted on deck slab thickness parameter which can become too sensitive for the obtained optimum solution. Deviations of slab thickness on both the upper and lower side of its optimum value have been considered reflecting its realistic possible ranges of variations during construction. In this procedure, the remaining parameters have been kept unchanged. For small deviations from the optimum value, compliance with the explicit and implicit constraints has been examined. Variations in the cost have also been estimated. It is obtained that without violating any constraint deck slab thickness obtained by the conventional method can be increased up to 25 mm whereas slab thickness obtained by cost optimization can be increased only up to 0.3 mm. The obtained result suggests that slab thickness becomes less sensitive in case of conventional method of design. Therefore, for realistic design purpose sensitivity should be conducted for any of the design procedure of girder and deck system.

Keywords: sensitivity analysis, optimum design, evolutionary operations, PC I-girder, deck system

Procedia PDF Downloads 131
16493 Maintenance Objective-Based Asset Maintenance Maturity Model

Authors: James M. Wakiru, Liliane Pintelon, Peter Muchiri, Peter Chemweno

Abstract:

The fast-changing business and operational environment are forcing organizations to adopt asset performance management strategies, not only to reduce costs but also maintain operational and production policies while addressing demand. To attain optimal asset performance management, a framework that ensures a continuous and systematic approach to analyzing an organization’s current maturity level and expected improvement regarding asset maintenance processes, strategies, technologies, capabilities, and systems is essential. Moreover, this framework while addressing maintenance-intensive organizations should consider the diverse business, operational and technical context (often dynamic) an organization is in and realistically prescribe or relate to the appropriate tools and systems the organization can potentially employ in the respective level, to improve and attain their maturity goals. This paper proposes an asset maintenance maturity model to assess the current capabilities, strength and weaknesses of maintenance processes an organization is using and analyze gaps for improvement via structuring set levels of achievement. At the epicentre of the proposed framework is the utilization of maintenance objective selected by an organization for various maintenance optimization programs. The framework adapts the Capability Maturity Model of assessing the maintenance process maturity levels in the organization.

Keywords: asset maintenance, maturity models, maintenance objectives, optimization

Procedia PDF Downloads 219
16492 The Bernstein Expansion for Exponentials in Taylor Functions: Approximation of Fixed Points

Authors: Tareq Hamadneh, Jochen Merker, Hassan Al-Zoubi

Abstract:

Bernstein's expansion for exponentials in Taylor functions provides lower and upper optimization values for the range of its original function. these values converge to the original functions if the degree is elevated or the domain subdivided. Taylor polynomial can be applied so that the exponential is a polynomial of finite degree over a given domain. Bernstein's basis has two main properties: its sum equals 1, and positive for all x 2 (0; 1). In this work, we prove the existence of fixed points for exponential functions in a given domain using the optimization values of Bernstein. The Bernstein basis of finite degree T over a domain D is defined non-negatively. Any polynomial p of degree t can be expanded into the Bernstein form of maximum degree t ≤ T, where we only need to compute the coefficients of Bernstein in order to optimize the original polynomial. The main property is that p(x) is approximated by the minimum and maximum Bernstein coefficients (Bernstein bound). If the bound is contained in the given domain, then we say that p(x) has fixed points in the same domain.

Keywords: Bernstein polynomials, Stability of control functions, numerical optimization, Taylor function

Procedia PDF Downloads 129
16491 Application of Failure Mode and Effects Analysis (FMEA) on the Virtual Process Hazard Analysis of Acetone Production Process

Authors: Princes Ann E. Prieto, Denise F. Alpuerto, John Rafael C. Unlayao, Neil Concibido, Monet Concepcion Maguyon-Detras

Abstract:

Failure Mode and Effects Analysis (FMEA) has been used in the virtual Process Hazard Analysis (PHA) of the Acetone production process through the dehydrogenation of isopropyl alcohol, for which very limited process risk assessment has been published. In this study, the potential failure modes, effects, and possible causes of selected major equipment in the process were identified. During the virtual FMEA mock sessions, the risks in the process were evaluated and recommendations to reduce and/or mitigate the process risks were formulated. The risk was estimated using the calculated risk priority number (RPN) and was classified into four (4) levels according to their effects on acetone production. Results of this study were also used to rank the criticality of equipment in the process based on the calculated criticality rating (CR). Bow tie diagrams were also created for the critical hazard scenarios identified in the study.

Keywords: chemical process safety, failure mode and effects analysis (FMEA), process hazard analysis (PHA), process safety management (PSM)

Procedia PDF Downloads 128
16490 Optimization of Floor Heating System in the Incompressible Turbulent Flow Using Constructal Theory

Authors: Karim Farahmandfar, Hamidolah Izadi, Mohammadreza Rezaei, Amin Ardali, Ebrahim Goshtasbi Rad, Khosro Jafarpoor

Abstract:

Statistics illustrates that the higher amount of annual energy consumption is related to surmounting the demand in buildings. Therefore, it is vital to economize the energy consumption and also find the solution with regard to this issue. One of the systems for the sake of heating the building is floor heating. As a matter of fact, floor heating performance is based on convection and radiation. Actually, in addition to creating a favorable heating condition, this method leads to energy saving. It is the goal of this article to outline the constructal theory and introduce the optimization method in branch networks for floor heating. There are several steps in order to gain this purpose. First of all, the pressure drop through the two points of the network is calculated. This pressure drop is as a function of pipes diameter and other parameters. After that, the amount of heat transfer is determined. Consequently, as a result of the combination of these two functions, the final function will be determined. It is necessary to mention that flow is laminar.

Keywords: constructal theory, optimization, floor heating system, turbulent flow

Procedia PDF Downloads 309
16489 A Holistic Workflow Modeling Method for Business Process Redesign

Authors: Heejung Lee

Abstract:

In a highly competitive environment, it becomes more important to shorten the whole business process while delivering or even enhancing the business value to the customers and suppliers. Although the workflow management systems receive much attention for its capacity to practically support the business process enactment, the effective workflow modeling method remain still challenging and the high degree of process complexity makes it more difficult to gain the short lead time. This paper presents a workflow structuring method in a holistic way that can reduce the process complexity using activity-needs and formal concept analysis, which eventually enhances the key performance such as quality, delivery, and cost in business process.

Keywords: workflow management, re-engineering, formal concept analysis, business process

Procedia PDF Downloads 405
16488 A Fast Optimizer for Large-scale Fulfillment Planning based on Genetic Algorithm

Authors: Choonoh Lee, Seyeon Park, Dongyun Kang, Jaehyeong Choi, Soojee Kim, Younggeun Kim

Abstract:

Market Kurly is the first South Korean online grocery retailer that guarantees same-day, overnight shipping. More than 1.6 million customers place an average of 4.7 million orders and add 3 to 14 products into a cart per month. The company has sold almost 30,000 kinds of various products in the past 6 months, including food items, cosmetics, kitchenware, toys for kids/pets, and even flowers. The company is operating and expanding multiple dry, cold, and frozen fulfillment centers in order to store and ship these products. Due to the scale and complexity of the fulfillment, pick-pack-ship processes are planned and operated in batches, and thus, the planning that decides the batch of the customers’ orders is a critical factor in overall productivity. This paper introduces a metaheuristic optimization method that reduces the complexity of batch processing in a fulfillment center. The method is an iterative genetic algorithm with heuristic creation and evolution strategies; it aims to group similar orders into pick-pack-ship batches to minimize the total number of distinct products. With a well-designed approach to create initial genes, the method produces streamlined plans, up to 13.5% less complex than the actual plans carried out in the company’s fulfillment centers in the previous months. Furthermore, our digital-twin simulations show that the optimized plans can reduce 3% of operation time for packing, which is the most complex and time-consuming task in the process. The optimization method implements a multithreading design on the Spring framework to support the company’s warehouse management systems in near real-time, finding a solution for 4,000 orders within 5 to 7 seconds on an AWS c5.2xlarge instance.

Keywords: fulfillment planning, genetic algorithm, online grocery retail, optimization

Procedia PDF Downloads 75
16487 Statistical Optimization of Vanillin Production by Pycnoporus Cinnabarinus 1181

Authors: Swarali Hingse, Shraddha Digole, Uday Annapure

Abstract:

The present study investigates the biotransformation of ferulic acid to vanillin by Pycnoporus cinnabarinus and its optimization using one-factor-at-a-time method as well as statistical approach. Effect of various physicochemical parameters and medium components was studied using one-factor-at-a-time method. Screening of the significant factors was carried out using L25 Taguchi orthogonal array and then these selected significant factors were further optimized using response surface methodology (RSM). Significant media components obtained using Taguchi L25 orthogonal array were glucose, KH2PO4 and yeast extract. Further, a Box Behnken design was used to investigate the interactive effects of the three most significant media components. The final medium obtained after optimization using RSM containing glucose (34.89 g/L), diammonium tartrate (1 g/L), yeast extract (1.47 g/L), MgSO4•7H2O (0.5 g/L), KH2PO4 (0.15 g/L), and CaCl2•2H2O (20 mg/L) resulted in amplification of vanillin production from 30.88 mg/L to 187.63 mg/L.

Keywords: ferulic acid, pycnoporus cinnabarinus, response surface methodology, vanillin

Procedia PDF Downloads 374