Search results for: production process optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 22537

Search results for: production process optimization

22117 Cloud Monitoring and Performance Optimization Ensuring High Availability and Security

Authors: Inayat Ur Rehman, Georgia Sakellari

Abstract:

Cloud computing has evolved into a vital technology for businesses, offering scalability, flexibility, and cost-effectiveness. However, maintaining high availability and optimal performance in the cloud is crucial for reliable services. This paper explores the significance of cloud monitoring and performance optimization in sustaining the high availability of cloud-based systems. It discusses diverse monitoring tools, techniques, and best practices for continually assessing the health and performance of cloud resources. The paper also delves into performance optimization strategies, including resource allocation, load balancing, and auto-scaling, to ensure efficient resource utilization and responsiveness. Addressing potential challenges in cloud monitoring and optimization, the paper offers insights into data security and privacy considerations. Through this thorough analysis, the paper aims to underscore the importance of cloud monitoring and performance optimization for ensuring a seamless and highly available cloud computing environment.

Keywords: cloud computing, cloud monitoring, performance optimization, high availability

Procedia PDF Downloads 65
22116 Methodology for Temporary Analysis of Production and Logistic Systems on the Basis of Distance Data

Authors: M. Mueller, M. Kuehn, M. Voelker

Abstract:

In small and medium-sized enterprises (SMEs), the challenge is to create a well-grounded and reliable basis for process analysis, optimization and planning due to a lack of data. SMEs have limited access to methods with which they can effectively and efficiently analyse processes and identify cause-and-effect relationships in order to generate the necessary database and derive optimization potential from it. The implementation of digitalization within the framework of Industry 4.0 thus becomes a particular necessity for SMEs. For these reasons, the abstract presents an analysis methodology that is subject to the objective of developing an SME-appropriate methodology for efficient, temporarily feasible data collection and evaluation in flexible production and logistics systems as a basis for process analysis and optimization. The overall methodology focuses on retrospective, event-based tracing and analysis of material flow objects. The technological basis consists of Bluetooth low energy (BLE)-based transmitters, so-called beacons, and smart mobile devices (SMD), e.g. smartphones as receivers, between which distance data can be measured and derived motion profiles. The distance is determined using the Received Signal Strength Indicator (RSSI), which is a measure of signal field strength between transmitter and receiver. The focus is the development of a software-based methodology for interpretation of relative movements of transmitters and receivers based on distance data. The main research is on selection and implementation of pattern recognition methods for automatic process recognition as well as methods for the visualization of relative distance data. Due to an existing categorization of the database regarding process types, classification methods (e.g. Support Vector Machine) from the field of supervised learning are used. The necessary data quality requires selection of suitable methods as well as filters for smoothing occurring signal variations of the RSSI, the integration of methods for determination of correction factors depending on possible signal interference sources (columns, pallets) as well as the configuration of the used technology. The parameter settings on which respective algorithms are based have a further significant influence on result quality of the classification methods, correction models and methods for visualizing the position profiles used. The accuracy of classification algorithms can be improved up to 30% by selected parameter variation; this has already been proven in studies. Similar potentials can be observed with parameter variation of methods and filters for signal smoothing. Thus, there is increased interest in obtaining detailed results on the influence of parameter and factor combinations on data quality in this area. The overall methodology is realized with a modular software architecture consisting of independently modules for data acquisition, data preparation and data storage. The demonstrator for initialization and data acquisition is available as mobile Java-based application. The data preparation, including methods for signal smoothing, are Python-based with the possibility to vary parameter settings and to store them in the database (SQLite). The evaluation is divided into two separate software modules with database connection: the achievement of an automated assignment of defined process classes to distance data using selected classification algorithms and the visualization as well as reporting in terms of a graphical user interface (GUI).

Keywords: event-based tracing, machine learning, process classification, parameter settings, RSSI, signal smoothing

Procedia PDF Downloads 131
22115 Analysis of Diabetes Patients Using Pearson, Cost Optimization, Control Chart Methods

Authors: Devatha Kalyan Kumar, R. Poovarasan

Abstract:

In this paper, we have taken certain important factors and health parameters of diabetes patients especially among children by birth (pediatric congenital) where using the above three metrics methods we are going to assess the importance of each attributes in the dataset and thereby determining the most highly responsible and co-related attribute causing diabetics among young patients. We use cost optimization, control chart and Spearmen methodologies for the real-time application of finding the data efficiency in this diabetes dataset. The Spearmen methodology is the correlation methodologies used in software development process to identify the complexity between the various modules of the software. Identifying the complexity is important because if the complexity is higher, then there is a higher chance of occurrence of the risk in the software. With the use of control; chart mean, variance and standard deviation of data are calculated. With the use of Cost optimization model, we find to optimize the variables. Hence we choose the Spearmen, control chart and cost optimization methods to assess the data efficiency in diabetes datasets.

Keywords: correlation, congenital diabetics, linear relationship, monotonic function, ranking samples, pediatric

Procedia PDF Downloads 256
22114 Utilization of Agro-Industrial Byproducts for Bacteriocin Production Using Newly Isolated Enterococcus faecium BS13

Authors: Vandana Bali, Manab B. Bera, Parmjit S. Panesar

Abstract:

Microbial production of antimicrobials as biopreservatives is the major area of focus nowadays due to increased interest of consumers towards natural and safe preservation of ready to eat food products. The agro-industrial byproduct based medium and optimized process conditions can contribute in economical production of bacteriocins. Keeping this in view, the present investigation was carried out on agro-industrial byproducts utilization for the production of bacteriocin using Enterococcus faecium BS13 isolated from local fermented food. Different agro-industrial byproduct based carbon sources (whey, potato starch liquor, kinnow peel, deoiledrice bran and molasses), nitrogen sources (soya okra, pea pod and corn steep liquor), metal ions and surfactants were tested for optimal bacteriocin production. The effect of various process parameters such as pH, temperature, inoculum level, agitation and time were also tested on bacteriocin production. The optimized medium containing whey, supplemented with 4%corn steep liquor and polysorbate-80 displayed maximum bacteriocin activity with 2% inoculum, at pH 6.5, temperature 40oC under shaking conditions (100 rpm).

Keywords: Bacteriocin, biopreservation, corn steep liquor, Enterococcus faecium, waste utilization, whey

Procedia PDF Downloads 238
22113 Energy Analysis of Sugarcane Production: A Case Study in Metehara Sugar Factory in Ethiopia

Authors: Wasihun Girma Hailemariam

Abstract:

Energy is one of the key elements required for every agricultural activity, especially for large scale agricultural production such as sugarcane cultivation which mostly is used to produce sugar and bioethanol from sugarcane. In such kinds of resource (energy) intensive activities, energy analysis of the production system and looking for other alternatives which can reduce energy inputs of the sugarcane production process are steps forward for resource management. The purpose of this study was to determine input energy (direct and indirect) per hectare of sugarcane production sector of Metehara sugar factory in Ethiopia. Total energy consumption of the production system was 61,642 MJ/ha-yr. This total input energy is a cumulative value of different inputs (direct and indirect inputs) in the production system. The contribution of these different inputs is discussed and a scenario of substituting the most influential input by other alternative input which can replace the original input in its nutrient content was discussed. In this study the most influential input for increased energy consumption was application of organic fertilizer which accounted for 50 % of the total energy consumption. Filter cake which is a residue from the sugar production in the factory was used to substitute the organic fertilizer and the reduction in the energy consumption of the sugarcane production was discussed

Keywords: energy analysis, organic fertilizer, resource management, sugarcane

Procedia PDF Downloads 158
22112 Design Optimisation of a Novel Cross Vane Expander-Compressor Unit for Refrigeration System

Authors: Y. D. Lim, K. S. Yap, K. T. Ooi

Abstract:

In recent years, environmental issue has been a hot topic in the world, especially the global warming effect caused by conventional non-environmentally friendly refrigerants has increased. Several studies of a more energy-efficient and environmentally friendly refrigeration system have been conducted in order to tackle the issue. In search of a better refrigeration system, CO2 refrigeration system has been proposed as a better option. However, the high throttling loss involved during the expansion process of the refrigeration cycle leads to a relatively low efficiency and thus the system is impractical. In order to improve the efficiency of the refrigeration system, it is suggested by replacing the conventional expansion valve in the refrigeration system with an expander. Based on this issue, a new type of expander-compressor combined unit, named Cross Vane Expander-Compressor (CVEC) was introduced to replace the compressor and the expansion valve of a conventional refrigeration system. A mathematical model was developed to calculate the performance of CVEC, and it was found that the machine is capable of saving the energy consumption of a refrigeration system by as much as 18%. Apart from energy saving, CVEC is also geometrically simpler and more compact. To further improve its efficiency, optimization study of the device is carried out. In this report, several design parameters of CVEC were chosen to be the variables of optimization study. This optimization study was done in a simulation program by using complex optimization method, which is a direct search, multi-variables and constrained optimization method. It was found that the main design parameters, which was shaft radius was reduced around 8% while the inner cylinder radius was remained unchanged at its lower limit after optimization. Furthermore, the port sizes were increased to their upper limit after optimization. The changes of these design parameters have resulted in reduction of around 12% in the total frictional loss and reduction of 4% in power consumption. Eventually, the optimization study has resulted in an improvement in the mechanical efficiency CVEC by 4% and improvement in COP by 6%.

Keywords: complex optimization method, COP, cross vane expander-compressor, CVEC, design optimization, direct search, energy saving, improvement, mechanical efficiency, multi variables

Procedia PDF Downloads 373
22111 Multi-Factor Optimization Method through Machine Learning in Building Envelope Design: Focusing on Perforated Metal Façade

Authors: Jinwooung Kim, Jae-Hwan Jung, Seong-Jun Kim, Sung-Ah Kim

Abstract:

Because the building envelope has a significant impact on the operation and maintenance stage of the building, designing the facade considering the performance can improve the performance of the building and lower the maintenance cost of the building. In general, however, optimizing two or more performance factors confronts the limits of time and computational tools. The optimization phase typically repeats infinitely until a series of processes that generate alternatives and analyze the generated alternatives achieve the desired performance. In particular, as complex geometry or precision increases, computational resources and time are prohibitive to find the required performance, so an optimization methodology is needed to deal with this. Instead of directly analyzing all the alternatives in the optimization process, applying experimental techniques (heuristic method) learned through experimentation and experience can reduce resource waste. This study proposes and verifies a method to optimize the double envelope of a building composed of a perforated panel using machine learning to the design geometry and quantitative performance. The proposed method is to achieve the required performance with fewer resources by supplementing the existing method which cannot calculate the complex shape of the perforated panel.

Keywords: building envelope, machine learning, perforated metal, multi-factor optimization, façade

Procedia PDF Downloads 224
22110 A Life Cycle Assessment (LCA) of Aluminum Production Process

Authors: Alaa Al Hawari, Mohammad Khader, Wael El Hasan, Mahmoud Alijla, Ammar Manawi, Abdelbaki Benamour

Abstract:

The production of aluminium alloys and ingots -starting from the processing of alumina to aluminium, and the final cast product- was studied using a Life Cycle Assessment (LCA) approach. The studied aluminium supply chain consisted of a carbon plant, a reduction plant, a casting plant, and a power plant. In the LCA model, the environmental loads of the different plants for the production of 1 ton of aluminium metal were investigated. The impact of the aluminium production was assessed in eight impact categories. The results showed that for all of the impact categories the power plant had the highest impact only in the cases of Human Toxicity Potential (HTP) the reduction plant had the highest impact and in the Marine Aquatic Eco-Toxicity Potential (MAETP) the carbon plant had the highest impact. Furthermore, the impact of the carbon plant and the reduction plant combined was almost the same as the impact of the power plant in the case of the Acidification Potential (AP). The carbon plant had a positive impact on the environment when it comes to the Eutrophication Potential (EP) due to the production of clean water in the process. The natural gas based power plant used in the case study had 8.4 times less negative impact on the environment when compared to the heavy fuel based power plant and 10.7 times less negative impact when compared to the hard coal based power plant.

Keywords: life cycle assessment, aluminium production, supply chain, ecological impacts

Procedia PDF Downloads 532
22109 Easy Way of Optimal Process-Storage Network Design

Authors: Gyeongbeom Yi

Abstract:

The purpose of this study is to introduce the analytic solution for determining the optimal capacity (lot-size) of a multiproduct, multistage production and inventory system to meet the finished product demand. Reasonable decision-making about the capacity of processes and storage units is an important subject for industry. The industrial solution for this subject is to use the classical economic lot sizing method, EOQ/EPQ (Economic Order Quantity/Economic Production Quantity) model, incorporated with practical experience. However, the unrealistic material flow assumption of the EOQ/EPQ model is not suitable for chemical plant design with highly interlinked processes and storage units. This study overcomes the limitation of the classical lot sizing method developed on the basis of the single product and single stage assumption. The superstructure of the plant considered consists of a network of serially and/or parallelly interlinked processes and storage units. The processes involve chemical reactions with multiple feedstock materials and multiple products as well as mixing, splitting or transportation of materials. The objective function for optimization is minimizing the total cost composed of setup and inventory holding costs as well as the capital costs of constructing processes and storage units. A novel production and inventory analysis method, PSW (Periodic Square Wave) model, is applied. The advantage of the PSW model comes from the fact that the model provides a set of simple analytic solutions in spite of a realistic description of the material flow between processes and storage units. The resulting simple analytic solution can greatly enhance the proper and quick investment decision for plant design and operation problem confronted in diverse economic situations.

Keywords: analytic solution, optimal design, process-storage network

Procedia PDF Downloads 331
22108 An Integration of Life Cycle Assessment and Techno-Economic Optimization in the Supply Chains

Authors: Yohanes Kristianto

Abstract:

The objective of this paper is to compose a sustainable supply chain that integrates product, process and networks design. An integrated life cycle assessment and techno-economic optimization is proposed that might deliver more economically feasible operations, minimizes environmental impacts and maximizes social contributions. Closed loop economy of the supply chain is achieved by reusing waste to be raw material of final products. Societal benefit is given by the supply chain by absorbing waste as source of raw material and opening new work opportunities. A case study of ethanol supply chain from rice straws is considered. The modeling results show that optimization within the scope of LCA is capable of minimizing both CO₂ emissions and energy and utility consumptions and thus enhancing raw materials utilization. Furthermore, the supply chain is capable of contributing to local economy through jobs creation. While the model is quite comprehensive, the future research recommendation on energy integration and global sustainability is proposed.

Keywords: life cycle assessment, techno-economic optimization, sustainable supply chains, closed loop economy

Procedia PDF Downloads 150
22107 The Optimization of the Parameters for Eco-Friendly Leaching of Precious Metals from Waste Catalyst

Authors: Silindile Gumede, Amir Hossein Mohammadi, Mbuyu Germain Ntunka

Abstract:

Goal 12 of the 17 Sustainable Development Goals (SDGs) encourages sustainable consumption and production patterns. This necessitates achieving the environmentally safe management of chemicals and all wastes throughout their life cycle and the proper disposal of pollutants and toxic waste. Fluid catalytic cracking (FCC) catalysts are widely used in the refinery to convert heavy feedstocks to lighter ones. During the refining processes, the catalysts are deactivated and discarded as hazardous toxic solid waste. Spent catalysts (SC) contain high-cost metal, and the recovery of metals from SCs is a tactical plan for supplying part of the demand for these substances and minimizing the environmental impacts. Leaching followed by solvent extraction, has been found to be the most efficient method to recover valuable metals with high purity from spent catalysts. However, the use of inorganic acids during the leaching process causes a secondary environmental issue. Therefore, it is necessary to explore other alternative efficient leaching agents that are economical and environmentally friendly. In this study, the waste catalyst was collected from a domestic refinery and was characterised using XRD, ICP, XRF, and SEM. Response surface methodology (RSM) and Box Behnken design were used to model and optimize the influence of some parameters affecting the acidic leaching process. The parameters selected in this investigation were the acid concentration, temperature, and leaching time. From the characterisation results, it was found that the spent catalyst consists of high concentrations of Vanadium (V) and Nickel (Ni); hence this study focuses on the leaching of Ni and V using a biodegradable acid to eliminate the formation of the secondary pollution.

Keywords: eco-friendly leaching, optimization, metal recovery, leaching

Procedia PDF Downloads 68
22106 Optimization Studies on Biosorption of Ni(II) and Cd(II) from Wastewater Using Pseudomonas putida in a Packed Bed Bioreactor

Authors: K.Narasimhulu, Y. Pydi Setty

Abstract:

The objective of this present study is the optimization of process parameters in biosorption of Ni(II) and Cd(II) ions by Pseudomonas putida using Response Surface Methodology in a Packed bed bioreactor. The experimental data were also tested with theoretical models to find the best fit model. The present paper elucidates RSM as an efficient approach for predictive model building and optimization of Ni(II) and Cd(II) ions using Pseudomonas putida. In packed bed biosorption studies, comparison of the breakthrough curves of Ni(II) and Cd(II) for Agar immobilized and PAA immobilized Pseudomonas putida at optimum conditions of flow rate of 300 mL/h, initial metal ion concentration of 100 mg/L and bed height of 20 cm with weight of biosorbent of 12 g, it was found that the Agar immobilized Pseudomonas putida showed maximum percent biosorption and bed saturation occurred at 20 minutes. Optimization results of Ni(II) and Cd(II) by Pseudomonas putida from the Design Expert software were obtained as bed height of 19.93 cm, initial metal ion concentration of 103.85 mg/L, and flow rate of 310.57 mL/h. The percent biosorption of Ni(II) and Cd(II) is 87.2% and 88.2% respectively. The predicted optimized parameters are in agreement with the experimental results.

Keywords: packed bed bioreactor, response surface mthodology, pseudomonas putida, biosorption, waste water

Procedia PDF Downloads 452
22105 Modeling and Optimization of Micro-Grid Using Genetic Algorithm

Authors: Mehrdad Rezaei, Reza Haghmaram, Nima Amjadi

Abstract:

This paper proposes an operating and cost optimization model for micro-grid (MG). This model takes into account emission costs of NOx, SO2, and CO2, together with the operation and maintenance costs. Wind turbines (WT), photovoltaic (PV) arrays, micro turbines (MT), fuel cells (FC), diesel engine generators (DEG) with different capacities are considered in this model. The aim of the optimization is minimizing operation cost according to constraints, supply demand and safety of the system. The proposed genetic algorithm (GA), with the ability to fine-tune its own settings, is used to optimize the micro-grid operation.

Keywords: micro-grid, optimization, genetic algorithm, MG

Procedia PDF Downloads 512
22104 Optimizing Cell Culture Performance in an Ambr15 Microbioreactor Using Dynamic Flux Balance and Computational Fluid Dynamic Modelling

Authors: William Kelly, Sorelle Veigne, Xianhua Li, Zuyi Huang, Shyamsundar Subramanian, Eugene Schaefer

Abstract:

The ambr15™ bioreactor is a single-use microbioreactor for cell line development and process optimization. The ambr system offers fully automatic liquid handling with the possibility of fed-batch operation and automatic control of pH and oxygen delivery. With operating conditions for large scale biopharmaceutical production properly scaled down, micro bioreactors such as the ambr15™ can potentially be used to predict the effect of process changes such as modified media or different cell lines. In this study, gassing rates and dilution rates were varied for a semi-continuous cell culture system in the ambr15™ bioreactor. The corresponding changes to metabolite production and consumption, as well as cell growth rate and therapeutic protein production were measured. Conditions were identified in the ambr15™ bioreactor that produced metabolic shifts and specific metabolic and protein production rates also seen in the corresponding larger (5 liter) scale perfusion process. A Dynamic Flux Balance model was employed to understand and predict the metabolic changes observed. The DFB model-predicted trends observed experimentally, including lower specific glucose consumption when CO₂ was maintained at higher levels (i.e. 100 mm Hg) in the broth. A Computational Fluid Dynamic (CFD) model of the ambr15™ was also developed, to understand transfer of O₂ and CO₂ to the liquid. This CFD model predicted gas-liquid flow in the bioreactor using the ANSYS software. The two-phase flow equations were solved via an Eulerian method, with population balance equations tracking the size of the gas bubbles resulting from breakage and coalescence. Reasonable results were obtained in that the Carbon Dioxide mass transfer coefficient (kLa) and the air hold up increased with higher gas flow rate. Volume-averaged kLa values at 500 RPM increased as the gas flow rate was doubled and matched experimentally determined values. These results form a solid basis for optimizing the ambr15™, using both CFD and FBA modelling approaches together, for use in microscale simulations of larger scale cell culture processes.

Keywords: cell culture, computational fluid dynamics, dynamic flux balance analysis, microbioreactor

Procedia PDF Downloads 282
22103 Developing Digital Twins of Steel Hull Processes

Authors: V. Ložar, N. Hadžić, T. Opetuk, R. Keser

Abstract:

The development of digital twins strongly depends on efficient algorithms and their capability to mirror real-life processes. Nowadays, such efforts are required to establish factories of the future faced with new demands of custom-made production. The ship hull processes face these challenges too. Therefore, it is important to implement design and evaluation approaches based on production system engineering. In this study, the recently developed finite state method is employed to describe the stell hull process as a platform for the implementation of digital twinning technology. The application is justified by comparing the finite state method with the analytical approach. This method is employed to rebuild a model of a real shipyard ship hull process using a combination of serial and splitting lines. The key performance indicators such as the production rate, work in process, probability of starvation, and blockade are calculated and compared to the corresponding results obtained through a simulation approach using the software tool Enterprise dynamics. This study confirms that the finite state method is a suitable tool for digital twinning applications. The conclusion highlights the advantages and disadvantages of methods employed in this context.

Keywords: digital twin, finite state method, production system engineering, shipyard

Procedia PDF Downloads 99
22102 Resource Efficiency within Current Production

Authors: Sarah Majid Ansari, Serjosha Wulf, Matthias Goerke

Abstract:

In times of global warming and the increasing shortage of resources, sustainable production is becoming more and more inevitable. Companies cannot only heighten their competitiveness but also contribute positively to environmental protection through efficient energy and resource consumption. Regarding this, technical solutions are often preferred during production, although organizational and process-related approaches also offer great potential. This project focuses on reducing resource usage, with a special emphasis on the human factor. It is the aspiration to develop a methodology that systematically implements and embeds suitable and individual measures and methods regarding resource efficiency throughout the entire production. The measures and methods established help employees handle resources and energy more sensitively. With this in mind, this paper also deals with the difficulties that can occur during the sensitization of employees and the implementation of these measures and methods. In addition, recommendations are given on how to avoid such difficulties.

Keywords: implementation, human factors, production plants, resource efficiency

Procedia PDF Downloads 481
22101 Research of Concentratibility of Low Quality Bauxite Raw Materials

Authors: Nadezhda Nikolaeva, Tatyana Alexandrova, Alexandr Alexandrov

Abstract:

Processing of high-silicon bauxite on the base of the traditional clinkering method is related to high power consumption and capital investments, which makes production of alumina from those ores non-competitive in terms of basic economic showings. For these reasons, development of technological solutions enabling to process bauxites with various chemical and mineralogical structures efficiently with low level of thermal power consumption is important. Flow sheet of the studies on washability of ores from the Timanskoe and the Severo-Onezhskoe deposits is on the base of the flotation method.

Keywords: low-quality bauxite, resource-saving technology, optimization, aluminum, conditioning of composition, separation characteristics

Procedia PDF Downloads 290
22100 A Hybrid Particle Swarm Optimization-Nelder- Mead Algorithm (PSO-NM) for Nelson-Siegel- Svensson Calibration

Authors: Sofia Ayouche, Rachid Ellaia, Rajae Aboulaich

Abstract:

Today, insurers may use the yield curve as an indicator evaluation of the profit or the performance of their portfolios; therefore, they modeled it by one class of model that has the ability to fit and forecast the future term structure of interest rates. This class of model is the Nelson-Siegel-Svensson model. Unfortunately, many authors have reported a lot of difficulties when they want to calibrate the model because the optimization problem is not convex and has multiple local optima. In this context, we implement a hybrid Particle Swarm optimization and Nelder Mead algorithm in order to minimize by least squares method, the difference between the zero-coupon curve and the NSS curve.

Keywords: optimization, zero-coupon curve, Nelson-Siegel-Svensson, particle swarm optimization, Nelder-Mead algorithm

Procedia PDF Downloads 430
22099 A Weighted Sum Particle Swarm Approach (WPSO) Combined with a Novel Feasibility-Based Ranking Strategy for Constrained Multi-Objective Optimization of Compact Heat Exchangers

Authors: Milad Yousefi, Moslem Yousefi, Ricarpo Poley, Amer Nordin Darus

Abstract:

Design optimization of heat exchangers is a very complicated task that has been traditionally carried out based on a trial-and-error procedure. To overcome the difficulties of the conventional design approaches especially when a large number of variables, constraints and objectives are involved, a new method based on a well-stablished evolutionary algorithm, particle swarm optimization (PSO), weighted sum approach and a novel constraint handling strategy is presented in this study. Since, the conventional constraint handling strategies are not effective and easy-to-implement in multi-objective algorithms, a novel feasibility-based ranking strategy is introduced which is both extremely user-friendly and effective. A case study from industry has been investigated to illustrate the performance of the presented approach. The results show that the proposed algorithm can find the near pareto-optimal with higher accuracy when it is compared to conventional non-dominated sorting genetic algorithm II (NSGA-II). Moreover, the difficulties of a trial-and-error process for setting the penalty parameters is solved in this algorithm.

Keywords: Heat exchanger, Multi-objective optimization, Particle swarm optimization, NSGA-II Constraints handling.

Procedia PDF Downloads 555
22098 Environmental Performance of Olive Oil Production in Greece

Authors: P. Tsarouhas, Ch. Achillas, D. Aidonis, D. Folinas, V. Maslis, N. Moussiopoulos

Abstract:

Agricultural production is a sector with high socioeconomic significance and key implications on employment and nutritional security. However, the impacts of agrifood production and consumption patterns on the environment are considerable, mainly due to the demand of large inputs of resources. This paper presents a case study of olive oil production in Greece, an important agri-product especially for countries in the Mediterranean basin. Life Cycle Analysis has been used to quantify the environmental performance of olive oil production. All key parameters that are associated with the life cycle of olive oil production are studied and environmental “hotspots” are diagnosed.

Keywords: LCA, olive oil production, environmental impact, case study, Greece

Procedia PDF Downloads 433
22097 Reducing the Computational Overhead of Metaheuristics Parameterization with Exploratory Landscape Analysis

Authors: Iannick Gagnon, Alain April

Abstract:

The performance of a metaheuristic on a given problem class depends on the class itself and the choice of parameters. Parameter tuning is the most time-consuming phase of the optimization process after the main calculations and it often nullifies the speed advantage of metaheuristics over traditional optimization algorithms. Several off-the-shelf parameter tuning algorithms are available, but when the objective function is expensive to evaluate, these can be prohibitively expensive to use. This paper presents a surrogate-like method for finding adequate parameters using fitness landscape analysis on simple benchmark functions and real-world objective functions. The result is a simple compound similarity metric based on the empirical correlation coefficient and a measure of convexity. It is then used to find the best benchmark functions to serve as surrogates. The near-optimal parameter set is then found using fractional factorial design. The real-world problem of NACA airfoil lift coefficient maximization is used as a preliminary proof of concept. The overall aim of this research is to reduce the computational overhead of metaheuristics parameterization.

Keywords: metaheuristics, stochastic optimization, particle swarm optimization, exploratory landscape analysis

Procedia PDF Downloads 153
22096 Reinforcement Learning Optimization: Unraveling Trends and Advancements in Metaheuristic Algorithms

Authors: Rahul Paul, Kedar Nath Das

Abstract:

The field of machine learning (ML) is experiencing rapid development, resulting in a multitude of theoretical advancements and extensive practical implementations across various disciplines. The objective of ML is to facilitate the ability of machines to perform cognitive tasks by leveraging knowledge gained from prior experiences and effectively addressing complex problems, even in situations that deviate from previously encountered instances. Reinforcement Learning (RL) has emerged as a prominent subfield within ML and has gained considerable attention in recent times from researchers. This surge in interest can be attributed to the practical applications of RL, the increasing availability of data, and the rapid advancements in computing power. At the same time, optimization algorithms play a pivotal role in the field of ML and have attracted considerable interest from researchers. A multitude of proposals have been put forth to address optimization problems or improve optimization techniques within the domain of ML. The necessity of a thorough examination and implementation of optimization algorithms within the context of ML is of utmost importance in order to provide guidance for the advancement of research in both optimization and ML. This article provides a comprehensive overview of the application of metaheuristic evolutionary optimization algorithms in conjunction with RL to address a diverse range of scientific challenges. Furthermore, this article delves into the various challenges and unresolved issues pertaining to the optimization of RL models.

Keywords: machine learning, reinforcement learning, loss function, evolutionary optimization techniques

Procedia PDF Downloads 75
22095 An Approach of High Scalable Production Capacity by Adaption of the Concept 'Everything as a Service'

Authors: Johannes Atug, Stefan Braunreuther, Gunther Reinhart

Abstract:

Volatile markets, as well as increasing global competition in manufacturing, lead to a high demand of flexible and agile production systems. These advanced production systems in turn conduct to high capital expenditure along with high investment risks. Developments in production regarding digitalization and cyber-physical systems result to a merger of informational- and operational technology. The approach of this paper is to benefit from this merger and present a framework of a production network with scalable production capacity and low capital expenditure by adaptation of the IT concept 'everything as a service' into the production environment.

Keywords: digital manufacturing system, everything as a service, reconfigurable production, value network

Procedia PDF Downloads 343
22094 Software Architecture Optimization Using Swarm Intelligence Techniques

Authors: Arslan Ellahi, Syed Amjad Hussain, Fawaz Saleem Bokhari

Abstract:

Optimization of software architecture can be done with respect to a quality attributes (QA). In this paper, there is an analysis of multiple research papers from different dimensions that have been used to classify those attributes. We have proposed a technique of swarm intelligence Meta heuristic ant colony optimization algorithm as a contribution to solve this critical optimization problem of software architecture. We have ranked quality attributes and run our algorithm on every QA, and then we will rank those on the basis of accuracy. At the end, we have selected the most accurate quality attributes. Ant colony algorithm is an effective algorithm and will perform best in optimizing the QA’s and ranking them.

Keywords: complexity, rapid evolution, swarm intelligence, dimensions

Procedia PDF Downloads 261
22093 High Pressure Delignification Process for Nanocrystalline Cellulose Production from Agro-Waste Biomass

Authors: Sakinul Islam, Nhol Kao, Sati Bhattacharya, Rahul Gupta

Abstract:

Nanocrystalline cellulose (NCC) has been widely used for miscellaneous applications due to its superior properties over other nanomaterials. However, the major problems associated with the production of NCC are long reaction time, low production rate and inefficient process. The mass production of NCC within a short period of time is still a great challenge. The main objective of this study is to produce NCC from rice husk agro waste biomass from a high pressure delignification process (HPDP), followed by bleaching and hydrolysis processes. The HPDP has not been explored for NCC production from rice husk biomass (RHB) until now. In order to produce NCC, powder rice husk (PRH) was placed into a stainless steel reactor at 80 ˚C under 5 bars. Aqueous solution of NaOH (4M) was used for the dissolution of lignin and other amorphous impurities from PRH. After certain experimental times (1h, 3.5h and 6h), bleaching and hydrolysis were carried out on delignified samples. NaOCl (20%) and H2SO4 (4M) solutions were used for bleaching and hydrolysis processes, respectively. The NCC suspension from hydrolysis was sonicated and neutralized by buffer solution for various characterisations. Finally NCC suspension was dried and analyzed by FTIR, XRD, SEM, AFM and TEM. The chemical composition of NCC and PRH was estimated by TAPPI (Technical Association of Pulp and Paper Industry) standard methods to observe the product purity. It was found that, the 6h of the HPDP was more efficient to produce good quality NCC than that at 1h and 3.5h due to low separation of non-cellulosic components from RHB. The analyses indicated the crystallinity of NCC to be 71 %, particle size of 20-50 nm (diameter) and 100-200 nm in length.

Keywords: nanocrystalline cellulose, NCC, high pressure delignification, bleaching, hydrolysis, agro-waste biomass

Procedia PDF Downloads 264
22092 The Study on Energy Saving in Clarification Process for Water Treatment Plant

Authors: Wiwat Onnakklum

Abstract:

Clarification is the turbidity removal process of water treatment plant. This paper was to study the factors affecting on energy consumption in order to control energy saving strategy. The factors studied were raw water turbidity in the range of 26-40 NTU and production rate in the range of 3.76-5.20 m³/sec. Clarifiers were sludge blanket and sludge recirculation clarifier. Experimental results found that the raw water turbidity was not affected significantly by energy consumption, while the production rate was affected significantly by energy consumption. Sludge blanket clarifier provided lower energy consumption than sludge recirculation clarifier about 32-37%. Subsequently, the operating pattern in production rate can be arranged to decreased energy consumption. The results showed that it can be reduced about 5.09 % of energy saving of clarification process about 754,655 Baht per year.

Keywords: sludge blanket clarifier, sludge recirculation clarifier, water treatment plant, energy

Procedia PDF Downloads 326
22091 Application Quality Function Deployment (QFD) Tool in Design of Aero Pumps Based on System Engineering

Authors: Z. Soleymani, M. Amirzadeh

Abstract:

Quality Function Deployment (QFD) was developed in 1960 in Japan and introduced in 1983 in America and Europe. The paper presents a real application of this technique in a way that the method of applying QFD in design and production aero fuel pumps has been considered. While designing a product and in order to apply system engineering process, the first step is identification customer needs then its transition to engineering parameters. Since each change in deign after production process leads to extra human costs and also increase in products quality risk, QFD can make benefits in sale by meeting customer expectations. Since the needs identified as well, the use of QFD tool can lead to increase in communications and less deviation in design and production phases, finally it leads to produce the products with defined technical attributes.

Keywords: customer voice, engineering parameters, gear pump, QFD

Procedia PDF Downloads 249
22090 Use of Transportation Networks to Optimize The Profit Dynamics of the Product Distribution

Authors: S. Jayasinghe, R. B. N. Dissanayake

Abstract:

Optimization modelling together with the Network models and Linear Programming techniques is a powerful tool in problem solving and decision making in real world applications. This study developed a mathematical model to optimize the net profit by minimizing the transportation cost. This model focuses the transportation among decentralized production plants to a centralized distribution centre and then the distribution among island wide agencies considering the customer satisfaction as a requirement. This company produces basically 9 types of food items with 82 different varieties and 4 types of non-food items with 34 different varieties. Among 6 production plants, 4 were located near the city of Mawanella and the other 2 were located in Galewala and Anuradhapura cities which are 80 km and 150 km away from Mawanella respectively. The warehouse located in the Mawanella was the main production plant and also the only distribution plant. This plant distributes manufactured products to 39 agencies island-wide. The average values and average amount of the goods for 6 consecutive months from May 2013 to October 2013 were collected and then average demand values were calculated. The following constraints are used as the necessary requirement to satisfy the optimum condition of the model; there was one source, 39 destinations and supply and demand for all the agencies are equal. Using transport cost for a kilometer, total transport cost was calculated. Then the model was formulated using distance and flow of the distribution. Network optimization and linear programming techniques were used to originate the model while excel solver is used in solving. Results showed that company requires total transport cost of Rs. 146, 943, 034.50 to fulfil the customers’ requirement for a month. This is very much less when compared with data without using the model. Model also proved that company can reduce their transportation cost by 6% when distributing to island-wide customers. Company generally satisfies their customers’ requirements by 85%. This satisfaction can be increased up to 97% by using this model. Therefore this model can be used by other similar companies in order to reduce the transportation cost.

Keywords: mathematical model, network optimization, linear programming

Procedia PDF Downloads 346
22089 Chemical Reaction Algorithm for Expectation Maximization Clustering

Authors: Li Ni, Pen ManMan, Li KenLi

Abstract:

Clustering is an intensive research for some years because of its multifaceted applications, such as biology, information retrieval, medicine, business and so on. The expectation maximization (EM) is a kind of algorithm framework in clustering methods, one of the ten algorithms of machine learning. Traditionally, optimization of objective function has been the standard approach in EM. Hence, research has investigated the utility of evolutionary computing and related techniques in the regard. Chemical Reaction Optimization (CRO) is a recently established method. So the property embedded in CRO is used to solve optimization problems. This paper presents an algorithm framework (EM-CRO) with modified CRO operators based on EM cluster problems. The hybrid algorithm is mainly to solve the problem of initial value sensitivity of the objective function optimization clustering algorithm. Our experiments mainly take the EM classic algorithm:k-means and fuzzy k-means as an example, through the CRO algorithm to optimize its initial value, get K-means-CRO and FKM-CRO algorithm. The experimental results of them show that there is improved efficiency for solving objective function optimization clustering problems.

Keywords: chemical reaction optimization, expection maimization, initia, objective function clustering

Procedia PDF Downloads 715
22088 Analysis of Tandem Detonator Algorithm Optimized by Quantum Algorithm

Authors: Tomasz Robert Kuczerski

Abstract:

The high complexity of the algorithm of the autonomous tandem detonator system creates an optimization problem due to the parallel operation of several machine states of the system. Many years of experience and classic analyses have led to a partially optimized model. Limitations on the energy resources of this class of autonomous systems make it necessary to search for more effective methods of optimisation. The use of the Quantum Approximate Optimization Algorithm (QAOA) in these studies shows the most promising results. With the help of multiple evaluations of several qubit quantum circuits, proper results of variable parameter optimization were obtained. In addition, it was observed that the increase in the number of assessments does not result in further efficient growth due to the increasing complexity of optimising variables. The tests confirmed the effectiveness of the QAOA optimization method.

Keywords: algorithm analysis, autonomous system, quantum optimization, tandem detonator

Procedia PDF Downloads 92