Search results for: optimization algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4760

Search results for: optimization algorithms

2990 Investigating Effects of Vehicle Speed and Road PSDs on Response of a 35-Ton Heavy Commercial Vehicle (HCV) Using Mathematical Modelling

Authors: Amal G. Kurian

Abstract:

The use of mathematical modeling has seen a considerable boost in recent times with the development of many advanced algorithms and mathematical modeling capabilities. The advantages this method has over other methods are that they are much closer to standard physics theories and thus represent a better theoretical model. They take lesser solving time and have the ability to change various parameters for optimization, which is a big advantage, especially in automotive industry. This thesis work focuses on a thorough investigation of the effects of vehicle speed and road roughness on a heavy commercial vehicle ride and structural dynamic responses. Since commercial vehicles are kept in operation continuously for longer periods of time, it is important to study effects of various physical conditions on the vehicle and its user. For this purpose, various experimental as well as simulation methodologies, are adopted ranging from experimental transfer path analysis to various road scenario simulations. To effectively investigate and eliminate several causes of unwanted responses, an efficient and robust technique is needed. Carrying forward this motivation, the present work focuses on the development of a mathematical model of a 4-axle configuration heavy commercial vehicle (HCV) capable of calculating responses of the vehicle on different road PSD inputs and vehicle speeds. Outputs from the model will include response transfer functions and PSDs and wheel forces experienced. A MATLAB code will be developed to implement the objectives in a robust and flexible manner which can be exploited further in a study of responses due to various suspension parameters, loading conditions as well as vehicle dimensions. The thesis work resulted in quantifying the effect of various physical conditions on ride comfort of the vehicle. An increase in discomfort is seen with velocity increase; also the effect of road profiles has a considerable effect on comfort of the driver. Details of dominant modes at each frequency are analysed and mentioned in work. The reduction in ride height or deflection of tire and suspension with loading along with load on each axle is analysed and it is seen that the front axle supports a greater portion of vehicle weight while more of payload weight comes on fourth and third axles. The deflection of the vehicle is seen to be well inside acceptable limits.

Keywords: mathematical modeling, HCV, suspension, ride analysis

Procedia PDF Downloads 242
2989 Optimization Approach to Integrated Production-Inventory-Routing Problem for Oxygen Supply Chains

Authors: Yena Lee, Vassilis M. Charitopoulos, Karthik Thyagarajan, Ian Morris, Jose M. Pinto, Lazaros G. Papageorgiou

Abstract:

With globalisation, the need to have better coordination of production and distribution decisions has become increasingly important for industrial gas companies in order to remain competitive in the marketplace. In this work, we investigate a problem that integrates production, inventory, and routing decisions in a liquid oxygen supply chain. The oxygen supply chain consists of production facilities, external third-party suppliers, and multiple customers, including hospitals and industrial customers. The product produced by the plants or sourced from the competitors, i.e., third-party suppliers, is distributed by a fleet of heterogenous vehicles to satisfy customer demands. The objective is to minimise the total operating cost involving production, third-party, and transportation costs. The key decisions for production include production and inventory levels and product amount from third-party suppliers. In contrast, the distribution decisions involve customer allocation, delivery timing, delivery amount, and vehicle routing. The optimisation of the coordinated production, inventory, and routing decisions is a challenging problem, especially when dealing with large-size problems. Thus, we present a two-stage procedure to solve the integrated problem efficiently. First, the problem is formulated as a mixed-integer linear programming (MILP) model by simplifying the routing component. The solution from the first-stage MILP model yields the optimal customer allocation, production and inventory levels, and delivery timing and amount. Then, we fix the previous decisions and solve a detailed routing. In the second stage, we propose a column generation scheme to address the computational complexity of the resulting detailed routing problem. A case study considering a real-life oxygen supply chain in the UK is presented to illustrate the capability of the proposed models and solution method. Furthermore, a comparison of the solutions from the proposed approach with the corresponding solutions provided by existing metaheuristic techniques (e.g., guided local search and tabu search algorithms) is presented to evaluate the efficiency.

Keywords: production planning, inventory routing, column generation, mixed-integer linear programming

Procedia PDF Downloads 104
2988 Optimal 3D Deployment and Path Planning of Multiple Uavs for Maximum Coverage and Autonomy

Authors: Indu Chandran, Shubham Sharma, Rohan Mehta, Vipin Kizheppatt

Abstract:

Unmanned aerial vehicles are increasingly being explored as the most promising solution to disaster monitoring, assessment, and recovery. Current relief operations heavily rely on intelligent robot swarms to capture the damage caused, provide timely rescue, and create road maps for the victims. To perform these time-critical missions, efficient path planning that ensures quick coverage of the area is vital. This study aims to develop a technically balanced approach to provide maximum coverage of the affected area in a minimum time using the optimal number of UAVs. A coverage trajectory is designed through area decomposition and task assignment. To perform efficient and autonomous coverage mission, solution to a TSP-based optimization problem using meta-heuristic approaches is designed to allocate waypoints to the UAVs of different flight capacities. The study exploits multi-agent simulations like PX4-SITL and QGroundcontrol through the ROS framework and visualizes the dynamics of UAV deployment to different search paths in a 3D Gazebo environment. Through detailed theoretical analysis and simulation tests, we illustrate the optimality and efficiency of the proposed methodologies.

Keywords: area coverage, coverage path planning, heuristic algorithm, mission monitoring, optimization, task assignment, unmanned aerial vehicles

Procedia PDF Downloads 202
2987 Optimization of Lercanidipine Nanocrystals Using Design of Experiments Approach

Authors: Dolly Gadhiya, Jayvadan Patel, Mihir Raval

Abstract:

Lercanidipine hydrochloride is a calcium channel blockers used for treating angina pectoris and hypertension. Lercanidipine is a BCS Class II drug having poor aqueous solubility. Absolute bioavailability of Lercanidipine is very low and the main reason ascribed for this is poor aqueous solubility of the drug. Design and formulatation of nanocrystals by media milling method was main focus of this study. In this present study preliminary optimization was carried out with one factor at a time (OFAT) approach. For this different parameters like size of milling beads, amount of zirconium beads, types of stabilizer, concentrations of stabilizer, concentrations of drug, stirring speeds and milling time were optimized on the basis of particle size, polydispersity index and zeta potential. From the OFAT model different levels for above parameters selected for Plackett - Burman Design (PBD). Plackett-Burman design having 13 runs involving 6 independent variables was carried out at higher and lower level. Based on statistical analysis of PBD it was found that concentration of stabilizer, concentration of drug and stirring speed have significant impact on particle size, PDI, zeta potential value and saturation solubility. These experimental designs for preparation of nanocrystals were applied successfully which shows increase in aqueous solubility and dissolution rate of Lercanidipine hydrochloride.

Keywords: Lercanidipine hydrochloride, nanocrystals, OFAT, Plackett Burman

Procedia PDF Downloads 194
2986 On Consolidated Predictive Model of the Natural History of Breast Cancer Considering Primary Tumor and Primary Distant Metastases Growth

Authors: Ella Tyuryumina, Alexey Neznanov

Abstract:

Finding algorithms to predict the growth of tumors has piqued the interest of researchers ever since the early days of cancer research. A number of studies were carried out as an attempt to obtain reliable data on the natural history of breast cancer growth. Mathematical modeling can play a very important role in the prognosis of tumor process of breast cancer. However, mathematical models describe primary tumor growth and metastases growth separately. Consequently, we propose a mathematical growth model for primary tumor and primary metastases which may help to improve predicting accuracy of breast cancer progression using an original mathematical model referred to CoM-IV and corresponding software. We are interested in: 1) modelling the whole natural history of primary tumor and primary metastases; 2) developing adequate and precise CoM-IV which reflects relations between PT and MTS; 3) analyzing the CoM-IV scope of application; 4) implementing the model as a software tool. The CoM-IV is based on exponential tumor growth model and consists of a system of determinate nonlinear and linear equations; corresponds to TNM classification. It allows to calculate different growth periods of primary tumor and primary metastases: 1) ‘non-visible period’ for primary tumor; 2) ‘non-visible period’ for primary metastases; 3) ‘visible period’ for primary metastases. The new predictive tool: 1) is a solid foundation to develop future studies of breast cancer models; 2) does not require any expensive diagnostic tests; 3) is the first predictor which makes forecast using only current patient data, the others are based on the additional statistical data. Thus, the CoM-IV model and predictive software: a) detect different growth periods of primary tumor and primary metastases; b) make forecast of the period of primary metastases appearance; c) have higher average prediction accuracy than the other tools; d) can improve forecasts on survival of BC and facilitate optimization of diagnostic tests. The following are calculated by CoM-IV: the number of doublings for ‘nonvisible’ and ‘visible’ growth period of primary metastases; tumor volume doubling time (days) for ‘nonvisible’ and ‘visible’ growth period of primary metastases. The CoM-IV enables, for the first time, to predict the whole natural history of primary tumor and primary metastases growth on each stage (pT1, pT2, pT3, pT4) relying only on primary tumor sizes. Summarizing: a) CoM-IV describes correctly primary tumor and primary distant metastases growth of IV (T1-4N0-3M1) stage with (N1-3) or without regional metastases in lymph nodes (N0); b) facilitates the understanding of the appearance period and manifestation of primary metastases.

Keywords: breast cancer, exponential growth model, mathematical modelling, primary metastases, primary tumor, survival

Procedia PDF Downloads 327
2985 Human-Machine Cooperation in Facial Comparison Based on Likelihood Scores

Authors: Lanchi Xie, Zhihui Li, Zhigang Li, Guiqiang Wang, Lei Xu, Yuwen Yan

Abstract:

Image-based facial features can be classified into category recognition features and individual recognition features. Current automated face recognition systems extract a specific feature vector of different dimensions from a facial image according to their pre-trained neural network. However, to improve the efficiency of parameter calculation, an algorithm generally reduces the image details by pooling. The operation will overlook the details concerned much by forensic experts. In our experiment, we adopted a variety of face recognition algorithms based on deep learning, compared a large number of naturally collected face images with the known data of the same person's frontal ID photos. Downscaling and manual handling were performed on the testing images. The results supported that the facial recognition algorithms based on deep learning detected structural and morphological information and rarely focused on specific markers such as stains and moles. Overall performance, distribution of genuine scores and impostor scores, and likelihood ratios were tested to evaluate the accuracy of biometric systems and forensic experts. Experiments showed that the biometric systems were skilled in distinguishing category features, and forensic experts were better at discovering the individual features of human faces. In the proposed approach, a fusion was performed at the score level. At the specified false accept rate, the framework achieved a lower false reject rate. This paper contributes to improving the interpretability of the objective method of facial comparison and provides a novel method for human-machine collaboration in this field.

Keywords: likelihood ratio, automated facial recognition, facial comparison, biometrics

Procedia PDF Downloads 118
2984 Control of a Quadcopter Using Genetic Algorithm Methods

Authors: Mostafa Mjahed

Abstract:

This paper concerns the control of a nonlinear system using two different methods, reference model and genetic algorithm. The quadcopter is a nonlinear unstable system, which is a part of aerial robots. It is constituted by four rotors placed at the end of a cross. The center of this cross is occupied by the control circuit. Its motions are governed by six degrees of freedom: three rotations around 3 axes (roll, pitch and yaw) and the three spatial translations. The control of such system is complex, because of nonlinearity of its dynamic representation and the number of parameters, which it involves. Numerous studies have been developed to model and stabilize such systems. The classical PID and LQ correction methods are widely used. If the latter represent the advantage to be simple because they are linear, they reveal the drawback to require the presence of a linear model to synthesize. It also implies the complexity of the established laws of command because the latter must be widened on all the domain of flight of these quadcopter. Note that, if the classical design methods are widely used to control aeronautical systems, the Artificial Intelligence methods as genetic algorithms technique receives little attention. In this paper, we suggest comparing two PID design methods. Firstly, the parameters of the PID are calculated according to the reference model. In a second phase, these parameters are established using genetic algorithms. By reference model, we mean that the corrected system behaves according to a reference system, imposed by some specifications: settling time, zero overshoot etc. Inspired from the natural evolution of Darwin's theory advocating the survival of the best, John Holland developed this evolutionary algorithm. Genetic algorithm (GA) possesses three basic operators: selection, crossover and mutation. We start iterations with an initial population. Each member of this population is evaluated through a fitness function. Our purpose is to correct the behavior of the quadcopter around three axes (roll, pitch and yaw) with 3 PD controllers. For the altitude, we adopt a PID controller.

Keywords: quadcopter, genetic algorithm, PID, fitness, model, control, nonlinear system

Procedia PDF Downloads 417
2983 MHD Boundary Layer Flow of a Nanofluid Past a Wedge Shaped Wick in Heat Pipe

Authors: Ziya Uddin

Abstract:

This paper deals with the theoretical and numerical investigation of magneto-hydrodynamic boundary layer flow of a nano fluid past a wedge shaped wick in heat pipe used for the cooling of electronic components and different type of machines. To incorporate the effect of nanoparticle diameter, concentration of nanoparticles in the pure fluid, nano thermal layer formed around the nanoparticle and Brownian motion of nano particles etc., appropriate models are used for the effective thermal and physical properties of nano fluids. To model the rotation of nano particles inside the base fluid, microfluidics theory is used. In this investigation ethylene glycol (EG) based nanofluids, are taken into account. The non-linear equations governing the flow and heat transfer are solved by using a very effective particle swarm optimization technique along with Runge-Kutta method. The values of heat transfer coefficient are found for different parameters involved in the formulation viz. nanoparticle concentration, nanoparticle size, magnetic field and wedge angle etc. It is found that the wedge angle, presence of magnetic field, nanoparticle size and nanoparticle concentration etc. have prominent effects on fluid flow and heat transfer characteristics for the considered configuration.

Keywords: nanofluids, wedge shaped wick, heat pipe, numerical modeling, particle swarm optimization, nanofluid applications, Heat transfer

Procedia PDF Downloads 376
2982 Optimizing and Evaluating Performance Quality Control of the Production Process of Disposable Essentials Using Approach Vague Goal Programming

Authors: Hadi Gholizadeh, Ali Tajdin

Abstract:

To have effective production planning, it is necessary to control the quality of processes. This paper aims at improving the performance of the disposable essentials process using statistical quality control and goal programming in a vague environment. That is expressed uncertainty because there is always a measurement error in the real world. Therefore, in this study, the conditions are examined in a vague environment that is a distance-based environment. The disposable essentials process in Kach Company was studied. Statistical control tools were used to characterize the existing process for four factor responses including the average of disposable glasses’ weights, heights, crater diameters, and volumes. Goal programming was then utilized to find the combination of optimal factors setting in a vague environment which is measured to apply uncertainty of the initial information when some of the parameters of the models are vague; also, the fuzzy regression model is used to predict the responses of the four described factors. Optimization results show that the process capability index values for disposable glasses’ average of weights, heights, crater diameters and volumes were improved. Such increasing the quality of the products and reducing the waste, which will reduce the cost of the finished product, and ultimately will bring customer satisfaction, and this satisfaction, will mean increased sales.

Keywords: goal programming, quality control, vague environment, disposable glasses’ optimization, fuzzy regression

Procedia PDF Downloads 217
2981 Review of the Model-Based Supply Chain Management Research in the Construction Industry

Authors: Aspasia Koutsokosta, Stefanos Katsavounis

Abstract:

This paper reviews the model-based qualitative and quantitative Operations Management research in the context of Construction Supply Chain Management (CSCM). Construction industry has been traditionally blamed for low productivity, cost and time overruns, waste, high fragmentation and adversarial relationships. The construction industry has been slower than other industries to employ the Supply Chain Management (SCM) concept and develop models that support the decision-making and planning. However the last decade there is a distinct shift from a project-based to a supply-based approach of construction management. CSCM comes up as a new promising management tool of construction operations and improves the performance of construction projects in terms of cost, time and quality. Modeling the Construction Supply Chain (CSC) offers the means to reap the benefits of SCM, make informed decisions and gain competitive advantage. Different modeling approaches and methodologies have been applied in the multi-disciplinary and heterogeneous research field of CSCM. The literature review reveals that a considerable percentage of CSC modeling accommodates conceptual or process models which discuss general management frameworks and do not relate to acknowledged soft OR methods. We particularly focus on the model-based quantitative research and categorize the CSCM models depending on their scope, mathematical formulation, structure, objectives, solution approach, software used and decision level. Although over the last few years there has been clearly an increase of research papers on quantitative CSC models, we identify that the relevant literature is very fragmented with limited applications of simulation, mathematical programming and simulation-based optimization. Most applications are project-specific or study only parts of the supply system. Thus, some complex interdependencies within construction are neglected and the implementation of the integrated supply chain management is hindered. We conclude this paper by giving future research directions and emphasizing the need to develop robust mathematical optimization models for the CSC. We stress that CSC modeling needs a multi-dimensional, system-wide and long-term perspective. Finally, prior applications of SCM to other industries have to be taken into account in order to model CSCs, but not without the consequential reform of generic concepts to match the unique characteristics of the construction industry.

Keywords: construction supply chain management, modeling, operations research, optimization, simulation

Procedia PDF Downloads 499
2980 Soybean Seed Composition Prediction From Standing Crops Using Planet Scope Satellite Imagery and Machine Learning

Authors: Supria Sarkar, Vasit Sagan, Sourav Bhadra, Meghnath Pokharel, Felix B.Fritschi

Abstract:

Soybean and their derivatives are very important agricultural commodities around the world because of their wide applicability in human food, animal feed, biofuel, and industries. However, the significance of soybean production depends on the quality of the soybean seeds rather than the yield alone. Seed composition is widely dependent on plant physiological properties, aerobic and anaerobic environmental conditions, nutrient content, and plant phenological characteristics, which can be captured by high temporal resolution remote sensing datasets. Planet scope (PS) satellite images have high potential in sequential information of crop growth due to their frequent revisit throughout the world. In this study, we estimate soybean seed composition while the plants are in the field by utilizing PlanetScope (PS) satellite images and different machine learning algorithms. Several experimental fields were established with varying genotypes and different seed compositions were measured from the samples as ground truth data. The PS images were processed to extract 462 hand-crafted vegetative and textural features. Four machine learning algorithms, i.e., partial least squares (PLSR), random forest (RFR), gradient boosting machine (GBM), support vector machine (SVM), and two recurrent neural network architectures, i.e., long short-term memory (LSTM) and gated recurrent unit (GRU) were used in this study to predict oil, protein, sucrose, ash, starch, and fiber of soybean seed samples. The GRU and LSTM architectures had two separate branches, one for vegetative features and the other for textures features, which were later concatenated together to predict seed composition. The results show that sucrose, ash, protein, and oil yielded comparable prediction results. Machine learning algorithms that best predicted the six seed composition traits differed. GRU worked well for oil (R-Squared: of 0.53) and protein (R-Squared: 0.36), whereas SVR and PLSR showed the best result for sucrose (R-Squared: 0.74) and ash (R-Squared: 0.60), respectively. Although, the RFR and GBM provided comparable performance, the models tended to extremely overfit. Among the features, vegetative features were found as the most important variables compared to texture features. It is suggested to utilize many vegetation indices for machine learning training and select the best ones by using feature selection methods. Overall, the study reveals the feasibility and efficiency of PS images and machine learning for plot-level seed composition estimation. However, special care should be given while designing the plot size in the experiments to avoid mixed pixel issues.

Keywords: agriculture, computer vision, data science, geospatial technology

Procedia PDF Downloads 125
2979 The Closed Cavity Façade (CCF): Optimization of CCF for Enhancing Energy Efficiency and Indoor Environmental Quality in Office Buildings

Authors: Michalis Michael, Mauro Overend

Abstract:

Buildings, in which we spend 87-90% of our time, act as a shelter protecting us from environmental conditions and weather phenomena. The building's overall performance is significantly dependent on the envelope’s glazing part, which is particularly critical as it is the most vulnerable part to heat gain and heat loss. However, conventional glazing technologies have relatively low-performance thermo-optical characteristics. In this regard, during winter, the heat losses due to the glazing part of a building envelope are significantly increased as well as the heat gains during the summer period. In this study, the contribution of an innovative glazing technology, namely Closed Cavity Façade (CCF) in improving energy efficiency and IEQ in office buildings is examined, aiming to optimize various design configurations of CCF. Using Energy Plus and IDA ICE packages, the performance of several CCF configurations and geometries for various climate types were investigated, aiming to identify the optimum solution. The model used for the simulations and optimization process was MATELab, a recently constructed outdoor test facility at the University of Cambridge (UK). The model was previously experimentally calibrated. The study revealed that the use of CCF technology instead of conventional double or triple glazing leads to important benefits. Particularly, the replacement of the traditional glazing units, used as the baseline, with the optimal configuration of CCF led to a decrease in energy consumption in the range of 18-37% (depending on the location). This mainly occurs due to integrating shading devices in the cavity and applying proper glass coatings and control strategies, which lead to improvement of thermal transmittance and g-value of the glazing. Since the solar gain through the façade is the main contributor to energy consumption during cooling periods, it was observed that a higher energy improvement is achieved in cooling-dominated locations. Furthermore, it was shown that a suitable selection of the constituents of a closed cavity façade, such as the colour and type of shading devices and the type of coatings, leads to an additional improvement of its thermal performance, avoiding overheating phenomena and consequently ensuring temperatures in the glass cavity below the critical value, and reducing the radiant discomfort providing extra benefits in terms of Indoor Environmental Quality (IEQ).

Keywords: building energy efficiency, closed cavity façade, optimization, occupants comfort

Procedia PDF Downloads 53
2978 Path Planning for Unmanned Aerial Vehicles in Constrained Environments for Locust Elimination

Authors: Aadiv Shah, Hari Nair, Vedant Mittal, Alice Cheeran

Abstract:

Present-day agricultural practices such as blanket spraying not only lead to excessive usage of pesticides but also harm the overall crop yield. This paper introduces an algorithm to optimize the traversal of an unmanned aerial vehicle (UAV) in constrained environments. The proposed system focuses on the agricultural application of targeted spraying for locust elimination. Given a satellite image of a farm, target zones that are prone to locust swarm formation are detected through the calculation of the normalized difference vegetation index (NDVI). This is followed by determining the optimal path for traversal of a UAV through these target zones using the proposed algorithm in order to perform pesticide spraying in the most efficient manner possible. Unlike the classic travelling salesman problem involving point-to-point optimization, the proposed algorithm determines an optimal path for multiple regions, independent of its geometry. Finally, the paper explores the idea of implementing reinforcement learning to model complex environmental behaviour and make the path planning mechanism for UAVs agnostic to external environment changes. This system not only presents a solution to the enormous losses incurred due to locust attacks but also an efficient way to automate agricultural practices across the globe in order to improve farmer ergonomics.

Keywords: locust, NDVI, optimization, path planning, reinforcement learning, UAV

Procedia PDF Downloads 240
2977 Synthetic Optimizing Control of Wind-Wave Hybrid Energy Conversion System

Authors: Lei Xue, Liye Zhao, Jundong Wang, Yu Xue

Abstract:

A hybrid energy conversion system composed of a floating offshore wind turbine (FOWT) and wave energy converters (WECs) may possibly reduce the levelized cost of energy, improving the platform dynamics and increasing the capacity to harvest energy. This paper investigates the aerodynamic performance and dynamic responses of the combined semi-submersible FOWT and point-absorber WECs in frequency and time domains using synthetic optimizing control under turbulent wind and irregular wave conditions. Individual pitch control is applied to the FOWT part, while spring–damping control is used on the WECs part, as well as the synergistic control effect of both are studied. The effect of the above control optimization is analyzed under several typical working conditions, such as below-rated wind speed, rated wind speed, and above-rated wind speed by OpenFAST and WEC-Sim software. Particularly, the wind-wave misalignment is also comparatively investigated, which has demonstrated the importance of applying proper integrated optimal control in this hybrid energy system. More specifically, the combination of individual pitch control and spring–damping control is able to mitigate the platform pitch motion and improve output power. However, the increase in blade root load needs to be considered which needs further investigations in the future.

Keywords: floating offshore wind turbine, wave energy converters, control optimization, individual pitch control, dynamic response

Procedia PDF Downloads 46
2976 Optimization of a Flexible Thermoelectric Generator for Energy Harvesting from Human Skin to Power Wearable Electronics

Authors: Dessalegn Abera Waktole, Boru Jia, Zhengxing Zuo, Wei Wang, Nianling Kuang

Abstract:

A flexible thermoelectric generator is one method for recycling waste heat. This research provides the optimum performance of a flexible thermoelectric generator with optimal geometric parameters and a detailed structural design. In this research, a numerical simulation and experiment were carried out to develop an efficient, flexible thermoelectric generator for energy harvesting from human skin. Heteromorphic electrodes and a polyimide substrate with a copper-printed circuit board were introduced into the structural design of a flexible thermoelectric generator. The heteromorphic electrode was used as a heat sink and component of a flexible thermoelectric generator to enhance the temperature difference within the thermoelectric legs. Both N-type and P-type thermoelectric legs were made of bismuth selenium telluride (Bi1.7Te3.7Se0.3) and bismuth antimony telluride (Bi0.4Sb1.6Te3). The output power of the flexible thermoelectric generator was analyzed under different heat source temperatures and heat dissipation conditions. The COMSOL Multiphysics 5.6 software was used to conduct the simulation, which was validated by experiment. It is recorded that the maximum power output of 232.064μW was obtained by considering different wind speed conditions, the ambient temperature of 20℃, and the heat source temperature of 36℃ under various load resistance conditions, which range from 0.24Ω to 0. 91Ω. According to this finding, heteromorphic electrodes have a significant impact on the performance of the device.

Keywords: flexible thermoelectric generator, optimization, performance, temperature gradient, waste heat recovery

Procedia PDF Downloads 124
2975 Supplier Risk Management: A Multivariate Statistical Modelling and Portfolio Optimization Based Approach for Supplier Delivery Performance Development

Authors: Jiahui Yang, John Quigley, Lesley Walls

Abstract:

In this paper, the authors develop a stochastic model regarding the investment in supplier delivery performance development from a buyer’s perspective. The authors propose a multivariate model through a Multinomial-Dirichlet distribution within an Empirical Bayesian inference framework, representing both the epistemic and aleatory uncertainties in deliveries. A closed form solution is obtained and the lower and upper bound for both optimal investment level and expected profit under uncertainty are derived. The theoretical properties provide decision makers with useful insights regarding supplier delivery performance improvement problems where multiple delivery statuses are involved. The authors also extend the model from a single supplier investment into a supplier portfolio, using a Lagrangian method to obtain a theoretical expression for an optimal investment level and overall expected profit. The model enables a buyer to know how the marginal expected profit/investment level of each supplier changes with respect to the budget and which supplier should be invested in when additional budget is available. An application of this model is illustrated in a simulation study. Overall, the main contribution of this study is to provide an optimal investment decision making framework for supplier development, taking into account multiple delivery statuses as well as multiple projects.

Keywords: decision making, empirical bayesian, portfolio optimization, supplier development, supply chain management

Procedia PDF Downloads 279
2974 Review on Crew Scheduling of Bus Transit: A Case Study in Kolkata

Authors: Sapan Tiwari, Namrata Ghosh

Abstract:

In urban mass transit, crew scheduling always plays a significant role. It deals with the formulation of work timetables for its staff so that an organization can meet the demand for its products or services. The efficient schedules of a specified timetable have an enormous impact on staff demand. It implies that an urban mass transit company's financial outcomes are strongly associated with planning operations in the region. The research aims to demonstrate the state of the crew scheduling studies and its practical implementation in mass transit businesses in metropolitan areas. First, there is a short overview of past studies in the field. Subsequently, the restrictions and problems with crew scheduling and some models, which have been developed to solve the related issues with their mathematical formulation, are defined. The comments are completed by a description of the solution opportunities provided by computer-aided scheduling program systems for operational use and exposures from urban mass transit organizations. Furthermore, Bus scheduling is performed using the Hungarian technique of problem-solving tasks and mathematical modeling. Afterward, the crew scheduling problem, which consists of developing duties using predefined tasks with set start and end times and places, is resolved. Each duty has to comply with a set line of work. The objective is to minimize a mixture of fixed expenses (number of duties) and varying costs. After the optimization of cost, the outcome of the research is that the same frequency can be provided with fewer buses and less workforce.

Keywords: crew scheduling, duty, optimization of cost, urban mass transit

Procedia PDF Downloads 142
2973 Process Optimization for 2205 Duplex Stainless Steel by Laser Metal Deposition

Authors: Siri Marthe Arbo, Afaf Saai, Sture Sørli, Mette Nedreberg

Abstract:

This work aims to establish a reliable approach for optimizing a Laser Metal Deposition (LMD) process for a critical maritime component, based on the material properties and structural performance required by the maritime industry. The component of interest is a water jet impeller, for which specific requirements for material properties are defined. The developed approach is based on the assessment of the effects of LMD process parameters on microstructure and material performance of standard AM 2205 duplex stainless steel powder. Duplex stainless steel offers attractive properties for maritime applications, combining high strength, enhanced ductility and excellent corrosion resistance due to the specific amounts of ferrite and austenite. These properties are strongly affected by the microstructural characteristics in addition to microstructural defects such as porosity and welding defects, all strongly influenced by the chosen LMD process parameters. In this study, the influence of deposition speed and heat input was evaluated. First, the influences of deposition speed and heat input on the microstructure characteristics, including ferrite/austenite fraction, amount of porosity and welding defects, were evaluated. Then, the achieved mechanical properties were evaluated by standard testing methods, measuring the hardness, tensile strength and elongation, bending force and impact energy. The measured properties were compared to the requirements of the water jet impeller. The results show that the required amounts of ferrite and austenite can be achieved directly by the LMD process without post-weld heat treatments. No intermetallic phases were observed in the material produced by the investigated process parameters. A high deposition speed was found to reduce the ductility due to the formation of welding defects. An increased heat input was associated with reduced strength due to the coarsening of the ferrite/austenite microstructure. The microstructure characterizations and measured mechanical performance demonstrate the great potential of the LMD process and generate a valuable database for the optimization of the LMD process for duplex stainless steels.

Keywords: duplex stainless steel, laser metal deposition, process optimization, microstructure, mechanical properties

Procedia PDF Downloads 208
2972 Bioinformatics High Performance Computation and Big Data

Authors: Javed Mohammed

Abstract:

Right now, bio-medical infrastructure lags well behind the curve. Our healthcare system is dispersed and disjointed; medical records are a bit of a mess; and we do not yet have the capacity to store and process the crazy amounts of data coming our way from widespread whole-genome sequencing. And then there are privacy issues. Despite these infrastructure challenges, some researchers are plunging into bio medical Big Data now, in hopes of extracting new and actionable knowledge. They are doing delving into molecular-level data to discover bio markers that help classify patients based on their response to existing treatments; and pushing their results out to physicians in novel and creative ways. Computer scientists and bio medical researchers are able to transform data into models and simulations that will enable scientists for the first time to gain a profound under-standing of the deepest biological functions. Solving biological problems may require High-Performance Computing HPC due either to the massive parallel computation required to solve a particular problem or to algorithmic complexity that may range from difficult to intractable. Many problems involve seemingly well-behaved polynomial time algorithms (such as all-to-all comparisons) but have massive computational requirements due to the large data sets that must be analyzed. High-throughput techniques for DNA sequencing and analysis of gene expression have led to exponential growth in the amount of publicly available genomic data. With the increased availability of genomic data traditional database approaches are no longer sufficient for rapidly performing life science queries involving the fusion of data types. Computing systems are now so powerful it is possible for researchers to consider modeling the folding of a protein or even the simulation of an entire human body. This research paper emphasizes the computational biology's growing need for high-performance computing and Big Data. It illustrates this article’s indispensability in meeting the scientific and engineering challenges of the twenty-first century, and how Protein Folding (the structure and function of proteins) and Phylogeny Reconstruction (evolutionary history of a group of genes) can use HPC that provides sufficient capability for evaluating or solving more limited but meaningful instances. This article also indicates solutions to optimization problems, and benefits Big Data and Computational Biology. The article illustrates the Current State-of-the-Art and Future-Generation Biology of HPC Computing with Big Data.

Keywords: high performance, big data, parallel computation, molecular data, computational biology

Procedia PDF Downloads 357
2971 Efficient Use of Power Light-Emitting Diode Chips in the Main Lighting System and in Generating Heat in Intelligent Buildings

Authors: Siamak Eskandari, Neda Ebadi

Abstract:

Among common electronic parts which have been invented and have made a great revolution in the lighting system through the world, certainly LEDs have no rival. These small parts with their very low power consumption, very dazzling and powerful light and small size and with their extremely high lifetime- compared to incandescent bulbs and compact fluorescent lamp (CFLs) have undoubtedly revolutionized the lighting industry of the world. Based on conducted studies and experiments, in addition to their acceptable light and low power consumption -compared to incandescent bulbs and CFLs-, they have very low and in some cases zero environmental pollution and negative effects on human beings. Because of their longevity, in the case of using high-quality circuits and proper and consistent use of LEDs in conventional and intelligent buildings, there will be no need to replace the burnout lamps, for a long time (10 years). In this study which was conducted on 10-watt power LEDs with suitable heatsink/cooling, considerable amount of heat was generated during lighting after 5 minutes and 45 seconds. The temperature rose to above 99 degrees Celsius and this amount of heat can raise the water temperature to 60 degrees Celsius and more. Based on conducted experiments, this can provide the heat required for bathing, washing, radiators (in cold seasons) easily and only by imposing very low cost and it will be a big step in the optimization of energy consumption in the future.

Keywords: energy, light, water, optimization of power LED

Procedia PDF Downloads 139
2970 Application of Argumentation for Improving the Classification Accuracy in Inductive Concept Formation

Authors: Vadim Vagin, Marina Fomina, Oleg Morosin

Abstract:

This paper contains the description of argumentation approach for the problem of inductive concept formation. It is proposed to use argumentation, based on defeasible reasoning with justification degrees, to improve the quality of classification models, obtained by generalization algorithms. The experiment’s results on both clear and noisy data are also presented.

Keywords: argumentation, justification degrees, inductive concept formation, noise, generalization

Procedia PDF Downloads 430
2969 Design Optimization of a Micro Compressor for Micro Gas Turbine Using Computational Fluid Dynamics

Authors: Kamran Siddique, Hiroyuki Asada, Yoshifumi Ogami

Abstract:

The use of Micro Gas Turbine (MGT) as the engine in Unmanned Aerobic Vehicles (UAVs) and power source in Robotics is widespread these days. Research has been conducted in the past decade or so to improve the performance of different components of MGT. This type of engine has interrelated components which have non-linear characteristics. Therefore, the overall engine performance depends on the individual engine element’s performance. Computational Fluid Dynamics (CFD) is one of the simulation method tools used to analyze or even optimize MGT system performance. In this study, the compressor of the MGT is designed, and performance optimization is being done using CFD. Performance of the micro compressor is improved in order to increase the overall performance of MGT. A high value of pressure ratio is to be achieved by studying the effect of change of different operating parameters like mass flow rate and revolutions per minute (RPM) and aerodynamical and geometrical parameters on the pressure ratio of the compressor. Two types of compressor designs are considered in this study; 3D centrifugal and ‘planar’ designs. For a 10 mm impeller, the planar model is the simplest compressor model with the ease in manufacturability. On the other hand, 3D centrifugal model, although more efficient, is very difficult to manufacture using current microfabrication resources. Therefore, the planar model is the best-suited model for a micro compressor. So. a planar micro compressor has been designed that has a good pressure ratio, and it is easy to manufacture using current microfabrication technologies. Future work is to fabricate the compressor to get experimental results and validate the theoretical model.

Keywords: computational fluid dynamics, microfabrication, MEMS, unmanned aerobic vehicles

Procedia PDF Downloads 134
2968 Unknown Groundwater Pollution Source Characterization in Contaminated Mine Sites Using Optimal Monitoring Network Design

Authors: H. K. Esfahani, B. Datta

Abstract:

Groundwater is one of the most important natural resources in many parts of the world; however it is widely polluted due to human activities. Currently, effective and reliable groundwater management and remediation strategies are obtained using characterization of groundwater pollution sources, where the measured data in monitoring locations are utilized to estimate the unknown pollutant source location and magnitude. However, accurately identifying characteristics of contaminant sources is a challenging task due to uncertainties in terms of predicting source flux injection, hydro-geological and geo-chemical parameters, and the concentration field measurement. Reactive transport of chemical species in contaminated groundwater systems, especially with multiple species, is a complex and highly non-linear geochemical process. Although sufficient concentration measurement data is essential to accurately identify sources characteristics, available data are often sparse and limited in quantity. Therefore, this inverse problem-solving method for characterizing unknown groundwater pollution sources is often considered ill-posed, complex and non- unique. Different methods have been utilized to identify pollution sources; however, the linked simulation-optimization approach is one effective method to obtain acceptable results under uncertainties in complex real life scenarios. With this approach, the numerical flow and contaminant transport simulation models are externally linked to an optimization algorithm, with the objective of minimizing the difference between measured concentration and estimated pollutant concentration at observation locations. Concentration measurement data are very important to accurately estimate pollution source properties; therefore, optimal design of the monitoring network is essential to gather adequate measured data at desired times and locations. Due to budget and physical restrictions, an efficient and effective approach for groundwater pollutant source characterization is to design an optimal monitoring network, especially when only inadequate and arbitrary concentration measurement data are initially available. In this approach, preliminary concentration observation data are utilized for preliminary source location, magnitude and duration of source activity identification, and these results are utilized for monitoring network design. Further, feedback information from the monitoring network is used as inputs for sequential monitoring network design, to improve the identification of unknown source characteristics. To design an effective monitoring network of observation wells, optimization and interpolation techniques are used. A simulation model should be utilized to accurately describe the aquifer properties in terms of hydro-geochemical parameters and boundary conditions. However, the simulation of the transport processes becomes complex when the pollutants are chemically reactive. Three dimensional transient flow and reactive contaminant transport process is considered. The proposed methodology uses HYDROGEOCHEM 5.0 (HGCH) as the simulation model for flow and transport processes with chemically multiple reactive species. Adaptive Simulated Annealing (ASA) is used as optimization algorithm in linked simulation-optimization methodology to identify the unknown source characteristics. Therefore, the aim of the present study is to develop a methodology to optimally design an effective monitoring network for pollution source characterization with reactive species in polluted aquifers. The performance of the developed methodology will be evaluated for an illustrative polluted aquifer sites, for example an abandoned mine site in Queensland, Australia.

Keywords: monitoring network design, source characterization, chemical reactive transport process, contaminated mine site

Procedia PDF Downloads 223
2967 Freight Time and Cost Optimization in Complex Logistics Networks, Using a Dimensional Reduction Method and K-Means Algorithm

Authors: Egemen Sert, Leila Hedayatifar, Rachel A. Rigg, Amir Akhavan, Olha Buchel, Dominic Elias Saadi, Aabir Abubaker Kar, Alfredo J. Morales, Yaneer Bar-Yam

Abstract:

The complexity of providing timely and cost-effective distribution of finished goods from industrial facilities to customers makes effective operational coordination difficult, yet effectiveness is crucial for maintaining customer service levels and sustaining a business. Logistics planning becomes increasingly complex with growing numbers of customers, varied geographical locations, the uncertainty of future orders, and sometimes extreme competitive pressure to reduce inventory costs. Linear optimization methods become cumbersome or intractable due to a large number of variables and nonlinear dependencies involved. Here we develop a complex systems approach to optimizing logistics networks based upon dimensional reduction methods and apply our approach to a case study of a manufacturing company. In order to characterize the complexity in customer behavior, we define a “customer space” in which individual customer behavior is described by only the two most relevant dimensions: the distance to production facilities over current transportation routes and the customer's demand frequency. These dimensions provide essential insight into the domain of effective strategies for customers; direct and indirect strategies. In the direct strategy, goods are sent to the customer directly from a production facility using box or bulk trucks. In the indirect strategy, in advance of an order by the customer, goods are shipped to an external warehouse near a customer using trains and then "last-mile" shipped by trucks when orders are placed. Each strategy applies to an area of the customer space with an indeterminate boundary between them. Specific company policies determine the location of the boundary generally. We then identify the optimal delivery strategy for each customer by constructing a detailed model of costs of transportation and temporary storage in a set of specified external warehouses. Customer spaces help give an aggregate view of customer behaviors and characteristics. They allow policymakers to compare customers and develop strategies based on the aggregate behavior of the system as a whole. In addition to optimization over existing facilities, using customer logistics and the k-means algorithm, we propose additional warehouse locations. We apply these methods to a medium-sized American manufacturing company with a particular logistics network, consisting of multiple production facilities, external warehouses, and customers along with three types of shipment methods (box truck, bulk truck and train). For the case study, our method forecasts 10.5% savings on yearly transportation costs and an additional 4.6% savings with three new warehouses.

Keywords: logistics network optimization, direct and indirect strategies, K-means algorithm, dimensional reduction

Procedia PDF Downloads 130
2966 Modeling Operating Theater Scheduling and Configuration: An Integrated Model in Health-Care Logistics

Authors: Sina Keyhanian, Abbas Ahmadi, Behrooz Karimi

Abstract:

We present a multi-objective binary programming model which considers surgical cases are scheduling among operating rooms and the configuration of surgical instruments in limited capacity hospital trays, simultaneously. Many mathematical models have been developed previously in the literature addressing different challenges in health-care logistics such as assigning operating rooms, leveling beds, etc. But what happens inside the operating rooms along with the inventory management of required instruments for various operations, and also their integration with surgical scheduling have been poorly discussed. Our model considers the minimization of movements between trays during a surgery which recalls the famous cell formation problem in group technology. This assumption can also provide a major potential contribution to robotic surgeries. The tray configuration problem which consumes surgical instruments requirement plan (SIRP) and sequence of surgical procedures based on required instruments (SIRO) is nested inside the bin packing problem. This modeling approach helps us understand that most of the same-output solutions will not be necessarily identical when it comes to the rearrangement of surgeries among rooms. A numerical example has been dealt with via a proposed nested simulated annealing (SA) optimization approach which provides insights about how various configurations inside a solution can alter the optimal condition.

Keywords: health-care logistics, hospital tray configuration, off-line bin packing, simulated annealing optimization, surgical case scheduling

Procedia PDF Downloads 267
2965 Optimization of Economic Order Quantity of Multi-Item Inventory Control Problem through Nonlinear Programming Technique

Authors: Prabha Rohatgi

Abstract:

To obtain an efficient control over a huge amount of inventory of drugs in pharmacy department of any hospital, generally, the medicines are categorized on the basis of their cost ‘ABC’ (Always Better Control), first and then categorize on the basis of their criticality ‘VED’ (Vital, Essential, desirable) for prioritization. About one-third of the annual expenditure of a hospital is spent on medicines. To minimize the inventory investment, the hospital management may like to keep the medicines inventory low, as medicines are perishable items. The main aim of each and every hospital is to provide better services to the patients under certain limited resources. To achieve the satisfactory level of health care services to outdoor patients, a hospital has to keep eye on the wastage of medicines because expiry date of medicines causes a great loss of money though it was limited and allocated for a particular period of time. The objectives of this study are to identify the categories of medicines requiring incentive managerial control. In this paper, to minimize the total inventory cost and the cost associated with the wastage of money due to expiry of medicines, an inventory control model is used as an estimation tool and then nonlinear programming technique is used under limited budget and fixed number of orders to be placed in a limited time period. Numerical computations have been given and shown that by using scientific methods in hospital services, we can give more effective way of inventory management under limited resources and can provide better health care services. The secondary data has been collected from a hospital to give empirical evidence.

Keywords: ABC-VED inventory classification, multi item inventory problem, nonlinear programming technique, optimization of EOQ

Procedia PDF Downloads 245
2964 Optimal Harmonic Filters Design of Taiwan High Speed Rail Traction System

Authors: Ying-Pin Chang

Abstract:

This paper presents a method for combining a particle swarm optimization with nonlinear time-varying evolution and orthogonal arrays (PSO-NTVEOA) in the planning of harmonic filters for the high speed railway traction system with specially connected transformers in unbalanced three-phase power systems. The objective is to minimize the cost of the filter, the filters loss, the total harmonic distortion of currents and voltages at each bus simultaneously. An orthogonal array is first conducted to obtain the initial solution set. The set is then treated as the initial training sample. Next, the PSO-NTVEOA method parameters are determined by using matrix experiments with an orthogonal array, in which a minimal number of experiments would have an effect that approximates the full factorial experiments. This PSO-NTVEOA method is then applied to design optimal harmonic filters in Taiwan High Speed Rail (THSR) traction system, where both rectifiers and inverters with IGBT are used. From the results of the illustrative examples, the feasibility of the PSO-NTVEOA to design an optimal passive harmonic filter of THSR system is verified and the design approach can greatly reduce the harmonic distortion. Three design schemes are compared that V-V connection suppressing the 3rd order harmonic, and Scott and Le Blanc connection for the harmonic improvement is better than the V-V connection.

Keywords: harmonic filters, particle swarm optimization, nonlinear time-varying evolution, orthogonal arrays, specially connected transformers

Procedia PDF Downloads 381
2963 Implementation of Proof of Work Using Ganache

Authors: Sakshi Singh, Shampa Chakraverty

Abstract:

One of the essential characteristics of Blockchain is the ability to validate the integrity of new transactions added to the Blockchain. Moreover, one of the essential consensus algorithms, Proof of Work, performs this job. In this work, we implemented the Proof of Work consensus method on the block formed by performing the transaction using Ganache. The primary goal of this implementation is to understand the process and record how Proof of Work works in reality on newly created blocks.

Keywords: proof of work, blockchain, ganache, smart contract

Procedia PDF Downloads 154
2962 Optimization Of Biogas Production Using Co-digestion Feedstocks Via Anaerobic Technologhy

Authors: E Tolufase

Abstract:

The demand, high costs and health implications of using energy derived from hydrocarbon compound have necessitated the continuous search for alternative source of energy. The World energy market is facing some challenges viz: depletion of fossil fuel reserves, population explosion, lack of energy security, economic and urbanization growth and also, in Nigeria some rural areas still depend largely on wood, charcoal, kerosene, petrol among others, as the sources of their energy. To overcome these short falls in energy supply and demand, as well as taking into consideration the risks from global climate change due to effect of greenhouse gas emissions and other pollutants from fossil fuels’ combustion, brought a lot of attention on efficiently harnessing the renewable energy sources. A very promising among the renewable energy resources for a clean energy technology for power production, vehicle and domestic usage is biogas. Therefore, optimization of biogas yield and quality is imperative. Hence, this study investigated yield and quality of biogas using low cost bio-digester and combination of various feed stocks referred to as co-digestion. Batch/Discontinuous Bio-digester type was used because it was cheap, easy, plausible and appropriate for different substrates used to get the desired results. Three substrates were used; cow dung, chicken droppings and lemon grass digested in five separate 21 litre digesters, A, B, C, D, and E and the gas collection system was designed using locally available materials. For single digestion we had; cow dung, chicken droppings, lemon grass, in Bio-digesters A, B, and C respectively, the co-digested three substrates in different mixed ratio 7:1:2 in digester D and E in ratio 5:3:2. The respective feed-stocks materials were collected locally, digested and analyzed in accordance with standard procedures. They were pre-fermented for a period of 10 days before being introduced into the digesters. They were digested for a retention period of 28 days, the physiochemical parameters namely; pressure, temperature, pH, volume of the gas collector system and volume of biogas produced were all closely monitored and recorded daily. The values of pH and temperature ranged 6.0 - 8.0, and 220C- 350C respectively. For the single substrate, bio-digester A(Cow dung only) produced biogas of total volume 0.1607m3(average volume of 0.0054m3 daily),while B (Chicken droppings ) produced 0.1722m3 (average of 0.0057m3 daily) and C (lemon grass) produced 0.1035m3 (average of 0.0035m3 daily). For the co-digested substrates in bio-digester D the total biogas produced was 0.2007m³ (average volume of 0.0067m³ daily) and bio-digester E produced 0.1991m³ (average volume of 0.0066m³ daily) It’s obvious from the results, that combining different substrates gave higher yields than when a singular feed stock was used and also mixing ratio played some roles in the yield improvement. Bio-digesters D and E contained the same substrates but mixed with different ratios, but higher yield was noticed in D with mixing ratio of 7:1:2 than in E with ratio 5:3:2.Therefore, co-digestion of substrates and mixing proportions are important factors for biogas production optimization.

Keywords: anaerobic, batch, biogas, biodigester, digestion, fermentation, optimization

Procedia PDF Downloads 10
2961 Parameters Estimation of Multidimensional Possibility Distributions

Authors: Sergey Sorokin, Irina Sorokina, Alexander Yazenin

Abstract:

We present a solution to the Maxmin u/E parameters estimation problem of possibility distributions in m-dimensional case. Our method is based on geometrical approach, where minimal area enclosing ellipsoid is constructed around the sample. Also we demonstrate that one can improve results of well-known algorithms in fuzzy model identification task using Maxmin u/E parameters estimation.

Keywords: possibility distribution, parameters estimation, Maxmin u\E estimator, fuzzy model identification

Procedia PDF Downloads 460