Search results for: test suite optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 12014

Search results for: test suite optimization

11324 Reliability Analysis of Variable Stiffness Composite Laminate Structures

Authors: A. Sohouli, A. Suleman

Abstract:

This study focuses on reliability analysis of variable stiffness composite laminate structures to investigate the potential structural improvement compared to conventional (straight fibers) composite laminate structures. A computational framework was developed which it consists of a deterministic design step and reliability analysis. The optimization part is Discrete Material Optimization (DMO) and the reliability of the structure is computed by Monte Carlo Simulation (MCS) after using Stochastic Response Surface Method (SRSM). The design driver in deterministic optimization is the maximum stiffness, while optimization method concerns certain manufacturing constraints to attain industrial relevance. These manufacturing constraints are the change of orientation between adjacent patches cannot be too large and the maximum number of successive plies of a particular fiber orientation should not be too high. Variable stiffness composites may be manufactured by Automated Fiber Machines (AFP) which provides consistent quality with good production rates. However, laps and gaps are the most important challenges to steer fibers that effect on the performance of the structures. In this study, the optimal curved fiber paths at each layer of composites are designed in the first step by DMO, and then the reliability analysis is applied to investigate the sensitivity of the structure with different standard deviations compared to the straight fiber angle composites. The random variables are material properties and loads on the structures. The results show that the variable stiffness composite laminate structures are much more reliable, even for high standard deviation of material properties, than the conventional composite laminate structures. The reason is that the variable stiffness composite laminates allow tailoring stiffness and provide the possibility of adjusting stress and strain distribution favorably in the structures.

Keywords: material optimization, Monte Carlo simulation, reliability analysis, response surface method, variable stiffness composite structures

Procedia PDF Downloads 498
11323 Influence of Radio Frequency Identification Technology at Cost of Supply Chain as a Driver for the Generation of Competitive Advantage

Authors: Mona Baniahmadi, Saied Haghanifar

Abstract:

Radio Frequency Identification (RFID) is regarded as a promising technology for the optimization of supply chain processes since it improves manufacturing and retail operations from forecasting demand for planning, managing inventory, and distribution. This study precisely aims at learning to know the RFID technology and at explaining how it can concretely be used for supply chain management and how it can help improving it in the case of Hejrat Company which is located in Iran and works on the distribution of medical drugs and cosmetics. This study uses some statistical analysis to calculate the expected benefits of an integrated RFID system on supply chain obtained through competitive advantages increases with decreasing cost factor. The study investigates how the cost of storage process, labor cost, the cost of missing goods, inventory management optimization, on-time delivery, order cost, lost sales and supply process optimization affect the performance of the integrated RFID supply chain regarding cost factors and provides a competitive advantage.

Keywords: cost, competitive advantage, radio frequency identification, supply chain

Procedia PDF Downloads 255
11322 A Linear Programming Approach to Assist Roster Construction Under a Salary Cap

Authors: Alex Contarino

Abstract:

Professional sports leagues often have a “free agency” period, during which teams may sign players with expiring contracts.To promote parity, many leagues operate under a salary cap that limits the amount teams can spend on player’s salaries in a given year. Similarly, in fantasy sports leagues, salary cap drafts are a popular method for selecting players. In order to sign a free agent in either setting, teams must bid against one another to buy the player’s services while ensuring the sum of their player’s salaries is below the salary cap. This paper models the bidding process for a free agent as a constrained optimization problem that can be solved using linear programming. The objective is to determine the largest bid that a team should offer the player subject to the constraint that the value of signing the player must exceed the value of using the salary cap elsewhere. Iteratively solving this optimization problem for each available free agent provides teams with an effective framework for maximizing the talent on their rosters. The utility of this approach is demonstrated for team sport roster construction and fantasy sport drafts, using recent data sets from both settings.

Keywords: linear programming, optimization, roster management, salary cap

Procedia PDF Downloads 98
11321 The Effect of Fetal Movement Counting on Maternal Antenatal Attachment

Authors: Esra Güney, Tuba Uçar

Abstract:

Aim: This study has been conducted for the purpose of determining the effects of fetal movement counting on antenatal maternal attachment. Material and Method: This research was conducted on the basis of the real test model with the pre-test /post-test control groups. The study population consists of pregnant women registered in the six different Family Health Centers located in the central Malatya districts of Yeşilyurt and Battalgazi. When power analysis is done, the sample size was calculated for each group of at least 55 pregnant women (55 tests, 55 controls). The data were collected by using Personal Information Form and MAAS (Maternal Antenatal Attachment Scale) between July 2015-June 2016. Fetal movement counting training was given to pregnant women by researchers in the experimental group after the pre-test data collection. No intervention was applied to the control group. Post-test data for both groups were collected after four weeks. Data were evaluated with percentage, chi-square arithmetic average, chi-square test and as for the dependent and independent group’s t test. Result: In the MAAS, the pre-test average of total scores in the experimental group is 70.78±6.78, control group is also 71.58±7.54 and so there was no significant difference in mean scores between the two groups (p>0.05). MAAS post-test average of total scores in the experimental group is 78.41±6.65, control group is also is 72.25±7.16 and so the mean scores between groups were found to have statistically significant difference (p<0.05). Conclusion: It was determined that fetal movement counting increases the maternal antenatal attachments.

Keywords: antenatal maternal attachment, fetal movement counting, pregnancy, midwifery

Procedia PDF Downloads 250
11320 Market Solvency Capital Requirement Minimization: How Non-linear Solvers Provide Portfolios Complying with Solvency II Regulation

Authors: Abraham Castellanos, Christophe Durville, Sophie Echenim

Abstract:

In this article, a portfolio optimization problem is performed in a Solvency II context: it illustrates how advanced optimization techniques can help to tackle complex operational pain points around the monitoring, control, and stability of Solvency Capital Requirement (SCR). The market SCR of a portfolio is calculated as a combination of SCR sub-modules. These sub-modules are the results of stress-tests on interest rate, equity, property, credit and FX factors, as well as concentration on counter-parties. The market SCR is non convex and non differentiable, which does not make it a natural optimization criteria candidate. In the SCR formulation, correlations between sub-modules are fixed, whereas risk-driven portfolio allocation is usually driven by the dynamics of the actual correlations. Implementing a portfolio construction approach that is efficient on both a regulatory and economic standpoint is not straightforward. Moreover, the challenge for insurance portfolio managers is not only to achieve a minimal SCR to reduce non-invested capital but also to ensure stability of the SCR. Some optimizations have already been performed in the literature, simplifying the standard formula into a quadratic function. But to our knowledge, it is the first time that the standard formula of the market SCR is used in an optimization problem. Two solvers are combined: a bundle algorithm for convex non- differentiable problems, and a BFGS (Broyden-Fletcher-Goldfarb- Shanno)-SQP (Sequential Quadratic Programming) algorithm, to cope with non-convex cases. A market SCR minimization is then performed with historical data. This approach results in significant reduction of the capital requirement, compared to a classical Markowitz approach based on the historical volatility. A comparative analysis of different optimization models (equi-risk-contribution portfolio, minimizing volatility portfolio and minimizing value-at-risk portfolio) is performed and the impact of these strategies on risk measures including market SCR and its sub-modules is evaluated. A lack of diversification of market SCR is observed, specially for equities. This was expected since the market SCR strongly penalizes this type of financial instrument. It was shown that this direct effect of the regulation can be attenuated by implementing constraints in the optimization process or minimizing the market SCR together with the historical volatility, proving the interest of having a portfolio construction approach that can incorporate such features. The present results are further explained by the Market SCR modelling.

Keywords: financial risk, numerical optimization, portfolio management, solvency capital requirement

Procedia PDF Downloads 100
11319 Hybrid Data-Driven Drilling Rate of Penetration Optimization Scheme Guided by Geological Formation and Historical Data

Authors: Ammar Alali, Mahmoud Abughaban, William Contreras Otalvora

Abstract:

Optimizing the drilling process for cost and efficiency requires the optimization of the rate of penetration (ROP). ROP is the measurement of the speed at which the wellbore is created, in units of feet per hour. It is the primary indicator of measuring drilling efficiency. Maximization of the ROP can indicate fast and cost-efficient drilling operations; however, high ROPs may induce unintended events, which may lead to nonproductive time (NPT) and higher net costs. The proposed ROP optimization solution is a hybrid, data-driven system that aims to improve the drilling process, maximize the ROP, and minimize NPT. The system consists of two phases: (1) utilizing existing geological and drilling data to train the model prior, and (2) real-time adjustments of the controllable dynamic drilling parameters [weight on bit (WOB), rotary speed (RPM), and pump flow rate (GPM)] that direct influence on the ROP. During the first phase of the system, geological and historical drilling data are aggregated. After, the top-rated wells, as a function of high instance ROP, are distinguished. Those wells are filtered based on NPT incidents, and a cross-plot is generated for the controllable dynamic drilling parameters per ROP value. Subsequently, the parameter values (WOB, GPM, RPM) are calculated as a conditioned mean based on physical distance, following Inverse Distance Weighting (IDW) interpolation methodology. The first phase is concluded by producing a model of drilling best practices from the offset wells, prioritizing the optimum ROP value. This phase is performed before the commencing of drilling. Starting with the model produced in phase one, the second phase runs an automated drill-off test, delivering live adjustments in real-time. Those adjustments are made by directing the driller to deviate two of the controllable parameters (WOB and RPM) by a small percentage (0-5%), following the Constrained Random Search (CRS) methodology. These minor incremental variations will reveal new drilling conditions, not explored before through offset wells. The data is then consolidated into a heat-map, as a function of ROP. A more optimum ROP performance is identified through the heat-map and amended in the model. The validation process involved the selection of a planned well in an onshore oil field with hundreds of offset wells. The first phase model was built by utilizing the data points from the top-performing historical wells (20 wells). The model allows drillers to enhance decision-making by leveraging existing data and blending it with live data in real-time. An empirical relationship between controllable dynamic parameters and ROP was derived using Artificial Neural Networks (ANN). The adjustments resulted in improved ROP efficiency by over 20%, translating to at least 10% saving in drilling costs. The novelty of the proposed system lays is its ability to integrate historical data, calibrate based geological formations, and run real-time global optimization through CRS. Those factors position the system to work for any newly drilled well in a developing field event.

Keywords: drilling optimization, geological formations, machine learning, rate of penetration

Procedia PDF Downloads 111
11318 Application of Hybrid Honey Bees Mating Optimization Algorithm in Multiuser Detection of Wireless Communication Systems

Authors: N. Larbi, F. Debbat

Abstract:

Wireless communication systems have changed dramatically and shown spectacular evolution over the past two decades. These radio technologies are engaged in a quest endless high-speed transmission coupled to a constant need to improve transmission quality. Various radio communication systems being developed use code division multiple access (CDMA) technique. This work analyses a hybrid honey bees mating optimization algorithm (HBMO) applied to multiuser detection (MuD) in CDMA communication systems. The HBMO is a swarm-based optimization algorithm, which simulates the mating process of real honey bees. We apply a hybridization of HBMO with simulated annealing (SA) in order to improve the solution generated by the HBMO. Simulation results show that the detection based on Hybrid HBMO, in term of bit error rate (BER), is viable option when compared with the classic detectors from literature under Rayleigh flat fading channel.

Keywords: BER, DS-CDMA multiuser detection, genetic algorithm, hybrid HBMO, simulated annealing

Procedia PDF Downloads 415
11317 On the Application of Heuristics of the Traveling Salesman Problem for the Task of Restoring the DNA Matrix

Authors: Boris Melnikov, Dmitrii Chaikovskii, Elena Melnikova

Abstract:

The traveling salesman problem (TSP) is a well-known optimization problem that seeks to find the shortest possible route that visits a set of points and returns to the starting point. In this paper, we apply some heuristics of the TSP for the task of restoring the DNA matrix. This restoration problem is often considered in biocybernetics. For it, we must recover the matrix of distances between DNA sequences if not all the elements of the matrix under consideration are known at the input. We consider the possibility of using this method in the testing of distance calculation algorithms between a pair of DNAs to restore the partially filled matrix.

Keywords: optimization problems, DNA matrix, partially filled matrix, traveling salesman problem, heuristic algorithms

Procedia PDF Downloads 133
11316 In Silico Exploration of Quinazoline Derivatives as EGFR Inhibitors for Lung Cancer: A Multi-Modal Approach Integrating QSAR-3D, ADMET, Molecular Docking, and Molecular Dynamics Analyses

Authors: Mohamed Moussaoui

Abstract:

A series of thirty-one potential inhibitors targeting the epidermal growth factor receptor kinase (EGFR), derived from quinazoline, underwent 3D-QSAR analysis using CoMFA and CoMSIA methodologies. The training and test sets of quinazoline derivatives were utilized to construct and validate the QSAR models, respectively, with dataset alignment performed using the lowest energy conformer of the most active compound. The best-performing CoMFA and CoMSIA models demonstrated impressive determination coefficients, with R² values of 0.981 and 0.978, respectively, and Leave One Out cross-validation determination coefficients, Q², of 0.645 and 0.729, respectively. Furthermore, external validation using a test set of five compounds yielded predicted determination coefficients, R² test, of 0.929 and 0.909 for CoMFA and CoMSIA, respectively. Building upon these promising results, eighteen new compounds were designed and assessed for drug likeness and ADMET properties through in silico methods. Additionally, molecular docking studies were conducted to elucidate the binding interactions between the selected compounds and the enzyme. Detailed molecular dynamics simulations were performed to analyze the stability, conformational changes, and binding interactions of the quinazoline derivatives with the EGFR kinase. These simulations provided deeper insights into the dynamic behavior of the compounds within the active site. This comprehensive analysis enhances the understanding of quinazoline derivatives as potential anti-cancer agents and provides valuable insights for lead optimization in the early stages of drug discovery, particularly for developing highly potent anticancer therapeutics

Keywords: 3D-QSAR, CoMFA, CoMSIA, ADMET, molecular docking, quinazoline, molecular dynamic, egfr inhibitors, lung cancer, anticancer

Procedia PDF Downloads 25
11315 Modeling and Optimization of Algae Oil Extraction Using Response Surface Methodology

Authors: I. F. Ejim, F. L. Kamen

Abstract:

Aims: In this experiment, algae oil extraction with a combination of n-hexane and ethanol was investigated. The effects of extraction solvent concentration, extraction time and temperature on the yield and quality of oil were studied using Response Surface Methodology (RSM). Experimental Design: Optimization of algae oil extraction using Box-Behnken design was used to generate 17 experimental runs in a three-factor-three-level design where oil yield, specific gravity, acid value and saponification value were evaluated as the response. Result: In this result, a minimum oil yield of 17% and maximum of 44% was realized. The optimum values for yield, specific gravity, acid value and saponification value from the overlay plot were 40.79%, 0.8788, 0.5056 mg KOH/g and 180.78 mg KOH/g respectively with desirability of 0.801. The maximum point prediction was yield 40.79% at solvent concentration 66.68 n-hexane, temperature of 40.0°C and extraction time of 4 hrs. Analysis of Variance (ANOVA) results showed that the linear and quadratic coefficient were all significant at p<0.05. The experiment was validated and results obtained were with the predicted values. Conclusion: Algae oil extraction was successfully optimized using RSM and its quality indicated it is suitable for many industrial uses.

Keywords: algae oil, response surface methodology, optimization, Box-Bohnken, extraction

Procedia PDF Downloads 315
11314 Using Jumping Particle Swarm Optimization for Optimal Operation of Pump in Water Distribution Networks

Authors: R. Rajabpour, N. Talebbeydokhti, M. H. Ahmadi

Abstract:

Carefully scheduling the operations of pumps can be resulted to significant energy savings. Schedules can be defined either implicit, in terms of other elements of the network such as tank levels, or explicit by specifying the time during which each pump is on/off. In this study, two new explicit representations based on time-controlled triggers were analyzed, where the maximum number of pump switches was established beforehand, and the schedule may contain fewer switches than the maximum. The optimal operation of pumping stations was determined using a Jumping Particle Swarm Optimization (JPSO) algorithm to achieve the minimum energy cost. The model integrates JPSO optimizer and EPANET hydraulic network solver. The optimal pump operation schedule of VanZyl water distribution system was determined using the proposed model and compared with those from Genetic and Ant Colony algorithms. The results indicate that the proposed model utilizing the JPSP algorithm outperformed the others and is a versatile management model for the operation of real-world water distribution system.

Keywords: JPSO, operation, optimization, water distribution system

Procedia PDF Downloads 224
11313 Aerodynamic Optimum Nose Shape Change of High-Speed Train by Design Variable Variation

Authors: Minho Kwak, Suhwan Yun, Choonsoo Park

Abstract:

Nose shape optimizations of high-speed train are performed for the improvement of aerodynamic characteristics. Based on the commercial train, KTX-Sancheon, multi-objective optimizations are conducted for the improvement of the side wind stability and the micro-pressure wave following the optimization for the reduction of aerodynamic drag. 3D nose shapes are modelled by the Vehicle Modeling Function. Aerodynamic drag and side wind stability are calculated by three-dimensional compressible Navier-Stokes solver, and micro pressure wave is done by axi-symmetric compressible Navier-Stokes solver. The Maxi-min Latin Hypercube Sampling method is used to extract sampling points to construct the approximation model. The kriging model is constructed for the approximation model and the NSGA-II algorithm was used as the multi-objective optimization algorithm. Nose length, nose tip height, and lower surface curvature are design variables. Because nose length is a dominant variable for aerodynamic characteristics of train nose, two optimization processes are progressed respectively with and without the design variable, nose length. Each pareto set was obtained and each optimized nose shape is selected respectively considering Honam high-speed rail line infrastructure in South Korea. Through the optimization process with the nose length, when compared to KTX Sancheon, aerodynamic drag was reduced by 9.0%, side wind stability was improved by 4.5%, micro-pressure wave was reduced by 5.4% whereas aerodynamic drag by 7.3%, side wind stability by 3.9%, micro-pressure wave by 3.9%, without the nose length. As a result of comparison between two optimized shapes, similar shapes are extracted other than the effect of nose length.

Keywords: aerodynamic characteristics, design variable, multi-objective optimization, train nose shape

Procedia PDF Downloads 334
11312 Maresin Like 1 Treatment: Curbing the Pathogenesis of Behavioral Dysfunction and Neurodegeneration in Alzheimer's Disease Mouse Model

Authors: Yan Lu, Song Hong, Janakiraman Udaiyappan, Aarti Nagayach, Quoc-Viet A. Duong, Masao Morita, Shun Saito, Yuichi Kobayashi, Yuhai, Zhao, Hongying Peng, Nicholas B. Pham, Walter J Lukiw, Christopher A. Vuong, Nicolas G. Bazan

Abstract:

Aims: Neurodegeneration and behavior dysfunction occurs in patients with Alzheimer's Disease (AD), and as the disease progresses many patients develop cognitive impairment. 5XFAD mouse model of AD is widely used to study AD pathogenesis and treatment. This study aimed to investigate the effect of maresin like 1 (MaR-L1) treatment in AD pathology using 5XFAD mice. Methods: We tested 12-month-old male 5XFAD mice and wild type control mice treated with MaR-L1 in a battery of behavioral tasks. We performed open field test, beam walking test, clasping test, inverted grid test, acetone test, marble burring test, elevated plus maze test, cross maze test and novel object recognition test. We also studied neuronal loss, amyloid β burden, and inflammation in the brains of 5XFAD mice using immunohistology and Western blotting. Results: MaR-L1 treatment to the 5XFAD mice showed improved cognitive function of 5XFAD mice. MaR-L1 showed decreased anxiety behavior in open field test and marble burring test, increased muscular strength in the beam walking test, clasping test and inverted grid test. Cognitive function was improved in MaR-L1 treated 5XFAD mice in the novel object recognition test. MaR-L1 prevented neuronal loss and aberrant inflammation. Conclusion: Our finding suggests that behavioral abnormalities were normalized by the administration of MaR-L1 and the neuroprotective role of MaR-L1 in the AD. It also indicates that MaR-L1 treatment is able to prevent and or ameliorate neuronal loss and aberrant inflammation. Further experiments to validate the results are warranted using other AD models in the future.

Keywords: Alzheimer's disease, motor and cognitive behavior, 5XFAD mice, Maresin Like 1, microglial cell, astrocyte, neurodegeneration, inflammation, resolution of inflammation

Procedia PDF Downloads 154
11311 Seismic Response Mitigation of Structures Using Base Isolation System Considering Uncertain Parameters

Authors: Rama Debbarma

Abstract:

The present study deals with the performance of Linear base isolation system to mitigate seismic response of structures characterized by random system parameters. This involves optimization of the tuning ratio and damping properties of the base isolation system considering uncertain system parameters. However, the efficiency of base isolator may reduce if it is not tuned to the vibrating mode it is designed to suppress due to unavoidable presence of system parameters uncertainty. With the aid of matrix perturbation theory and first order Taylor series expansion, the total probability concept is used to evaluate the unconditional response of the primary structures considering random system parameters. For this, the conditional second order information of the response quantities are obtained in random vibration framework using state space formulation. Subsequently, the maximum unconditional root mean square displacement of the primary structures is used as the objective function to obtain optimum damping parameters Numerical study is performed to elucidate the effect of parameters uncertainties on the optimization of parameters of linear base isolator and system performance.

Keywords: linear base isolator, earthquake, optimization, uncertain parameters

Procedia PDF Downloads 407
11310 Bounded Solution Method for Geometric Programming Problem with Varying Parameters

Authors: Abdullah Ali H. Ahmadini, Firoz Ahmad, Intekhab Alam

Abstract:

Geometric programming problem (GPP) is a well-known non-linear optimization problem having a wide range of applications in many engineering problems. The structure of GPP is quite dynamic and easily fit to the various decision-making processes. The aim of this paper is to highlight the bounded solution method for GPP with special reference to variation among right-hand side parameters. Thus this paper is taken the advantage of two-level mathematical programming problems and determines the solution of the objective function in a specified interval called lower and upper bounds. The beauty of the proposed bounded solution method is that it does not require sensitivity analyses of the obtained optimal solution. The value of the objective function is directly calculated under varying parameters. To show the validity and applicability of the proposed method, a numerical example is presented. The system reliability optimization problem is also illustrated and found that the value of the objective function lies between the range of lower and upper bounds, respectively. At last, conclusions and future research are depicted based on the discussed work.

Keywords: varying parameters, geometric programming problem, bounded solution method, system reliability optimization

Procedia PDF Downloads 119
11309 Modeling and Optimizing of Sinker Electric Discharge Machine Process Parameters on AISI 4140 Alloy Steel by Central Composite Rotatable Design Method

Authors: J. Satya Eswari, J. Sekhar Babub, Meena Murmu, Govardhan Bhat

Abstract:

Electrical Discharge Machining (EDM) is an unconventional manufacturing process based on removal of material from a part by means of a series of repeated electrical sparks created by electric pulse generators at short intervals between a electrode tool and the part to be machined emmersed in dielectric fluid. In this paper, a study will be performed on the influence of the factors of peak current, pulse on time, interval time and power supply voltage. The output responses measured were material removal rate (MRR) and surface roughness. Finally, the parameters were optimized for maximum MRR with the desired surface roughness. RSM involves establishing mathematical relations between the design variables and the resulting responses and optimizing the process conditions. RSM is not free from problems when it is applied to multi-factor and multi-response situations. Design of experiments (DOE) technique to select the optimum machining conditions for machining AISI 4140 using EDM. The purpose of this paper is to determine the optimal factors of the electro-discharge machining (EDM) process investigate feasibility of design of experiment techniques. The work pieces used were rectangular plates of AISI 4140 grade steel alloy. The study of optimized settings of key machining factors like pulse on time, gap voltage, flushing pressure, input current and duty cycle on the material removal, surface roughness is been carried out using central composite design. The objective is to maximize the Material removal rate (MRR). Central composite design data is used to develop second order polynomial models with interaction terms. The insignificant coefficients’ are eliminated with these models by using student t test and F test for the goodness of fit. CCD is first used to establish the determine the optimal factors of the electro-discharge machining (EDM) for maximizing the MRR. The responses are further treated through a objective function to establish the same set of key machining factors to satisfy the optimization problem of the electro-discharge machining (EDM) process. The results demonstrate the better performance of CCD data based RSM for optimizing the electro-discharge machining (EDM) process.

Keywords: electric discharge machining (EDM), modeling, optimization, CCRD

Procedia PDF Downloads 326
11308 Numerical Simulation on Deformation Behaviour of Additively Manufactured AlSi10Mg Alloy

Authors: Racholsan Raj Nirmal, B. S. V. Patnaik, R. Jayaganthan

Abstract:

The deformation behaviour of additively manufactured AlSi10Mg alloy under low strains, high strain rates and elevated temperature conditions is essential to analyse and predict its response against dynamic loading such as impact and thermomechanical fatigue. The constitutive relation of Johnson-Cook is used to capture the strain rate sensitivity and thermal softening effect in AlSi10Mg alloy. Johnson-Cook failure model is widely used for exploring damage mechanics and predicting the fracture in many materials. In this present work, Johnson-Cook material and damage model parameters for additively manufactured AlSi10Mg alloy have been determined numerically from four types of uniaxial tensile test. Three different uniaxial tensile tests with dynamic strain rates (0.1, 1, 10, 50, and 100 s-1) and elevated temperature tensile test with three different temperature conditions (450 K, 500 K and 550 K) were performed on 3D printed AlSi10Mg alloy in ABAQUS/Explicit. Hexahedral elements are used to discretize tensile specimens and fracture energy value of 43.6 kN/m was used for damage initiation. Levenberg Marquardt optimization method was used for the evaluation of Johnson-Cook model parameters. It was observed that additively manufactured AlSi10Mg alloy has shown relatively higher strain rate sensitivity and lower thermal stability as compared to the other Al alloys.

Keywords: ABAQUS, additive manufacturing, AlSi10Mg, Johnson-Cook model

Procedia PDF Downloads 143
11307 Optimization of the Mechanical Performance of Fused Filament Fabrication Parts

Authors: Iván Rivet, Narges Dialami, Miguel Cervera, Michele Chiumenti

Abstract:

Process parameters in Additive Manufacturing (AM) play a critical role in the mechanical performance of the final component. In order to find the input configuration that guarantees the optimal performance of the printed part, the process-performance relationship must be found. Fused Filament Fabrication (FFF) is the selected demonstrative AM technology due to its great popularity in the industrial manufacturing world. A material model that considers the different printing patterns present in a FFF part is used. A voxelized mesh is built from the manufacturing toolpaths described in the G-Code file. An Adaptive Mesh Refinement (AMR) based on the octree strategy is used in order to reduce the complexity of the mesh while maintaining its accuracy. High-fidelity and cost-efficient Finite Element (FE) simulations are performed and the influence of key process parameters in the mechanical performance of the component is analyzed. A robust optimization process based on appropriate failure criteria is developed to find the printing direction that leads to the optimal mechanical performance of the component. The Tsai-Wu failure criterion is implemented due to the orthotropy and heterogeneity constitutive nature of FFF components and because of the differences between the strengths in tension and compression. The optimization loop implements a modified version of an Anomaly Detection (AD) algorithm and uses the computed metrics to obtain the optimal printing direction. The developed methodology is verified with a case study on an industrial demonstrator.

Keywords: additive manufacturing, optimization, printing direction, mechanical performance, voxelization

Procedia PDF Downloads 42
11306 Overall Stability of Welded Q460GJ Steel Box Columns: Experimental Study and Numerical Simulations

Authors: Zhou Xiong, Kang Shao Bo, Yang Bo

Abstract:

To date, high-performance structural steel has been widely used for columns in construction practices due to its significant advantages over conventional steel. However, the same design approach with conventional steel columns is still adopted in the design of high-performance steel columns. As a result, its superior properties cannot be fully considered in design. This paper conducts a test and finite element analysis on the overall stability behaviour of welded Q460GJ steel box columns. In the test, four steel columns with different slenderness and width-to-thickness ratio were compressed under an axial compression testing machine. And finite element models were established in which material nonlinearity and residual stress distributions of test columns were included. Then, comparisons were made between test results and finite element result, it showed that finite element analysis results are agree well with the test result. It means that the test and finite element model are reliable. Then, we compared the test result with the design value calculated by current code, the result showed that Q460GJ steel box columns have the higher overall buckling capacity than the design value. It is necessary to update the design curves for Q460GJ steel columns so that the overall stability capacity of Q460GJ box columns can be designed appropriately.

Keywords: axial compression, box columns, global buckling, numerical simulations, Q460GJ steel

Procedia PDF Downloads 388
11305 Optimization and Retrofitting for an Egyptian Refinery Water Network

Authors: Mohamed Mousa

Abstract:

Sacristies in the supply of freshwater, strict regulations on discharging wastewater and the support to encourage sustainable development by water minimization techniques leads to raise the interest of water reusing, regeneration, and recycling. Water is considered a vital element in chemical industries. In this study, an optimization model will be developed to determine the optimal design of refinery’s water network system via source interceptor sink that involves several network alternatives, then a Mixed-Integer Non-Linear programming (MINLP) was used to obtain the optimal network superstructure based on flowrates, the concentration of contaminants, etc. The main objective of the model is to reduce the fixed cost of piping installation interconnections, reducing the operating cots of all streams within the refiner’s water network, and minimize the concentration of pollutants to comply with the environmental regulations. A real case study for one of the Egyptian refineries was studied by GAMS / BARON global optimization platform, and the water network had been retrofitted and optimized, leading to saving around 195 m³/ hr. of freshwater with a total reduction reaches to 26 %.

Keywords: freshwater minimization, modelling, GAMS, BARON, water network design, wastewater reudction

Procedia PDF Downloads 207
11304 Analysis of a CO₂ Two-Phase Ejector Performances with Taguchi and Anova Optimization

Authors: Karima Megdouli

Abstract:

The ejector, a central element within the CO₂ transcritical ejection refrigeration system, holds significant importance in enhancing refrigeration capacity and minimizing compressor power usage. This study's objective is to introduce a technique for enhancing the effectiveness of the CO₂ transcritical two-phase ejector, utilizing Taguchi and ANOVA analysis. The investigation delves into the impact of geometric parameters, secondary flow temperature, and primary flow pressure on the efficiency of the ejector. Results indicate that employing a combination of Taguchi and ANOVA offers increased reliability and superior performance when optimizing the design of the CO₂ two-phase ejector.

Keywords: ejector, supersonic, Taguchi, ANOVA, optimization

Procedia PDF Downloads 59
11303 A Different Approach to Optimize Fuzzy Membership Functions with Extended FIR Filter

Authors: Jun-Ho Chung, Sung-Hyun Yoo, In-Hwan Choi, Hyun-Kook Lee, Moon-Kyu Song, Choon-Ki Ahn

Abstract:

The extended finite impulse response (EFIR) filter is addressed to optimize membership functions (MFs) of the fuzzy model that has strong nonlinearity. MFs are important parts of the fuzzy logic system (FLS) and, thus optimizing MFs of FLS is one of approaches to improve the performance of output. We employ the EFIR as an alternative optimization option to nonlinear fuzzy model. The performance of EFIR is demonstrated on a fuzzy cruise control via a numerical example.

Keywords: fuzzy logic system, optimization, membership function, extended FIR filter

Procedia PDF Downloads 703
11302 Data Clustering Algorithm Based on Multi-Objective Periodic Bacterial Foraging Optimization with Two Learning Archives

Authors: Chen Guo, Heng Tang, Ben Niu

Abstract:

Clustering splits objects into different groups based on similarity, making the objects have higher similarity in the same group and lower similarity in different groups. Thus, clustering can be treated as an optimization problem to maximize the intra-cluster similarity or inter-cluster dissimilarity. In real-world applications, the datasets often have some complex characteristics: sparse, overlap, high dimensionality, etc. When facing these datasets, simultaneously optimizing two or more objectives can obtain better clustering results than optimizing one objective. However, except for the objectives weighting methods, traditional clustering approaches have difficulty in solving multi-objective data clustering problems. Due to this, evolutionary multi-objective optimization algorithms are investigated by researchers to optimize multiple clustering objectives. In this paper, the Data Clustering algorithm based on Multi-objective Periodic Bacterial Foraging Optimization with two Learning Archives (DC-MPBFOLA) is proposed. Specifically, first, to reduce the high computing complexity of the original BFO, periodic BFO is employed as the basic algorithmic framework. Then transfer the periodic BFO into a multi-objective type. Second, two learning strategies are proposed based on the two learning archives to guide the bacterial swarm to move in a better direction. On the one hand, the global best is selected from the global learning archive according to the convergence index and diversity index. On the other hand, the personal best is selected from the personal learning archive according to the sum of weighted objectives. According to the aforementioned learning strategies, a chemotaxis operation is designed. Third, an elite learning strategy is designed to provide fresh power to the objects in two learning archives. When the objects in these two archives do not change for two consecutive times, randomly initializing one dimension of objects can prevent the proposed algorithm from falling into local optima. Fourth, to validate the performance of the proposed algorithm, DC-MPBFOLA is compared with four state-of-art evolutionary multi-objective optimization algorithms and one classical clustering algorithm on evaluation indexes of datasets. To further verify the effectiveness and feasibility of designed strategies in DC-MPBFOLA, variants of DC-MPBFOLA are also proposed. Experimental results demonstrate that DC-MPBFOLA outperforms its competitors regarding all evaluation indexes and clustering partitions. These results also indicate that the designed strategies positively influence the performance improvement of the original BFO.

Keywords: data clustering, multi-objective optimization, bacterial foraging optimization, learning archives

Procedia PDF Downloads 122
11301 A Geochemical Perspective on A-Type Granites of Khanak and Devsar Areas, Haryana, India: Implications for Petrogenesis

Authors: Naresh Kumar, Radhika Sharma, A. K. Singh

Abstract:

Granites from Khanak and Devsar areas, a part of Malani Igneous Suite (MIS) were investigated for their geochemical characteristics to understand the petrogenetic aspect of the research area. Neoproterozoic rocks of MIS are well exposed in Jhunjhunu, Jodhpur, Pali, Barmer, Jalor, Jaisalmer districts of Rajasthan and Bhiwani district of Haryana and also occur at Kirana hills of Pakistan. The MIS predominantly consists of acidic volcanic with acidic plutonic (granite of various types), mafic volcanic, mafic intrusive and minor amount of pyroclasts. Based on the field and petrographical studies, 28 samples were selected and analyzed for geochemical analysis of major, trace and rare earth elements at the Wadia Institute of Himalayan Geology, Dehradun by X-Ray Fluorescence Spectrometer (XRF) and ICP-MS (Inductively Coupled Plasma- Mass Spectrometry). Granites from the studied areas are categorized as grey, green and pink. Khanak granites consist of quartz, k-feldspar, plagioclase, and biotite as essential minerals and hematite, zircon, annite, monazite & rutile as accessory minerals. In Devsar granites, plagioclase is replaced by perthite and occurs as dominantly. Geochemically, granites from Khanak and Devsar areas exhibit typical A-type granites characteristics with their enrichment in SiO2, Na2O+K2O, Fe/Mg, Rb, Zr, Y, Th, U, REE (except Eu) and significant depletion in MgO, CaO, Sr, P, Ti, Ni, Cr, V and Eu suggested about A-type affinities in Northwestern Peninsular India. The amount of heat production (HP) in green and grey granites of Devsar area varies upto 9.68 & 11.70 μWm-3 and total heat generation unit (HGU) i.e. 23.04 & 27.86 respectively. Pink granites of Khanak area display a higher enrichment of HP (16.53 μWm-3) and HGU (39.37) than the granites from Devsar area. Overall, they have much higher values of HP and HGU than the average value of continental crust (3.8 HGU), which imply a possible linear relationship among the surface heat flow and crustal heat generation in the rocks of MIS. Chondrite-normalized REE patterns show enriched LREE, moderate to strong negative Eu anomalies and more or less flat heavy REE. In primitive mantle-normalized multi-element variation diagrams, the granites show pronounced depletions in the high-field-strength elements (HFSE) Nb, Zr, Sr, P, and Ti. Geochemical characteristics (major, trace and REE) along with the use of various discrimination schemes revealed their probable correspondence to magma derived from the crustal origin by a different degree of partial melting.

Keywords: A-type granite, neoproterozoic, Malani igneous suite, Khanak, Devsar

Procedia PDF Downloads 261
11300 Distributed System Computing Resource Scheduling Algorithm Based on Deep Reinforcement Learning

Authors: Yitao Lei, Xingxiang Zhai, Burra Venkata Durga Kumar

Abstract:

As the quantity and complexity of computing in large-scale software systems increase, distributed system computing becomes increasingly important. The distributed system realizes high-performance computing by collaboration between different computing resources. If there are no efficient resource scheduling resources, the abuse of distributed computing may cause resource waste and high costs. However, resource scheduling is usually an NP-hard problem, so we cannot find a general solution. However, some optimization algorithms exist like genetic algorithm, ant colony optimization, etc. The large scale of distributed systems makes this traditional optimization algorithm challenging to work with. Heuristic and machine learning algorithms are usually applied in this situation to ease the computing load. As a result, we do a review of traditional resource scheduling optimization algorithms and try to introduce a deep reinforcement learning method that utilizes the perceptual ability of neural networks and the decision-making ability of reinforcement learning. Using the machine learning method, we try to find important factors that influence the performance of distributed system computing and help the distributed system do an efficient computing resource scheduling. This paper surveys the application of deep reinforcement learning on distributed system computing resource scheduling proposes a deep reinforcement learning method that uses a recurrent neural network to optimize the resource scheduling, and proposes the challenges and improvement directions for DRL-based resource scheduling algorithms.

Keywords: resource scheduling, deep reinforcement learning, distributed system, artificial intelligence

Procedia PDF Downloads 90
11299 Modeling and Validation of Microspheres Generation in the Modified T-Junction Device

Authors: Lei Lei, Hongbo Zhang, Donald J. Bergstrom, Bing Zhang, K. Y. Song, W. J. Zhang

Abstract:

This paper presents a model for a modified T-junction device for microspheres generation. The numerical model is developed using a commercial software package: COMSOL Multiphysics. In order to test the accuracy of the numerical model, multiple variables, such as the flow rate of cross-flow, fluid properties, structure, and geometry of the microdevice are applied. The results from the model are compared with the experimental results in the diameter of the microsphere generated. The comparison shows a good agreement. Therefore the model is useful in further optimization of the device and feedback control of microsphere generation if any.

Keywords: CFD modeling, validation, microsphere generation, modified T-junction

Procedia PDF Downloads 684
11298 Design and Optimization of Flow Field for Cavitation Reduction of Valve Sleeves

Authors: Kamal Upadhyay, Zhou Hua, Yu Rui

Abstract:

This paper aims to improve the streamline linked with the flow field and cavitation on the valve sleeve. We observed that local pressure fluctuation produces a low-pressure zone, central to the formation of vapor volume fraction within the valve chamber led to air-bubbles (or cavities). Thus, it allows simultaneously to a severe negative impact on the inner surface and lifespan of the valve sleeves. Cavitation reduction is a vitally important issue to pressure control valves. The optimization of the flow field is proposed in this paper to reduce the cavitation of valve sleeves. In this method, the inner wall of the valve sleeve is changed from a cylindrical surface to the conical surface, leading to the decline of the fluid flow velocity and the rise of the outlet pressure. Besides, the streamline is distributed inside the sleeve uniformly. Thus, the bubble generation is lessened. The fluid models are built and analysis of flow field distribution, pressure, vapor volume and velocity was carried out using computational fluid dynamics (CFD) and numerical technique. The results indicate that this structure can suppress the cavitation of valve sleeves effectively.

Keywords: streamline, cavitation, optimization, computational fluid dynamics

Procedia PDF Downloads 125
11297 LanE-change Path Planning of Autonomous Driving Using Model-Based Optimization, Deep Reinforcement Learning and 5G Vehicle-to-Vehicle Communications

Authors: William Li

Abstract:

Lane-change path planning is a crucial and yet complex task in autonomous driving. The traditional path planning approach based on a system of carefully-crafted rules to cover various driving scenarios becomes unwieldy as more and more rules are added to deal with exceptions and corner cases. This paper proposes to divide the entire path planning to two stages. In the first stage the ego vehicle travels longitudinally in the source lane to reach a safe state. In the second stage the ego vehicle makes lateral lane-change maneuver to the target lane. The paper derives the safe state conditions based on lateral lane-change maneuver calculation to ensure collision free in the second stage. To determine the acceleration sequence that minimizes the time to reach a safe state in the first stage, the paper proposes three schemes, namely, kinetic model based optimization, deep reinforcement learning, and 5G vehicle-to-vehicle (V2V) communications. The paper investigates these schemes via simulation. The model-based optimization is sensitive to the model assumptions. The deep reinforcement learning is more flexible in handling scenarios beyond the model assumed by the optimization. The 5G V2V eliminates uncertainty in predicting future behaviors of surrounding vehicles by sharing driving intents and enabling cooperative driving.

Keywords: lane change, path planning, autonomous driving, deep reinforcement learning, 5G, V2V communications, connected vehicles

Procedia PDF Downloads 201
11296 Investigating the Properties of Asphalt and Asphalt Mixture Based on the Effect of Waste Toner

Authors: Prince Igor Itoua, Daquan Sun, Shihui Shen

Abstract:

This study aimed at investigating the properties of asphalt and mix asphalt based on the effects of waste toner sources (WT1 and WT2) with 8% dosage waste toner powders (WT). The test results included penetration, softening points, ductility, G*sinδ, G*/sinδ, Ideal cracking test(IDEAL-CT), and Ideal shear rutting test(IDEAL-RT). The results showed that the base binder with WT2 had a significantly higher viscosity value compared to the WT1 modified binder, and thus, higher energy for mixing and compaction is needed. Fur-thermore, the results of penetration, softening points, G*sinδ, and G*/sinδ were all affected by waste toner type. In terms of asphalt mixture, the IDEAL-RT test revealed that the addition of waste toner improved the rutting resistance of the asphalt mixture regardless of toner type. Further, CTindex values for waste toner-modified asphalt mixtures show no significant difference. Above all, WT-modified asphalt mixtures produced by the wet process have better rutting performance.

Keywords: waste toner, waste toner modified asphalt, asphalt mixture properties, IDEAL-RT test, IDEAL-CT test

Procedia PDF Downloads 64
11295 Analysis and Optimized Design of a Packaged Liquid Chiller

Authors: Saeed Farivar, Mohsen Kahrom

Abstract:

The purpose of this work is to develop a physical simulation model for the purpose of studying the effect of various design parameters on the performance of packaged-liquid chillers. This paper presents a steady-state model for predicting the performance of package-Liquid chiller over a wide range of operation condition. The model inputs are inlet conditions; geometry and output of model include system performance variable such as power consumption, coefficient of performance (COP) and states of refrigerant through the refrigeration cycle. A computer model that simulates the steady-state cyclic performance of a vapor compression chiller is developed for the purpose of performing detailed physical design analysis of actual industrial chillers. The model can be used for optimizing design and for detailed energy efficiency analysis of packaged liquid chillers. The simulation model takes into account presence of all chiller components such as compressor, shell-and-tube condenser and evaporator heat exchangers, thermostatic expansion valve and connection pipes and tubing’s by thermo-hydraulic modeling of heat transfer, fluids flow and thermodynamics processes in each one of the mentioned components. To verify the validity of the developed model, a 7.5 USRT packaged-liquid chiller is used and a laboratory test stand for bringing the chiller to its standard steady-state performance condition is build. Experimental results obtained from testing the chiller in various load and temperature conditions is shown to be in good agreement with those obtained from simulating the performance of the chiller using the computer prediction model. An entropy-minimization-based optimization analysis is performed based on the developed analytical performance model of the chiller. The variation of design parameters in construction of shell-and-tube condenser and evaporator heat exchangers are studied using the developed performance and optimization analysis and simulation model and a best-match condition between the physical design and construction of chiller heat exchangers and its compressor is found to exist. It is expected that manufacturers of chillers and research organizations interested in developing energy-efficient design and analysis of compression chillers can take advantage of the presented study and its results.

Keywords: optimization, packaged liquid chiller, performance, simulation

Procedia PDF Downloads 260