Search results for: resilience optimization model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19702

Search results for: resilience optimization model

18352 An Integrated Approach for Optimizing Drillable Parameters to Increase Drilling Performance: A Real Field Case Study

Authors: Hamidoddin Yousife

Abstract:

Drilling optimization requires a prediction of drilling rate of penetration (ROP) since it provides a significant reduction in drilling costs. There are several factors that can have an impact on the ROP, both controllable and uncontrollable. Numerous drilling penetration rate models have been considered based on drilling parameters. This papers considered the effect of proper drilling parameter selection such as bit, Mud Type, applied weight on bit (WOB), Revolution per minutes (RPM), and flow rate on drilling optimization and drilling cost reduction. A predicted analysis is used in real-time drilling performance to determine the optimal drilling operation. As a result of these modeling studies, the real data collected from three directional wells at Azadegan oil fields, Iran, was verified and adjusted to determine the drillability of a specific formation. Simulation results and actual drilling results show significant improvements in inaccuracy. Once simulations had been validated, optimum drilling parameters and equipment specifications were determined by varying weight on bit (WOB), rotary speed (RPM), hydraulics (hydraulic pressure), and bit specification for each well until the highest drilling rate was achieved. To evaluate the potential operational and economic benefits of optimizing results, a qualitative and quantitative analysis of the data was performed.

Keywords: drlling, cost, optimization, parameters

Procedia PDF Downloads 170
18351 Developing Medium Term Maintenance Plan For Road Networks

Authors: Helen S. Ghali, Haidy S. Ghali, Salma Ibrahim, Ossama Hosny, Hatem S. Elbehairy

Abstract:

Infrastructure systems are essential assets in any community; accordingly, authorities aim to maximize its life span while minimizing the life cycle cost. This requires studying the asset conditions throughout its operation and forming a cost-efficient maintenance strategy plan. The objective of this study is to develop a highway management system that provides medium-term maintenance plans with the minimum life cycle cost subject to budget constraints. The model is applied to data collected for the highway network in India with the aim to output a 5-year maintenance plan strategy from 2019 till 2023. The main element considered is the surface coarse, either rigid or flexible pavement. The model outputs a 5-year maintenance plan for each segment given the budget constraint while maximizing the new pavement condition rating and minimizing its life cycle cost.

Keywords: infrastructure, asset management, optimization, maintenance plan

Procedia PDF Downloads 219
18350 A Theoretical Hypothesis on Ferris Wheel Model of University Social Responsibility

Authors: Le Kang

Abstract:

According to the nature of the university, as a free and responsible academic community, USR is based on a different foundation —academic responsibility, so the Pyramid and the IC Model of CSR could not fully explain the most distinguished feature of USR. This paper sought to put forward a new model— Ferris Wheel Model, to illustrate the nature of USR and the process of achievement. The Ferris Wheel Model of USR shows the university creates a balanced, fairness and neutrality systemic structure to afford social responsibilities; that makes the organization could obtain a synergistic effect to achieve more extensive interests of stakeholders and wider social responsibilities.

Keywords: USR, achievement model, ferris wheel model, social responsibilities

Procedia PDF Downloads 725
18349 Agent-Based Simulation for Supply Chain Transport Corridors

Authors: Kamalendu Pal

Abstract:

Supply chains are the spinal cord of trade and commerce. Their logistics use different transport corridors on regular basis for operational purpose. The international supply chain transport corridors include different infrastructure elements (e.g. weighbridge, package handling equipment, border clearance authorities, and so on) in supply chains. This paper presents the use of multi-agent systems (MAS) to model and simulate some aspects of transportation corridors, and in particular the area of weighbridge resource optimization for operational profit generation purpose. An underlying multi-agent model provides a means of modeling the relationships among stakeholders in order to enable coordination in a transport corridor environment. Simulations of the costs of container unloading, reloading, and waiting time for queuing up tracks have been carried out using data sets. Results of the simulation provide the potential guidance in making decisions about optimal service resource allocation in a trade corridor.

Keywords: multi-agent systems, simulation, supply chain, transport corridor, weighbridge

Procedia PDF Downloads 354
18348 A New Perspective: The Use of Low-Cost Phase Change Material in Building Envelope System

Authors: Andrey A. Chernousov, Ben Y. B. Chan

Abstract:

The use of the low-cost paraffinic phase change material can be rather effective in smart building envelopes in the South China region. Particular attention has to be paid to the PCM optimization as an exploitation conditions and the envelope insulation changes its thermal characteristics. The studied smart building envelope consists of a reinforced aluminum exterior, polymeric insulation foam, phase change material and reinforced interior gypsum board. A prototype sample was tested to validate the numerical scheme using EnergryPlus software. Three scenarios of insulation thermal resistance loss (ΔR/R = 0%, 25%, 50%) were compared with the different PCM thicknesses (tP=0, 1, 2.5, 5 mm). The comparisons were carried out for a west facing enveloped office building (50 storey). PCM optimization was applied to find the maximum efficiency for the different ΔR/R cases. It was found, during the optimization, that the PCM is an important smart component, lowering the peak energy demand up to 2.7 times. The results are not influenced by the insulation aging in terms of ΔR/R during long-term exploitation. In hot and humid climates like Hong Kong, the insulation core of the smart systems is recommended to be laminated completely. This can be very helpful in achieving an acceptable payback period.

Keywords: smart building envelope, thermal performance, phase change material, energy efficiency, large-scale sandwich panel

Procedia PDF Downloads 733
18347 Minimum-Fuel Optimal Trajectory for Reusable First-Stage Rocket Landing Using Particle Swarm Optimization

Authors: Kevin Spencer G. Anglim, Zhenyu Zhang, Qingbin Gao

Abstract:

Reusable launch vehicles (RLVs) present a more environmentally-friendly approach to accessing space when compared to traditional launch vehicles that are discarded after each flight. This paper studies the recyclable nature of RLVs by presenting a solution method for determining minimum-fuel optimal trajectories using principles from optimal control theory and particle swarm optimization (PSO). This problem is formulated as a minimum-landing error powered descent problem where it is desired to move the RLV from a fixed set of initial conditions to three different sets of terminal conditions. However, unlike other powered descent studies, this paper considers the highly nonlinear effects caused by atmospheric drag, which are often ignored for studies on the Moon or on Mars. Rather than optimizing the controls directly, the throttle control is assumed to be bang-off-bang with a predetermined thrust direction for each phase of flight. The PSO method is verified in a one-dimensional comparison study, and it is then applied to the two-dimensional cases, the results of which are illustrated.

Keywords: minimum-fuel optimal trajectory, particle swarm optimization, reusable rocket, SpaceX

Procedia PDF Downloads 278
18346 Research on Pilot Sequence Design Method of Multiple Input Multiple Output Orthogonal Frequency Division Multiplexing System Based on High Power Joint Criterion

Authors: Linyu Wang, Jiahui Ma, Jianhong Xiang, Hanyu Jiang

Abstract:

For the pilot design of the sparse channel estimation model in Multiple Input Multiple Output Orthogonal Frequency Division Multiplexing (MIMO-OFDM) systems, the observation matrix constructed according to the matrix cross-correlation criterion, total correlation criterion and other optimization criteria are not optimal, resulting in inaccurate channel estimation and high bit error rate at the receiver. This paper proposes a pilot design method combining high-power sum and high-power variance criteria, which can more accurately estimate the channel. First, the pilot insertion position is designed according to the high-power variance criterion under the condition of equal power. Then, according to the high power sum criterion, the pilot power allocation is converted into a cone programming problem, and the power allocation is carried out. Finally, the optimal pilot is determined by calculating the weighted sum of the high power sum and the high power variance. Compared with the traditional pilot frequency, under the same conditions, the constructed MIMO-OFDM system uses the optimal pilot frequency for channel estimation, and the communication bit error rate performance obtains a gain of 6~7dB.

Keywords: MIMO-OFDM, pilot optimization, compressed sensing, channel estimation

Procedia PDF Downloads 149
18345 Energy Efficient Retrofitting and Optimization of Dual Mixed Refrigerant Natural Gas Liquefaction Process

Authors: Muhammad Abdul Qyyum, Kinza Qadeer, Moonyong Lee

Abstract:

Globally, liquefied natural gas (LNG) has drawn interest as a green energy source in comparison with other fossil fuels, mainly because of its ease of transport and low carbon dioxide emissions. It is expected that demand for LNG will grow steadily over the next few decades. In addition, because the demand for clean energy is increasing, LNG production facilities are expanding into new natural gas reserves across the globe. However, LNG production is an energy and cost intensive process because of the huge power requirements for compression and refrigeration. Therefore, one of the major challenges in the LNG industry is to improve the energy efficiency of existing LNG processes through economic and ecological strategies. The advancement in expansion devices such as two-phase cryogenic expander (TPE) and cryogenic hydraulic turbine (HT) were exploited for energy and cost benefits in natural gas liquefaction. Retrofitting the conventional Joule–Thompson (JT) valve with TPE and HT have the potential to improve the energy efficiency of LNG processes. This research investigated the potential feasibility of the retrofitting of a dual mixed refrigerant (DMR) process by replacing the isenthalpic expansion with isentropic expansion corresponding to energy efficient LNG production. To fully take the potential benefit of the proposed process retrofitting, the proposed DMR schemes were optimized by using a Coggins optimization approach, which was implemented in Microsoft Visual Studio (MVS) environment and linked to the rigorous HYSYS® model. The results showed that the required energy of the proposed isentropic expansion based DMR process could be saved up to 26.5% in comparison with the conventional isenthalpic based DMR process using the JT valves. Utilization of the recovered energy into boosting the natural gas feed pressure could further improve the energy efficiency of the LNG process up to 34% as compared to the base case. This work will help the process engineers to overcome the challenges relating to energy efficiency and safety concerns of LNG processes. Furthermore, the proposed retrofitting scheme can also be implemented to improve the energy efficiency of other isenthalpic expansion based energy intensive cryogenic processes.

Keywords: cryogenic liquid turbine, Coggins optimization, dual mixed refrigerant, energy efficient LNG process, two-phase expander

Procedia PDF Downloads 147
18344 Optimum Design of Dual-Purpose Outriggers in Tall Buildings

Authors: Jiwon Park, Jihae Hur, Kukjae Kim, Hansoo Kim

Abstract:

In this study, outriggers, which are horizontal structures connecting a building core to distant columns to increase the lateral stiffness of a tall building, are used to reduce differential axial shortening in a tall building. Therefore, the outriggers in tall buildings are used to serve the dual purposes of reducing the lateral displacement and reducing the differential axial shortening. Since the location of the outrigger greatly affects the effectiveness of the outrigger in terms of the lateral displacement at the top of the tall building and the maximum differential axial shortening, the optimum locations of the dual-purpose outriggers can be determined by an optimization method. Because the floors where the outriggers are installed are given as integer numbers, the conventional gradient-based optimization methods cannot be directly used. In this study, a piecewise quadratic interpolation method is used to resolve the integrality requirement posed by the optimum locations of the dual-purpose outriggers. The optimal solutions for the dual-purpose outriggers are searched by linear scalarization which is a popular method for multi-objective optimization problems. It was found that increasing the number of outriggers reduced the maximum lateral displacement and the maximum differential axial shortening. It was also noted that the optimum locations for reducing the lateral displacement and reducing the differential axial shortening were different. Acknowledgment: This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Science and ICT (NRF-2017R1A2B4010043) and financially supported by Korea Ministry of Land, Infrastructure and Transport(MOLIT) as U-City Master and Doctor Course Grant Program.

Keywords: concrete structure, optimization, outrigger, tall building

Procedia PDF Downloads 177
18343 Reliability Enhancement by Parameter Design in Ferrite Magnet Process

Authors: Won Jung, Wan Emri

Abstract:

Ferrite magnet is widely used in many automotive components such as motors and alternators. Magnets used inside the components must be in good quality to ensure the high level of performance. The purpose of this study is to design input parameters that optimize the ferrite magnet production process to ensure the quality and reliability of manufactured products. Design of Experiments (DOE) and Statistical Process Control (SPC) are used as mutual supplementations to optimize the process. DOE and SPC are quality tools being used in the industry to monitor and improve the manufacturing process condition. These tools are practically used to maintain the process on target and within the limits of natural variation. A mixed Taguchi method is utilized for optimization purpose as a part of DOE analysis. SPC with proportion data is applied to assess the output parameters to determine the optimal operating conditions. An example of case involving the monitoring and optimization of ferrite magnet process was presented to demonstrate the effectiveness of this approach. Through the utilization of these tools, reliable magnets can be produced by following the step by step procedures of proposed framework. One of the main contributions of this study was producing the crack free magnets by applying the proposed parameter design.

Keywords: ferrite magnet, crack, reliability, process optimization, Taguchi method

Procedia PDF Downloads 518
18342 A Tutorial on Model Predictive Control for Spacecraft Maneuvering Problem with Theory, Experimentation and Applications

Authors: O. B. Iskender, K. V. Ling, V. Dubanchet, L. Simonini

Abstract:

This paper discusses the recent advances and future prospects of spacecraft position and attitude control using Model Predictive Control (MPC). First, the challenges of the space missions are summarized, in particular, taking into account the errors, uncertainties, and constraints imposed by the mission, spacecraft and, onboard processing capabilities. The summary of space mission errors and uncertainties provided in categories; initial condition errors, unmodeled disturbances, sensor, and actuator errors. These previous constraints are classified into two categories: physical and geometric constraints. Last, real-time implementation capability is discussed regarding the required computation time and the impact of sensor and actuator errors based on the Hardware-In-The-Loop (HIL) experiments. The rationales behind the scenarios’ are also presented in the scope of space applications as formation flying, attitude control, rendezvous and docking, rover steering, and precision landing. The objectives of these missions are explained, and the generic constrained MPC problem formulations are summarized. Three key design elements used in MPC design: the prediction model, the constraints formulation and the objective cost function are discussed. The prediction models can be linear time invariant or time varying depending on the geometry of the orbit, whether it is circular or elliptic. The constraints can be given as linear inequalities for input or output constraints, which can be written in the same form. Moreover, the recent convexification techniques for the non-convex geometrical constraints (i.e., plume impingement, Field-of-View (FOV)) are presented in detail. Next, different objectives are provided in a mathematical framework and explained accordingly. Thirdly, because MPC implementation relies on finding in real-time the solution to constrained optimization problems, computational aspects are also examined. In particular, high-speed implementation capabilities and HIL challenges are presented towards representative space avionics. This covers an analysis of future space processors as well as the requirements of sensors and actuators on the HIL experiments outputs. The HIL tests are investigated for kinematic and dynamic tests where robotic arms and floating robots are used respectively. Eventually, the proposed algorithms and experimental setups are introduced and compared with the authors' previous work and future plans. The paper concludes with a conjecture that MPC paradigm is a promising framework at the crossroads of space applications while could be further advanced based on the challenges mentioned throughout the paper and the unaddressed gap.

Keywords: convex optimization, model predictive control, rendezvous and docking, spacecraft autonomy

Procedia PDF Downloads 111
18341 Model Predictive Control of Three Phase Inverter for PV Systems

Authors: Irtaza M. Syed, Kaamran Raahemifar

Abstract:

This paper presents a model predictive control (MPC) of a utility interactive three phase inverter (TPI) for a photovoltaic (PV) system at commercial level. The proposed model uses phase locked loop (PLL) to synchronize TPI with the power electric grid (PEG) and performs MPC control in a dq reference frame. TPI model consists of boost converter (BC), maximum power point tracking (MPPT) control, and a three leg voltage source inverter (VSI). Operational model of VSI is used to synthesize sinusoidal current and track the reference. Model is validated using a 35.7 kW PV system in Matlab/Simulink. Implementation and results show simplicity and accuracy, as well as reliability of the model.

Keywords: model predictive control, three phase voltage source inverter, PV system, Matlab/simulink

Procedia PDF Downloads 596
18340 Enhancing Project Management Performance in Prefabricated Building Construction under Uncertainty: A Comprehensive Approach

Authors: Niyongabo Elyse

Abstract:

Prefabricated building construction is a pioneering approach that combines design, production, and assembly to attain energy efficiency, environmental sustainability, and economic feasibility. Despite continuous development in the industry in China, the low technical maturity of standardized design, factory production, and construction assembly introduces uncertainties affecting prefabricated component production and on-site assembly processes. This research focuses on enhancing project management performance under uncertainty to help enterprises navigate these challenges and optimize project resources. The study introduces a perspective on how uncertain factors influence the implementation of prefabricated building construction projects. It proposes a theoretical model considering project process management ability, adaptability to uncertain environments, and collaboration ability of project participants. The impact of uncertain factors is demonstrated through case studies and quantitative analysis, revealing constraints on implementation time, cost, quality, and safety. To address uncertainties in prefabricated component production scheduling, a fuzzy model is presented, expressing processing times in interval values. The model utilizes a cooperative co-evolution evolution algorithm (CCEA) to optimize scheduling, demonstrated through a real case study showcasing reduced project duration and minimized effects of processing time disturbances. Additionally, the research addresses on-site assembly construction scheduling, considering the relationship between task processing times and assigned resources. A multi-objective model with fuzzy activity durations is proposed, employing a hybrid cooperative co-evolution evolution algorithm (HCCEA) to optimize project scheduling. Results from real case studies indicate improved project performance in terms of duration, cost, and resilience to processing time delays and resource changes. The study also introduces a multistage dynamic process control model, utilizing IoT technology for real-time monitoring during component production and construction assembly. This approach dynamically adjusts schedules when constraints arise, leading to enhanced project management performance, as demonstrated in a real prefabricated housing project. Key contributions include a fuzzy prefabricated components production scheduling model, a multi-objective multi-mode resource-constrained construction project scheduling model with fuzzy activity durations, a multi-stage dynamic process control model, and a cooperative co-evolution evolution algorithm. The integrated mathematical model addresses the complexity of prefabricated building construction project management, providing a theoretical foundation for practical decision-making in the field.

Keywords: prefabricated construction, project management performance, uncertainty, fuzzy scheduling

Procedia PDF Downloads 51
18339 Optimal Design of 3-Way Reversing Valve Considering Cavitation Effect

Authors: Myeong-Gon Lee, Yang-Gyun Kim, Tae-Young Kim, Seung-Ho Han

Abstract:

The high-pressure valve uses one set of 2-way valves for the purpose of reversing fluid direction. If there is no accurate control device for the 2-way valves, lots of surging can be generated. The surging is a kind of pressure ripple that occurs in rapid changes of fluid motions under inaccurate valve control. To reduce the surging effect, a 3-way reversing valve can be applied which provides a rapid and precise change of water flow directions without any accurate valve control system. However, a cavitation occurs due to a complicated internal trim shape of the 3-way reversing valve. The cavitation causes not only noise and vibration but also decreasing the efficiency of valve-operation, in which the bubbles generated below the saturated vapor pressure are collapsed rapidly at higher pressure zone. The shape optimization of the 3-way reversing valve to minimize the cavitation effect is necessary. In this study, the cavitation index according to the international standard ISA was introduced to estimate macroscopically the occurrence of the cavitation effect. Computational fluid dynamic analysis was carried out, and the cavitation effect was quantified by means of the percent of cavitation converted from calculated results of vapor volume fraction. In addition, the shape optimization of the 3-way reversing valve was performed by taking into account of the percent of cavitation.

Keywords: 3-Way reversing valve, cavitation, shape optimization, vapor volume fraction

Procedia PDF Downloads 372
18338 Bi-Objective Optimization for Sustainable Supply Chain Network Design in Omnichannel

Authors: Veerpaul Maan, Gaurav Mishra

Abstract:

The evolution of omnichannel has revolutionized the supply chain of the organizations by enhancing customer shopping experience. For these organizations need to develop well-integrated multiple distribution channels to leverage the benefits of omnichannel. To adopt an omnichannel system in the supply chain has resulted in structuring and reconfiguring the practices of the traditional supply chain distribution network. In this paper a multiple distribution supply chain network (MDSCN) have been proposed which integrates online giants with a local retailers distribution network in uncertain environment followed by sustainability. To incorporate sustainability, an additional objective function is added to reduce the carbon content through minimizing the travel distance of the product. Through this proposed model, customers are free to access product and services as per their choice of channels which increases their convenience, reach and satisfaction. Further, a numerical illustration is being shown along with interpretation of results to validate the proposed model.

Keywords: sustainable supply chain network, omnichannel, multiple distribution supply chain network, integrate multiple distribution channels

Procedia PDF Downloads 223
18337 Model-Based Process Development for the Comparison of a Radial Riveting and Roller Burnishing Process in Mechanical Joining Technology

Authors: Tobias Beyer, Christoph Friedrich

Abstract:

Modern simulation methodology using finite element models is nowadays a recognized tool for product design/optimization. Likewise, manufacturing process design is increasingly becoming the focus of simulation methodology in order to enable sustainable results based on reduced real-life tests here as well. In this article, two process simulations -radial riveting and roller burnishing- used for mechanical joining of components are explained. In the first step, the required boundary conditions are developed and implemented in the respective simulation models. This is followed by process space validation. With the help of the validated models, the interdependencies of the input parameters are investigated and evaluated by means of sensitivity analyses. Limit case investigations are carried out and evaluated with the aid of the process simulations. Likewise, a comparison of the two joining methods to each other becomes possible.

Keywords: FEM, model-based process development, process simulation, radial riveting, roller burnishing, sensitivity analysis

Procedia PDF Downloads 112
18336 The Possibility of Solving a 3x3 Rubik’s Cube under 3 Seconds

Authors: Chung To Kong, Siu Ming Yiu

Abstract:

Rubik's cube was invented in 1974. Since then, speedcubers all over the world try their best to break the world record again and again. The newest record is 3.47 seconds. There are many factors that affect the timing, including turns per second (tps), algorithm, finger trick, hardware of the cube. In this paper, the lower bound of the cube solving time will be discussed using convex optimization. Extended analysis of the world records will be used to understand how to improve the timing. With the understanding of each part of the solving step, the paper suggests a list of speed improvement techniques. Based on the analysis of the world record, there is a high possibility that the 3 seconds mark will be broken soon.

Keywords: Rubik's Cube, speed, finger trick, optimization

Procedia PDF Downloads 208
18335 Proxisch: An Optimization Approach of Large-Scale Unstable Proxy Servers Scheduling

Authors: Xiaoming Jiang, Jinqiao Shi, Qingfeng Tan, Wentao Zhang, Xuebin Wang, Muqian Chen

Abstract:

Nowadays, big companies such as Google, Microsoft, which have adequate proxy servers, have perfectly implemented their web crawlers for a certain website in parallel. But due to lack of expensive proxy servers, it is still a puzzle for researchers to crawl large amounts of information from a single website in parallel. In this case, it is a good choice for researchers to use free public proxy servers which are crawled from the Internet. In order to improve efficiency of web crawler, the following two issues should be considered primarily: (1) Tasks may fail owing to the instability of free proxy servers; (2) A proxy server will be blocked if it visits a single website frequently. In this paper, we propose Proxisch, an optimization approach of large-scale unstable proxy servers scheduling, which allow anyone with extremely low cost to run a web crawler efficiently. Proxisch is designed to work efficiently by making maximum use of reliable proxy servers. To solve second problem, it establishes a frequency control mechanism which can ensure the visiting frequency of any chosen proxy server below the website’s limit. The results show that our approach performs better than the other scheduling algorithms.

Keywords: proxy server, priority queue, optimization algorithm, distributed web crawling

Procedia PDF Downloads 213
18334 Effect of Progressive Type-I Right Censoring on Bayesian Statistical Inference of Simple Step–Stress Acceleration Life Testing Plan under Weibull Life Distribution

Authors: Saleem Z. Ramadan

Abstract:

This paper discusses the effects of using progressive Type-I right censoring on the design of the Simple Step Accelerated Life testing using Bayesian approach for Weibull life products under the assumption of cumulative exposure model. The optimization criterion used in this paper is to minimize the expected pre-posterior variance of the PTH percentile time of failures. The model variables are the stress changing time and the stress value for the first step. A comparison between the conventional and the progressive Type-I right censoring is provided. The results have shown that the progressive Type-I right censoring reduces the cost of testing on the expense of the test precision when the sample size is small. Moreover, the results have shown that using strong priors or large sample size reduces the sensitivity of the test precision to the censoring proportion. Hence, the progressive Type-I right censoring is recommended in these cases as progressive Type-I right censoring reduces the cost of the test and doesn't affect the precision of the test a lot. Moreover, the results have shown that using direct or indirect priors affects the precision of the test.

Keywords: reliability, accelerated life testing, cumulative exposure model, Bayesian estimation, progressive type-I censoring, Weibull distribution

Procedia PDF Downloads 507
18333 Model Observability – A Monitoring Solution for Machine Learning Models

Authors: Amreth Chandrasehar

Abstract:

Machine Learning (ML) Models are developed and run in production to solve various use cases that help organizations to be more efficient and help drive the business. But this comes at a massive development cost and lost business opportunities. According to the Gartner report, 85% of data science projects fail, and one of the factors impacting this is not paying attention to Model Observability. Model Observability helps the developers and operators to pinpoint the model performance issues data drift and help identify root cause of issues. This paper focuses on providing insights into incorporating model observability in model development and operationalizing it in production.

Keywords: model observability, monitoring, drift detection, ML observability platform

Procedia PDF Downloads 112
18332 Optimization of Wear during Dry Sliding Wear of AISI 1042 Steel Using Response Surface Methodology

Authors: Sukant Mehra, Parth Gupta, Varun Arora, Sarvoday Singh, Amit Kohli

Abstract:

The study was emphasised on dry sliding wear behavior of AISI 1042 steel. Dry sliding wear tests were performed using pin-on-disk apparatus under normal loads of 5, 7.5 and 10 kgf and at speeds 600, 750 and 900 rpm. Response surface methodology (RSM) was utilized for finding optimal values of process parameter and experiment was based on rotatable, central composite design (CCD). It was found that the wear followed linear pattern with the load and rpm. The obtained optimal process parameters have been predicted and verified by confirmation experiments.

Keywords: central composite design (CCD), optimization, response surface methodology (RSM), wear

Procedia PDF Downloads 577
18331 Multi Objective Optimization for Two-Sided Assembly Line Balancing

Authors: Srushti Bhatt, M. B. Kiran

Abstract:

Two-sided assembly line balancing problem is yet to be addressed simply to compete for the global market for manufacturers. The task assigned in an ordered sequence to get optimum performance of the system is known as assembly line balancing problem mainly classified as single and two sided. It is very challenging in manufacturing industries to balance two-sided assembly line, wherein the set of sequential workstations the task operations are performed in two sides of the line. The conflicting major objective in two-sided assembly line balancing problem is either to maximize /minimize the performance parameters. The present study emphases on combining different evolutionary algorithm; ant colony, Tabu search and petri net method; and compares their results of an algorithm for solving two-sided assembly line balancing problem. The concept of multi objective optimization of performance parameters is now a day adopted to make a decision involving more than one objective function to be simultaneously optimized. The optimum result can be expected among the selected methods using multi-objective optimization. The performance parameters considered in the present study are a number of workstation, slickness and smoothness index. The simulation of the assembly line balancing problem provides optimal results of classical and practical problems.

Keywords: Ant colony, petri net, tabu search, two sided ALBP

Procedia PDF Downloads 278
18330 All-or-None Principle and Weakness of Hodgkin-Huxley Mathematical Model

Authors: S. A. Sadegh Zadeh, C. Kambhampati

Abstract:

Mathematical and computational modellings are the necessary tools for reviewing, analysing, and predicting processes and events in the wide spectrum range of scientific fields. Therefore, in a field as rapidly developing as neuroscience, the combination of these two modellings can have a significant role in helping to guide the direction the field takes. The paper combined mathematical and computational modelling to prove a weakness in a very precious model in neuroscience. This paper is intended to analyse all-or-none principle in Hodgkin-Huxley mathematical model. By implementation the computational model of Hodgkin-Huxley model and applying the concept of all-or-none principle, an investigation on this mathematical model has been performed. The results clearly showed that the mathematical model of Hodgkin-Huxley does not observe this fundamental law in neurophysiology to generating action potentials. This study shows that further mathematical studies on the Hodgkin-Huxley model are needed in order to create a model without this weakness.

Keywords: all-or-none, computational modelling, mathematical model, transmembrane voltage, action potential

Procedia PDF Downloads 617
18329 The Lived Experience of Risk and Protective Contexts of Blind Successful University Students in Sidist Kilo Campus

Authors: Zelalem Markos Borko

Abstract:

The quality of life of people with blindness is significantly influenced by the level of resilience they possess. A qualitative approach of the descriptive phenomenological design was employed to address basic study objectives. The researcher purposely selected three blind graduate students from Sidist Kilo Campus and conducted a semi-structured interview to gather data. Data were analyzed by using thematic coding techniques. The present study found that personal characteristics such as commitment, living hope, motivation, positive self-esteem, self-confidence, and communication have shaped resiliency for successful university students with visual disabilities. The finding showed that the school environment is the place in which blind students had developed/experienced social, psychological, and economical competency and hope for their academic and entire life success. Furthermore, the finding showed that blind students had experienced individual, family, school, and community-related risks in the success track. Therefore, governmental and non-governmental organizations should provide training for students with visual impairments that focus on the individual traits that shape resilience for academic success, such as commitment, living hope, motivation, positive self-esteem, self-confidence, and communication and also community-oriented training should be to break the social stigma and discriminations for the individuals with the visual impairment.

Keywords: blind students, risk and protective factors, lived experience, success

Procedia PDF Downloads 82
18328 Model Order Reduction of Complex Airframes Using Component Mode Synthesis for Dynamic Aeroelasticity Load Analysis

Authors: Paul V. Thomas, Mostafa S. A. Elsayed, Denis Walch

Abstract:

Airframe structural optimization at different design stages results in new mass and stiffness distributions which modify the critical design loads envelop. Determination of aircraft critical loads is an extensive analysis procedure which involves simulating the aircraft at thousands of load cases as defined in the certification requirements. It is computationally prohibitive to use a Global Finite Element Model (GFEM) for the load analysis, hence reduced order structural models are required which closely represent the dynamic characteristics of the GFEM. This paper presents the implementation of Component Mode Synthesis (CMS) method for the generation of high fidelity Reduced Order Model (ROM) of complex airframes. Here, sub-structuring technique is used to divide the complex higher order airframe dynamical system into a set of subsystems. Each subsystem is reduced to fewer degrees of freedom using matrix projection onto a carefully chosen reduced order basis subspace. The reduced structural matrices are assembled for all the subsystems through interface coupling and the dynamic response of the total system is solved. The CMS method is employed to develop the ROM of a Bombardier Aerospace business jet which is coupled with an aerodynamic model for dynamic aeroelasticity loads analysis under gust turbulence. Another set of dynamic aeroelastic loads is also generated employing a stick model of the same aircraft. Stick model is the reduced order modelling methodology commonly used in the aerospace industry based on stiffness generation by unitary loading application. The extracted aeroelastic loads from both models are compared against those generated employing the GFEM. Critical loads Modal participation factors and modal characteristics of the different ROMs are investigated and compared against those of the GFEM. Results obtained show that the ROM generated using Craig Bampton CMS reduction process has a superior dynamic characteristics compared to the stick model.

Keywords: component mode synthesis, craig bampton reduction method, dynamic aeroelasticity analysis, model order reduction

Procedia PDF Downloads 210
18327 Transformer Fault Diagnostic Predicting Model Using Support Vector Machine with Gradient Decent Optimization

Authors: R. O. Osaseri, A. R. Usiobaifo

Abstract:

The power transformer which is responsible for the voltage transformation is of great relevance in the power system and oil-immerse transformer is widely used all over the world. A prompt and proper maintenance of the transformer is of utmost importance. The dissolved gasses content in power transformer, oil is of enormous importance in detecting incipient fault of the transformer. There is a need for accurate prediction of the incipient fault in transformer oil in order to facilitate the prompt maintenance and reducing the cost and error minimization. Study on fault prediction and diagnostic has been the center of many researchers and many previous works have been reported on the use of artificial intelligence to predict incipient failure of transformer faults. In this study machine learning technique was employed by using gradient decent algorithms and Support Vector Machine (SVM) in predicting incipient fault diagnosis of transformer. The method focuses on creating a system that improves its performance on previous result and historical data. The system design approach is basically in two phases; training and testing phase. The gradient decent algorithm is trained with a training dataset while the learned algorithm is applied to a set of new data. This two dataset is used to prove the accuracy of the proposed model. In this study a transformer fault diagnostic model based on Support Vector Machine (SVM) and gradient decent algorithms has been presented with a satisfactory diagnostic capability with high percentage in predicting incipient failure of transformer faults than existing diagnostic methods.

Keywords: diagnostic model, gradient decent, machine learning, support vector machine (SVM), transformer fault

Procedia PDF Downloads 324
18326 Parametric Appraisal of Robotic Arc Welding of Mild Steel Material by Principal Component Analysis-Fuzzy with Taguchi Technique

Authors: Amruta Rout, Golak Bihari Mahanta, Gunji Bala Murali, Bibhuti Bhusan Biswal, B. B. V. L. Deepak

Abstract:

The use of industrial robots for performing welding operation is one of the chief sign of contemporary welding in these days. The weld joint parameter and weld process parameter modeling is one of the most crucial aspects of robotic welding. As weld process parameters affect the weld joint parameters differently, a multi-objective optimization technique has to be utilized to obtain optimal setting of weld process parameter. In this paper, a hybrid optimization technique, i.e., Principal Component Analysis (PCA) combined with fuzzy logic has been proposed to get optimal setting of weld process parameters like wire feed rate, welding current. Gas flow rate, welding speed and nozzle tip to plate distance. The weld joint parameters considered for optimization are the depth of penetration, yield strength, and ultimate strength. PCA is a very efficient multi-objective technique for converting the correlated and dependent parameters into uncorrelated and independent variables like the weld joint parameters. Also in this approach, no need for checking the correlation among responses as no individual weight has been assigned to responses. Fuzzy Inference Engine can efficiently consider these aspects into an internal hierarchy of it thereby overcoming various limitations of existing optimization approaches. At last Taguchi method is used to get the optimal setting of weld process parameters. Therefore, it has been concluded the hybrid technique has its own advantages which can be used for quality improvement in industrial applications.

Keywords: robotic arc welding, weld process parameters, weld joint parameters, principal component analysis, fuzzy logic, Taguchi method

Procedia PDF Downloads 180
18325 Optimization of Lercanidipine Nanocrystals Using Design of Experiments Approach

Authors: Dolly Gadhiya, Jayvadan Patel, Mihir Raval

Abstract:

Lercanidipine hydrochloride is a calcium channel blockers used for treating angina pectoris and hypertension. Lercanidipine is a BCS Class II drug having poor aqueous solubility. Absolute bioavailability of Lercanidipine is very low and the main reason ascribed for this is poor aqueous solubility of the drug. Design and formulatation of nanocrystals by media milling method was main focus of this study. In this present study preliminary optimization was carried out with one factor at a time (OFAT) approach. For this different parameters like size of milling beads, amount of zirconium beads, types of stabilizer, concentrations of stabilizer, concentrations of drug, stirring speeds and milling time were optimized on the basis of particle size, polydispersity index and zeta potential. From the OFAT model different levels for above parameters selected for Plackett - Burman Design (PBD). Plackett-Burman design having 13 runs involving 6 independent variables was carried out at higher and lower level. Based on statistical analysis of PBD it was found that concentration of stabilizer, concentration of drug and stirring speed have significant impact on particle size, PDI, zeta potential value and saturation solubility. These experimental designs for preparation of nanocrystals were applied successfully which shows increase in aqueous solubility and dissolution rate of Lercanidipine hydrochloride.

Keywords: Lercanidipine hydrochloride, nanocrystals, OFAT, Plackett Burman

Procedia PDF Downloads 206
18324 Machine Learning Approaches Based on Recency, Frequency, Monetary (RFM) and K-Means for Predicting Electrical Failures and Voltage Reliability in Smart Cities

Authors: Panaya Sudta, Wanchalerm Patanacharoenwong, Prachya Bumrungkun

Abstract:

As With the evolution of smart grids, ensuring the reliability and efficiency of electrical systems in smart cities has become crucial. This paper proposes a distinct approach that combines advanced machine learning techniques to accurately predict electrical failures and address voltage reliability issues. This approach aims to improve the accuracy and efficiency of reliability evaluations in smart cities. The aim of this research is to develop a comprehensive predictive model that accurately predicts electrical failures and voltage reliability in smart cities. This model integrates RFM analysis, K-means clustering, and LSTM networks to achieve this objective. The research utilizes RFM analysis, traditionally used in customer value assessment, to categorize and analyze electrical components based on their failure recency, frequency, and monetary impact. K-means clustering is employed to segment electrical components into distinct groups with similar characteristics and failure patterns. LSTM networks are used to capture the temporal dependencies and patterns in customer data. This integration of RFM, K-means, and LSTM results in a robust predictive tool for electrical failures and voltage reliability. The proposed model has been tested and validated on diverse electrical utility datasets. The results show a significant improvement in prediction accuracy and reliability compared to traditional methods, achieving an accuracy of 92.78% and an F1-score of 0.83. This research contributes to the proactive maintenance and optimization of electrical infrastructures in smart cities. It also enhances overall energy management and sustainability. The integration of advanced machine learning techniques in the predictive model demonstrates the potential for transforming the landscape of electrical system management within smart cities. The research utilizes diverse electrical utility datasets to develop and validate the predictive model. RFM analysis, K-means clustering, and LSTM networks are applied to these datasets to analyze and predict electrical failures and voltage reliability. The research addresses the question of how accurately electrical failures and voltage reliability can be predicted in smart cities. It also investigates the effectiveness of integrating RFM analysis, K-means clustering, and LSTM networks in achieving this goal. The proposed approach presents a distinct, efficient, and effective solution for predicting and mitigating electrical failures and voltage issues in smart cities. It significantly improves prediction accuracy and reliability compared to traditional methods. This advancement contributes to the proactive maintenance and optimization of electrical infrastructures, overall energy management, and sustainability in smart cities.

Keywords: electrical state prediction, smart grids, data-driven method, long short-term memory, RFM, k-means, machine learning

Procedia PDF Downloads 58
18323 Parallel 2-Opt Local Search on GPU

Authors: Wen-Bao Qiao, Jean-Charles Créput

Abstract:

To accelerate the solution for large scale traveling salesman problems (TSP), a parallel 2-opt local search algorithm with simple implementation based on Graphics Processing Unit (GPU) is presented and tested in this paper. The parallel scheme is based on technique of data decomposition by dynamically assigning multiple K processors on the integral tour to treat K edges’ 2-opt local optimization simultaneously on independent sub-tours, where K can be user-defined or have a function relationship with input size N. We implement this algorithm with doubly linked list on GPU. The implementation only requires O(N) memory. We compare this parallel 2-opt local optimization against sequential exhaustive 2-opt search along integral tour on TSP instances from TSPLIB with more than 10000 cities.

Keywords: parallel 2-opt, double links, large scale TSP, GPU

Procedia PDF Downloads 628