Search results for: accelerated failure time model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 32388

Search results for: accelerated failure time model

29328 Fast Switching Mechanism for Multicasting Failure in OpenFlow Networks

Authors: Alaa Allakany, Koji Okamura

Abstract:

Multicast technology is an efficient and scalable technology for data distribution in order to optimize network resources. However, in the IP network, the responsibility for management of multicast groups is distributed among network routers, which causes some limitations such as delays in processing group events, high bandwidth consumption and redundant tree calculation. Software Defined Networking (SDN) represented by OpenFlow presented as a solution for many problems, in SDN the control plane and data plane are separated by shifting the control and management to a remote centralized controller, and the routers are used as a forwarder only. In this paper we will proposed fast switching mechanism for solving the problem of link failure in multicast tree based on Tabu Search heuristic algorithm and modifying the functions of OpenFlow switch to fasts switch to the pack up sub tree rather than sending to the controller. In this work we will implement multicasting OpenFlow controller, this centralized controller is a core part in our multicasting approach, which is responsible for 1- constructing the multicast tree, 2- handling the multicast group events and multicast state maintenance. And finally modifying OpenFlow switch functions for fasts switch to pack up paths. Forwarders, forward the multicast packet based on multicast routing entries which were generated by the centralized controller. Tabu search will be used as heuristic algorithm for construction near optimum multicast tree and maintain multicast tree to still near optimum in case of join or leave any members from multicast group (group events).

Keywords: multicast tree, software define networks, tabu search, OpenFlow

Procedia PDF Downloads 263
29327 Analytical Solutions of Josephson Junctions Dynamics in a Resonant Cavity for Extended Dicke Model

Authors: S.I.Mukhin, S. Seidov, A. Mukherjee

Abstract:

The Dicke model is a key tool for the description of correlated states of quantum atomic systems, excited by resonant photon absorption and subsequently emitting spontaneous coherent radiation in the superradiant state. The Dicke Hamiltonian (DH) is successfully used for the description of the dynamics of the Josephson Junction (JJ) array in a resonant cavity under applied current. In this work, we have investigated a generalized model, which is described by DH with a frustrating interaction term. This frustrating interaction term is explicitly the infinite coordinated interaction between all the spin half in the system. In this work, we consider an array of N superconducting islands, each divided into two sub-islands by a Josephson Junction, taken in a charged qubit / Cooper Pair Box (CPB) condition. The array is placed inside the resonant cavity. One important aspect of the problem lies in the dynamical nature of the physical observables involved in the system, such as condensed electric field and dipole moment. It is important to understand how these quantities behave with time to define the quantum phase of the system. The Dicke model without frustrating term is solved to find the dynamical solutions of the physical observables in analytic form. We have used Heisenberg’s dynamical equations for the operators and on applying newly developed Rotating Holstein Primakoff (HP) transformation and DH we have arrived at the four coupled nonlinear dynamical differential equations for the momentum and spin component operators. It is possible to solve the system analytically using two-time scales. The analytical solutions are expressed in terms of Jacobi's elliptic functions for the metastable ‘bound luminosity’ dynamic state with the periodic coherent beating of the dipoles that connect the two double degenerate dipolar ordered phases discovered previously. In this work, we have proceeded the analysis with the extended DH with a frustrating interaction term. Inclusion of the frustrating term involves complexity in the system of differential equations and it gets difficult to solve analytically. We have solved semi-classical dynamic equations using the perturbation technique for small values of Josephson energy EJ. Because the Hamiltonian contains parity symmetry, thus phase transition can be found if this symmetry is broken. Introducing spontaneous symmetry breaking term in the DH, we have derived the solutions which show the occurrence of finite condensate, showing quantum phase transition. Our obtained result matches with the existing results in this scientific field.

Keywords: Dicke Model, nonlinear dynamics, perturbation theory, superconductivity

Procedia PDF Downloads 134
29326 The Impact of Size of the Regional Economic Blocs to the Country’s Flows of Trade: Evidence from COMESA, EAC and Tanzania

Authors: Mosses E. Lufuke, Lorna M. Kamau

Abstract:

This paper attempted to assess whether the size of the regional economic bloc has an impact to the flow of trade to a particular country. Two different sized blocs (COMESA and EAC) and one country (Tanzania) have been used as the point of references. Using the results from of the analyses, the paper also was anticipated to establish whether it was rational for Tanzania to withdraw its membership from COMESA (the larger bloc) to join EAC (the small one). Gravity model has been used to estimate the relationship between the variables, from which the bilateral trade flows between Tanzania and the eighteen member countries of the two blocs (COMESA and EAC) was employed for the time between 2000 and 2013. In the model, the dummy variable for regional bloc (bloc) at which the Tanzania trade partner countries belong are also added to the model to understand which trade bloc exhibit higher trade flow with Tanzania. From the findings, it was noted that over the period of study (2000-2013) Tanzania acknowledged more than 257% of trade volume in EAC than in COMESA. Conclusive, it was noted that the flow of trade is explained by many other variables apart from the size of regional bloc; and that the size by itself offer insufficient evidence in causality relationship. The paper therefore remain neutral on such staggered switching decision since more analyses are required to establish the country’s trade flow, especially when if it had been in multiple membership of COMESA and EAC.

Keywords: economic bloc, flow of trade, size of bloc, switching

Procedia PDF Downloads 248
29325 Optimization of Reaction Parameters' Influences on Production of Bio-Oil from Fast Pyrolysis of Oil Palm Empty Fruit Bunch Biomass in a Fluidized Bed Reactor

Authors: Chayanoot Sangwichien, Taweesak Reungpeerakul, Kyaw Thu

Abstract:

Oil palm mills in Southern Thailand produced a large amount of biomass solid wastes. Lignocellulose biomass is the main source for production of biofuel which can be combined or used as an alternative to fossil fuels. Biomass composed of three main constituents of cellulose, hemicellulose, and lignin. Thermochemical conversion process applied to produce biofuel from biomass. Pyrolysis of biomass is the best way to thermochemical conversion of biomass into pyrolytic products (bio-oil, gas, and char). Operating parameters play an important role to optimize the product yields from fast pyrolysis of biomass. This present work concerns with the modeling of reaction kinetics parameters for fast pyrolysis of empty fruit bunch in the fluidized bed reactor. A global kinetic model used to predict the product yields from fast pyrolysis of empty fruit bunch. The reaction temperature and vapor residence time parameters are mainly affected by product yields of EFB pyrolysis. The reaction temperature and vapor residence time parameters effects on empty fruit bunch pyrolysis are considered at the reaction temperature in the range of 450-500˚C and at a vapor residence time of 2 s, respectively. The optimum simulated bio-oil yield of 53 wt.% obtained at the reaction temperature and vapor residence time of 450˚C and 2 s, 500˚C and 1 s, respectively. The simulated data are in good agreement with the reported experimental data. These simulated data can be applied to the performance of experiment work for the fast pyrolysis of biomass.

Keywords: kinetics, empty fruit bunch, fast pyrolysis, modeling

Procedia PDF Downloads 214
29324 Intelligent Platform for Photovoltaic Park Operation and Maintenance

Authors: Andreas Livera, Spyros Theocharides, Michalis Florides, Charalambos Anastassiou

Abstract:

A main challenge in the quest for ensuring quality of operation, especially for photovoltaic (PV) systems, is to safeguard the reliability and optimal performance by detecting and diagnosing potential failures and performance losses at early stages or before the occurrence through real-time monitoring, supervision, fault detection, and predictive maintenance. The purpose of this work is to present the functionalities and results related to the development and validation of a software platform for PV assets diagnosis and maintenance. The platform brings together proprietary hardware sensors and software algorithms to enable the early detection and prediction of the most common and critical faults in PV systems. It was validated using field measurements from operating PV systems. The results showed the effectiveness of the platform for detecting faults and losses (e.g., inverter failures, string disconnections, and potential induced degradation) at early stages, forecasting PV power production while also providing recommendations for maintenance actions. Increased PV energy yield production and revenue can be thus achieved while also minimizing operation and maintenance (O&M) costs.

Keywords: failure detection and prediction, operation and maintenance, performance monitoring, photovoltaic, platform, recommendations, predictive maintenance

Procedia PDF Downloads 49
29323 A Lightweight Pretrained Encrypted Traffic Classification Method with Squeeze-and-Excitation Block and Sharpness-Aware Optimization

Authors: Zhiyan Meng, Dan Liu, Jintao Meng

Abstract:

Dependable encrypted traffic classification is crucial for improving cybersecurity and handling the growing amount of data. Large language models have shown that learning from large datasets can be effective, making pre-trained methods for encrypted traffic classification popular. However, attention-based pre-trained methods face two main issues: their large neural parameters are not suitable for low-computation environments like mobile devices and real-time applications, and they often overfit by getting stuck in local minima. To address these issues, we developed a lightweight transformer model, which reduces the computational parameters through lightweight vocabulary construction and Squeeze-and-Excitation Block. We use sharpness-aware optimization to avoid local minima during pre-training and capture temporal features with relative positional embeddings. Our approach keeps the model's classification accuracy high for downstream tasks. We conducted experiments on four datasets -USTC-TFC2016, VPN 2016, Tor 2016, and CICIOT 2022. Even with fewer than 18 million parameters, our method achieves classification results similar to methods with ten times as many parameters.

Keywords: sharpness-aware optimization, encrypted traffic classification, squeeze-and-excitation block, pretrained model

Procedia PDF Downloads 30
29322 Displacement Based Design of a Dual Structural System

Authors: Romel Cordova Shedan

Abstract:

The traditional seismic design is the methodology of Forced Based Design (FBD). The Displacement Based Design (DBD) is a seismic design that considers structural damage to achieve a failure mechanism of the structure before the collapse. It is easier to quantify damage of a structure with displacements rather than forces. Therefore, a structure to achieve an inelastic displacement design with good ductility, it is necessary to be damaged. The first part of this investigation is about differences between the methodologies of DBD and FBD with some DBD advantages. In the second part, there is a study case about a dual building 5-story, which is regular in plan and elevation. The building is located in a seismic zone, which acceleration in firm soil is 45% of the acceleration of gravity. Then it is applied both methodologies into the study case to compare its displacements, shear forces and overturning moments. In the third part, the Dynamic Time History Analysis (DTHA) is done, to compare displacements with DBD and FBD methodologies. Three accelerograms were used and the magnitude of the acceleration scaled to be spectrum compatible with design spectrum. Then, using ASCE 41-13 guidelines, the hinge plastics were assigned to structure. Finally, both methodologies results about study case are compared. It is important to take into account that the seismic performance level of the building for DBD is greater than FBD method. This is due to drifts of DBD are in the order of 2.0% and 2.5% comparing with FBD drifts of 0.7%. Therefore, displacements of DBD is greater than the FBD method. Shear forces of DBD result greater than FBD methodology. These strengths of DBD method ensures that structure achieves design inelastic displacements, because those strengths were obtained due to a displacement spectrum reduction factor which depends on damping and ductility of the dual system. Also, the displacements for the study case for DBD results to be greater than FBD and DTHA. In that way, it proves that the seismic performance level of the building for DBD is greater than FBD method. Due to drifts of DBD which are in the order of 2.0% and 2.5% compared with little FBD drifts of 0.7%.

Keywords: displacement-based design, displacement spectrum reduction factor, dynamic time history analysis, forced based design

Procedia PDF Downloads 229
29321 Comparative Study of Sorption of Cr Ions and Dye Bezaktiv Yellow HE-4G with the Use of Adsorbents Natural Mixture of Olive Stone and Date Pits from Aqueous Solution

Authors: H. Aksas, H. Babaci, K. Louhab

Abstract:

In this paper, a comparative study of the adsorption of Chromium and dyes, onto mixture biosorbents, olive stones and date pits at different percentage was investigated in aqueous solution. The study of various parameters: Effect of contact time, pH, temperature and initial concentration shows that these materials possess a high affinity for the adsorption of chromium for the adsorption of dye bezaktiv yellow HE-4G. To deepen the comparative study of the adsorption of chromium and dye with the use of different blends of olive stones and date pits, the following models are studied: Langmuir, Freundlich isotherms and Dubinin- Radushkvich (D-R) were used as the adsorption equilibrium data model. Langmuir isotherm model was the most suitable for the adsorption of the dye bezaktiv HE-4G and the D-R model is most suitable for adsorption Chrome. The pseudo-first-order model, pseudo-second order and intraparticle diffusion were used to describe the adsorption kinetics. The apparent activation energy was found to be less than 8KJ/mol, which is characteristic of a controlled chemical reaction for the adsorption of two materials. t was noticed that adsorption of chromium and dye BEZAKTIV HE-YELLOW 4G follows the kinetics of the pseudo second order. The study of the effect of temperature was quantified by calculating various thermodynamic parameters such as Gibbs free energy, enthalpy and entropy changes. The resulting thermodynamic parameters indicate the endothermic nature of the adsorption of Cr (VI) ions and the dye Bezaktiv HE-4G. But these materials are very good adsorbents, as they represent a low cost. in addition, it has been noticed that the greater the quantity of olive stone in the mixture increases, the adsorption ability of the dye or chromium increases.

Keywords: chromium ions, anions dye, sorption, mixed adsorbents, olive stone, date pits

Procedia PDF Downloads 228
29320 Reverse Logistics End of Life Products Acquisition and Sorting

Authors: Badli Shah Mohd Yusoff, Khairur Rijal Jamaludin, Rozetta Dollah

Abstract:

The emerging of reverse logistics and product recovery management is an important concept in reconciling economic and environmental objectives through recapturing values of the end of life product returns. End of life products contains valuable modules, parts, residues and materials that can create value if recovered efficiently. The main objective of this study is to explore and develop a model to recover as much of the economic value as reasonably possible to find the optimality of return acquisition and sorting to meet demand and maximize profits over time. In this study, the benefits that can be obtained for remanufacturer is to develop demand forecasting of used products in the future with uncertainty of returns and quality of products. Formulated based on a generic disassembly tree, the proposed model focused on three reverse logistics activity, namely refurbish, remanufacture and disposal incorporating all plausible means quality levels of the returns. While stricter sorting policy, constitute to the decrease amount of products to be refurbished or remanufactured and increases the level of discarded products. Numerical experiments carried out to investigate the characteristics and behaviour of the proposed model with mathematical programming model using Lingo 16.0 for medium-term planning of return acquisition, disassembly (refurbish or remanufacture) and disposal activities. Moreover, the model seeks an analysis a number of decisions relating to trade off management system to maximize revenue from the collection of use products reverse logistics services through refurbish and remanufacture recovery options. The results showed that full utilization in the sorting process leads the system to obtain less quantity from acquisition with minimal overall cost. Further, sensitivity analysis provides a range of possible scenarios to consider in optimizing the overall cost of refurbished and remanufactured products.

Keywords: core acquisition, end of life, reverse logistics, quality uncertainty

Procedia PDF Downloads 303
29319 Characterization of Aerosol Droplet in Absorption Columns to Avoid Amine Emissions

Authors: Hammad Majeed, Hanna Knuutila, Magne Hilestad, Hallvard Svendsen

Abstract:

Formation of aerosols can cause serious complications in industrial exhaust gas CO2 capture processes. SO3 present in the flue gas can cause aerosol formation in an absorption based capture process. Small mist droplets and fog formed can normally not be removed in conventional demisting equipment because their submicron size allows the particles or droplets to follow the gas flow. As a consequence of this aerosol based emissions in the order of grams per Nm3 have been identified from PCCC plants. In absorption processes aerosols are generated by spontaneous condensation or desublimation processes in supersaturated gas phases. Undesired aerosol development may lead to amine emissions many times larger than what would be encountered in a mist free gas phase in PCCC development. It is thus of crucial importance to understand the formation and build-up of these aerosols in order to mitigate the problem.Rigorous modelling of aerosol dynamics leads to a system of partial differential equations. In order to understand mechanics of a particle entering an absorber an implementation of the model is created in Matlab. The model predicts the droplet size, the droplet internal variable profiles and the mass transfer fluxes as function of position in the absorber. The Matlab model is based on a subclass method of weighted residuals for boundary value problems named, orthogonal collocation method. The model comprises a set of mass transfer equations for transferring components and the essential diffusion reaction equations to describe the droplet internal profiles for all relevant constituents. Also included is heat transfer across the interface and inside the droplet. This paper presents results describing the basic simulation tool for the characterization of aerosols formed in CO2 absorption columns and gives examples as to how various entering droplets grow or shrink through an absorber and how their composition changes with respect to time. Below are given some preliminary simulation results for an aerosol droplet composition and temperature profiles. Results: As an example a droplet of initial size of 3 microns, initially containing a 5M MEA, solution is exposed to an atmosphere free of MEA. Composition of the gas phase and temperature is changing with respect to time throughout the absorber.

Keywords: amine solvents, emissions, global climate change, simulation and modelling, aerosol generation

Procedia PDF Downloads 265
29318 Advanced Model for Calculation of the Neutral Axis Shifting and the Wall Thickness Distribution in Rotary Draw Bending Processes

Authors: B. Engel, H. Hassan

Abstract:

Rotary draw bending is a method which is being used in tube forming. In the tube bending process, the neutral axis moves towards the inner arc and the wall thickness distribution changes for tube’s cross section. Thinning takes place in the outer arc of the tube (extrados) due to the stretching of the material, whereas thickening occurs in the inner arc of the tube (intrados) due to the comparison of the material. The calculations of the wall thickness distribution, neutral axis shifting, and strain distribution have not been accurate enough, so far. The previous model (the geometrical model) describes the neutral axis shifting and wall thickness distribution. The geometrical of the tube, bending radius and bending angle are considered in the geometrical model, while the influence of the material properties of the tube forming are ignored. The advanced model is a modification of the previous model using material properties that depends on the correction factor. The correction factor is a purely empirically determined factor. The advanced model was compared with the Finite element simulation (FE simulation) using a different bending factor (Bf=bending radius/ diameter of the tube), wall thickness (Wf=diameter of the tube/ wall thickness), and material properties (strain hardening exponent). Finite element model of rotary draw bending has been performed in PAM-TUBE program (version: 2012). Results from the advanced model resemble the FE simulation and the experimental test.

Keywords: rotary draw bending, material properties, neutral axis shifting, wall thickness distribution

Procedia PDF Downloads 397
29317 Implementation of an Image Processing System Using Artificial Intelligence for the Diagnosis of Malaria Disease

Authors: Mohammed Bnebaghdad, Feriel Betouche, Malika Semmani

Abstract:

Image processing become more sophisticated over time due to technological advances, especially artificial intelligence (AI) technology. Currently, AI image processing is used in many areas, including surveillance, industry, science, and medicine. AI in medical image processing can help doctors diagnose diseases faster, with minimal mistakes, and with less effort. Among these diseases is malaria, which remains a major public health challenge in many parts of the world. It affects millions of people every year, particularly in tropical and subtropical regions. Early detection of malaria is essential to prevent serious complications and reduce the burden of the disease. In this paper, we propose and implement a scheme based on AI image processing to enhance malaria disease diagnosis through automated analysis of blood smear images. The scheme is based on the convolutional neural network (CNN) method. So, we have developed a model that classifies infected and uninfected single red cells using images available on Kaggle, as well as real blood smear images obtained from the Central Laboratory of Medical Biology EHS Laadi Flici (formerly El Kettar) in Algeria. The real images were segmented into individual cells using the watershed algorithm in order to match the images from the Kaagle dataset. The model was trained and tested, achieving an accuracy of 99% and 97% accuracy for new real images. This validates that the model performs well with new real images, although with slightly lower accuracy. Additionally, the model has been embedded in a Raspberry Pi4, and a graphical user interface (GUI) was developed to visualize the malaria diagnostic results and facilitate user interaction.

Keywords: medical image processing, malaria parasite, classification, CNN, artificial intelligence

Procedia PDF Downloads 20
29316 Towards Computational Fluid Dynamics Based Methodology to Accelerate Bioprocess Scale Up and Scale Down

Authors: Vishal Kumar Singh

Abstract:

Bioprocess development is a time-constrained activity aimed at harnessing the full potential of culture performance in an ambience that is not natural to cells. Even with the use of chemically defined media and feeds, a significant amount of time is devoted in identifying the apt operating parameters. In addition, the scale-up of these processes is often accompanied by loss of antibody titer and product quality, which further delays the commercialization of the drug product. In such a scenario, the investigation of this disparity of culture performance is done by further experimentation at a smaller scale that is representative of at-scale production bioreactors. These scale-down model developments are also time-intensive. In this study, a computation fluid dynamics-based multi-objective scaling approach has been illustrated to speed up the process transfer. For the implementation of this approach, a transient multiphase water-air system has been studied in Ansys CFX to visualize the air bubble distribution and volumetric mass transfer coefficient (kLa) profiles, followed by the design of experiment based parametric optimization approach to define the operational space. The proposed approach is completely in silico and requires minimum experimentation, thereby rendering a high throughput to the overall process development.

Keywords: bioprocess development, scale up, scale down, computation fluid dynamics, multi-objective, Ansys CFX, design of experiment

Procedia PDF Downloads 82
29315 Cybersecurity Assessment of Decentralized Autonomous Organizations in Smart Cities

Authors: Claire Biasco, Thaier Hayajneh

Abstract:

A smart city is the integration of digital technologies in urban environments to enhance the quality of life. Smart cities capture real-time information from devices, sensors, and network data to analyze and improve city functions such as traffic analysis, public safety, and environmental impacts. Current smart cities face controversy due to their reliance on real-time data tracking and surveillance. Internet of Things (IoT) devices and blockchain technology are converging to reshape smart city infrastructure away from its centralized model. Connecting IoT data to blockchain applications would create a peer-to-peer, decentralized model. Furthermore, blockchain technology powers the ability for IoT device data to shift from the ownership and control of centralized entities to individuals or communities with Decentralized Autonomous Organizations (DAOs). In the context of smart cities, DAOs can govern cyber-physical systems to have a greater influence over how urban services are being provided. This paper will explore how the core components of a smart city now apply to DAOs. We will also analyze different definitions of DAOs to determine their most important aspects in relation to smart cities. Both categorizations will provide a solid foundation to conduct a cybersecurity assessment of DAOs in smart cities. It will identify the benefits and risks of adopting DAOs as they currently operate. The paper will then provide several mitigation methods to combat cybersecurity risks of DAO integrations. Finally, we will give several insights into what challenges will be faced by DAO and blockchain spaces in the coming years before achieving a higher level of maturity.

Keywords: blockchain, IoT, smart city, DAO

Procedia PDF Downloads 121
29314 Analysis and Prediction of Netflix Viewing History Using Netflixlatte as an Enriched Real Data Pool

Authors: Amir Mabhout, Toktam Ghafarian, Amirhossein Farzin, Zahra Makki, Sajjad Alizadeh, Amirhossein Ghavi

Abstract:

The high number of Netflix subscribers makes it attractive for data scientists to extract valuable knowledge from the viewers' behavioural analyses. This paper presents a set of statistical insights into viewers' viewing history. After that, a deep learning model is used to predict the future watching behaviour of the users based on previous watching history within the Netflixlatte data pool. Netflixlatte in an aggregated and anonymized data pool of 320 Netflix viewers with a length 250 000 data points recorded between 2008-2022. We observe insightful correlations between the distribution of viewing time and the COVID-19 pandemic outbreak. The presented deep learning model predicts future movie and TV series viewing habits with an average loss of 0.175.

Keywords: data analysis, deep learning, LSTM neural network, netflix

Procedia PDF Downloads 251
29313 Subpixel Corner Detection for Monocular Camera Linear Model Research

Authors: Guorong Sui, Xingwei Jia, Fei Tong, Xiumin Gao

Abstract:

Camera calibration is a fundamental issue of high precision noncontact measurement. And it is necessary to analyze and study the reliability and application range of its linear model which is often used in the camera calibration. According to the imaging features of monocular cameras, a camera model which is based on the image pixel coordinates and three dimensional space coordinates is built. Using our own customized template, the image pixel coordinate is obtained by the subpixel corner detection method. Without considering the aberration of the optical system, the feature extraction and linearity analysis of the line segment in the template are performed. Moreover, the experiment is repeated 11 times by constantly varying the measuring distance. At last, the linearity of the camera is achieved by fitting 11 groups of data. The camera model measurement results show that the relative error does not exceed 1%, and the repeated measurement error is not more than 0.1 mm magnitude. Meanwhile, it is found that the model has some measurement differences in the different region and object distance. The experiment results show this linear model is simple and practical, and have good linearity within a certain object distance. These experiment results provide a powerful basis for establishment of the linear model of camera. These works will have potential value to the actual engineering measurement.

Keywords: camera linear model, geometric imaging relationship, image pixel coordinates, three dimensional space coordinates, sub-pixel corner detection

Procedia PDF Downloads 277
29312 An Overbooking Model for Car Rental Service with Different Types of Cars

Authors: Naragain Phumchusri, Kittitach Pongpairoj

Abstract:

Overbooking is a very useful revenue management technique that could help reduce costs caused by either undersales or oversales. In this paper, we propose an overbooking model for two types of cars that can minimize the total cost for car rental service. With two types of cars, there is an upgrade possibility for lower type to upper type. This makes the model more complex than one type of cars scenario. We have found that convexity can be proved in this case. Sensitivity analysis of the parameters is conducted to observe the effects of relevant parameters on the optimal solution. Model simplification is proposed using multiple linear regression analysis, which can help estimate the optimal overbooking level using appropriate independent variables. The results show that the overbooking level from multiple linear regression model is relatively close to the optimal solution (with the adjusted R-squared value of at least 72.8%). To evaluate the performance of the proposed model, the total cost was compared with the case where the decision maker uses a naïve method for the overbooking level. It was found that the total cost from optimal solution is only 0.5 to 1 percent (on average) lower than the cost from regression model, while it is approximately 67% lower than the cost obtained by the naïve method. It indicates that our proposed simplification method using regression analysis can effectively perform in estimating the overbooking level.

Keywords: overbooking, car rental industry, revenue management, stochastic model

Procedia PDF Downloads 172
29311 Interpreting Privacy Harms from a Non-Economic Perspective

Authors: Christopher Muhawe, Masooda Bashir

Abstract:

With increased Internet Communication Technology(ICT), the virtual world has become the new normal. At the same time, there is an unprecedented collection of massive amounts of data by both private and public entities. Unfortunately, this increase in data collection has been in tandem with an increase in data misuse and data breach. Regrettably, the majority of data breach and data misuse claims have been unsuccessful in the United States courts for the failure of proof of direct injury to physical or economic interests. The requirement to express data privacy harms from an economic or physical stance negates the fact that not all data harms are physical or economic in nature. The challenge is compounded by the fact that data breach harms and risks do not attach immediately. This research will use a descriptive and normative approach to show that not all data harms can be expressed in economic or physical terms. Expressing privacy harms purely from an economic or physical harm perspective negates the fact that data insecurity may result into harms which run counter the functions of privacy in our lives. The promotion of liberty, selfhood, autonomy, promotion of human social relations and the furtherance of the existence of a free society. There is no economic value that can be placed on these functions of privacy. The proposed approach addresses data harms from a psychological and social perspective.

Keywords: data breach and misuse, economic harms, privacy harms, psychological harms

Procedia PDF Downloads 195
29310 Hybrid Quasi-Steady Thermal Lattice Boltzmann Model for Studying the Behavior of Oil in Water Emulsions Used in Machining Tool Cooling and Lubrication

Authors: W. Hasan, H. Farhat, A. Alhilo, L. Tamimi

Abstract:

Oil in water (O/W) emulsions are utilized extensively for cooling and lubricating cutting tools during parts machining. A robust Lattice Boltzmann (LBM) thermal-surfactants model, which provides a useful platform for exploring complex emulsions’ characteristics under variety of flow conditions, is used here for the study of the fluid behavior during conventional tools cooling. The transient thermal capabilities of the model are employed for simulating the effects of the flow conditions of O/W emulsions on the cooling of cutting tools. The model results show that the temperature outcome is slightly affected by reversing the direction of upper plate (workpiece). On the other hand, an important increase in effective viscosity is seen which supports better lubrication during the work.

Keywords: hybrid lattice Boltzmann method, Gunstensen model, thermal, surfactant-covered droplet, Marangoni stress

Procedia PDF Downloads 304
29309 The High Efficiency of Cationic Azo Dye Removal Using Raw, Purified and Pillared Clay from Algerian Clay

Authors: Amina Ramdani, Abdelkader Kadeche, Zoubida Taleb, Safia Taleb

Abstract:

The aim of this present study is to evaluate the adsorption capacity of a dye, Malachite green, on a local Algerian montmorillonite clay mineral (raw, purified and Cr-pillared). Various parameters influencing the dye adsorption process ie contact time, adsorbent dose, initial concentration of dye, pH of the solution and temperature. Cr pillared clay has been obtained with a better surface character than purified and natural clay. An increase in basal spacing from 12.45 Å (Mont-Na) to 22.88 Å (Mont-PLCr), surface area from 67 m2 /g (Mont-Na) to 102 m2 /g (Mont-PLCr). The experimental results show that the dye adsorption kinetic were fast: 5 min for Cr-pillared clay mineral, and 30 min for raw and purified clay mineral (RC and Mont-Na). The removal efficiency on Mont-PLCr (98.64%) is greater than that of Mont-Na (86.20%) and RC (82.09%). The acidity and basicity of the medium considerably affect the adsorption of the dye. It attained its maximum at pH 4.8. The equilibrium and kinetic data were found to fit well the Langmuir model and the pseudo-second-order model.

Keywords: Dye removal, pillared clay, isotherm, kinetic

Procedia PDF Downloads 166
29308 Preparation of Superparamagnetic Functionalized Magnetite Nanoparticles for Magnetically Separable Catalysis

Authors: Priya Arora, Jaspreet K. Rajput

Abstract:

Superparamagnetism has accelerated the research and use of more economical and ecological magnetically separable catalysts due to their more efficient isolation by response to an applied magnetic field. Magnetite nanomaterials coated by SiO2 shell have received a great deal of focus in the last decades as it provides high stability and also easy further surface functionalization depending upon the application. In this protocol, Fe3O4 magnetic nanoparticles have been synthesized by co-precipitation combined with sonication method. Further, Fe3O4 superparamagnetic nanoparticles have been functionalized by various moieties to serve as efficient catalysts for multicomponent reactions. The functionalized nanoparticles were characterized by techniques like Fourier transform infrared spectroscopy (FT-IR), transmission electron microscopy (TEM), scanning electron microscopy (SEM), vibrating sample magnetometer (VSM), thermogravimetric analysis (TGA), Brunauer-Emmett-Teller (BET) surface area analysis. The as prepared nanocatalysts can be reused for several times without any significant loss in its activity. The utilization of magnetic nanoparticles as catalysts for this reaction is one approach i.e. green, inexpensive, facile and widely applicable.

Keywords: functionalized, magnetite, multicomponent reactions, superparamagnetic

Procedia PDF Downloads 340
29307 Developing and Evaluating Clinical Risk Prediction Models for Coronary Artery Bypass Graft Surgery

Authors: Mohammadreza Mohebbi, Masoumeh Sanagou

Abstract:

The ability to predict clinical outcomes is of great importance to physicians and clinicians. A number of different methods have been used in an effort to accurately predict these outcomes. These methods include the development of scoring systems based on multivariate statistical modelling, and models involving the use of classification and regression trees. The process usually consists of two consecutive phases, namely model development and external validation. The model development phase consists of building a multivariate model and evaluating its predictive performance by examining calibration and discrimination, and internal validation. External validation tests the predictive performance of a model by assessing its calibration and discrimination in different but plausibly related patients. A motivate example focuses on prediction modeling using a sample of patients undergone coronary artery bypass graft (CABG) has been used for illustrative purpose and a set of primary considerations for evaluating prediction model studies using specific quality indicators as criteria to help stakeholders evaluate the quality of a prediction model study has been proposed.

Keywords: clinical prediction models, clinical decision rule, prognosis, external validation, model calibration, biostatistics

Procedia PDF Downloads 297
29306 Analysis of Sweat Evaporation and Heat Transfer on Skin Surface: A Pointwise Numerical Study

Authors: Utsav Swarnkar, Rabi Pathak, Rina Maiti

Abstract:

This study aims to investigate the thermoregulatory role of sweating by comprehensively analyzing the evaporation process and its thermal cooling impact on local skin temperature at various time intervals. Traditional experimental methods struggle to fully capture these intricate phenomena. Therefore, numerical simulations play a crucial role in assessing sweat production rates and associated thermal cooling. This research utilizes transient computational fluid dynamics (CFD) to enhance our understanding of the evaporative cooling process on human skin. We conducted a simulation employing the k-w SST turbulence model. This simulation includes a scenario where sweat evaporation occurs over the skin surface, and at particular time intervals, temperatures at different locations have been observed and its effect explained. During this study, sweat evaporation was monitored on the skin surface following the commencement of the simulation. Subsequent to the simulation, various observations were made regarding temperature fluctuations at specific points over time intervals. It was noted that points situated closer to the periphery of the droplets exhibited higher levels of heat transfer and lower temperatures, whereas points within the droplets displayed contrasting trends.

Keywords: CFD, sweat, evaporation, multiphase flow, local heat loss

Procedia PDF Downloads 67
29305 Physics-Informed Convolutional Neural Networks for Reservoir Simulation

Authors: Jiangxia Han, Liang Xue, Keda Chen

Abstract:

Despite the significant progress over the last decades in reservoir simulation using numerical discretization, meshing is complex. Moreover, the high degree of freedom of the space-time flow field makes the solution process very time-consuming. Therefore, we present Physics-Informed Convolutional Neural Networks(PICNN) as a hybrid scientific theory and data method for reservoir modeling. Besides labeled data, the model is driven by the scientific theories of the underlying problem, such as governing equations, boundary conditions, and initial conditions. PICNN integrates governing equations and boundary conditions into the network architecture in the form of a customized convolution kernel. The loss function is composed of data matching, initial conditions, and other measurable prior knowledge. By customizing the convolution kernel and minimizing the loss function, the neural network parameters not only fit the data but also honor the governing equation. The PICNN provides a methodology to model and history-match flow and transport problems in porous media. Numerical results demonstrate that the proposed PICNN can provide an accurate physical solution from a limited dataset. We show how this method can be applied in the context of a forward simulation for continuous problems. Furthermore, several complex scenarios are tested, including the existence of data noise, different work schedules, and different good patterns.

Keywords: convolutional neural networks, deep learning, flow and transport in porous media, physics-informed neural networks, reservoir simulation

Procedia PDF Downloads 143
29304 Artificial Steady-State-Based Nonlinear MPC for Wheeled Mobile Robot

Authors: M. H. Korayem, Sh. Ameri, N. Yousefi Lademakhi

Abstract:

To ensure the stability of closed-loop nonlinear model predictive control (NMPC) within a finite horizon, there is a need for appropriate design terminal ingredients, which can be a time-consuming and challenging effort. Otherwise, in order to ensure the stability of the control system, it is necessary to consider an infinite predictive horizon. Increasing the prediction horizon increases computational demand and slows down the implementation of the method. In this study, a new technique has been proposed to ensure system stability without terminal ingredients. This technique has been employed in the design of the NMPC algorithm, leading to a reduction in the computational complexity of designing terminal ingredients and computational burden. The studied system is a wheeled mobile robot (WMR) subjected to non-holonomic constraints. Simulation has been investigated for two problems: trajectory tracking and adjustment mode.

Keywords: wheeled mobile robot, nonlinear model predictive control, stability, without terminal ingredients

Procedia PDF Downloads 91
29303 Modeling and Simulation of a CMOS-Based Analog Function Generator

Authors: Madina Hamiane

Abstract:

Modelling and simulation of an analogy function generator is presented based on a polynomial expansion model. The proposed function generator model is based on a 10th order polynomial approximation of any of the required functions. The polynomial approximations of these functions can then be implemented using basic CMOS circuit blocks. In this paper, a circuit model is proposed that can simultaneously generate many different mathematical functions. The circuit model is designed and simulated with HSPICE and its performance is demonstrated through the simulation of a number of non-linear functions.

Keywords: modelling and simulation, analog function generator, polynomial approximation, CMOS transistors

Procedia PDF Downloads 459
29302 Robust Optimisation Model and Simulation-Particle Swarm Optimisation Approach for Vehicle Routing Problem with Stochastic Demands

Authors: Mohanad Al-Behadili, Djamila Ouelhadj

Abstract:

In this paper, a specific type of vehicle routing problem under stochastic demand (SVRP) is considered. This problem is of great importance because it models for many of the real world vehicle routing applications. This paper used a robust optimisation model to solve the problem along with the novel Simulation-Particle Swarm Optimisation (Sim-PSO) approach. The proposed Sim-PSO approach is based on the hybridization of the Monte Carlo simulation technique with the PSO algorithm. A comparative study between the proposed model and the Sim-PSO approach against other solution methods in the literature has been given in this paper. This comparison including the Analysis of Variance (ANOVA) to show the ability of the model and solution method in solving the complicated SVRP. The experimental results show that the proposed model and Sim-PSO approach has a significant impact on the obtained solution by providing better quality solutions comparing with well-known algorithms in the literature.

Keywords: stochastic vehicle routing problem, robust optimisation model, Monte Carlo simulation, particle swarm optimisation

Procedia PDF Downloads 277
29301 Plotting of an Ideal Logic versus Resource Outflow Graph through Response Analysis on a Strategic Management Case Study Based Questionnaire

Authors: Vinay A. Sharma, Shiva Prasad H. C.

Abstract:

The initial stages of any project are often observed to be in a mixed set of conditions. Setting up the project is a tough task, but taking the initial decisions is rather not complex, as some of the critical factors are yet to be introduced into the scenario. These simple initial decisions potentially shape the timeline and subsequent events that might later be plotted on it. Proceeding towards the solution for a problem is the primary objective in the initial stages. The optimization in the solutions can come later, and hence, the resources deployed towards attaining the solution are higher than what they would have been in the optimized versions. A ‘logic’ that counters the problem is essentially the core of the desired solution. Thus, if the problem is solved, the deployment of resources has led to the required logic being attained. As the project proceeds along, the individuals working on the project face fresh challenges as a team and are better accustomed to their surroundings. The developed, optimized solutions are then considered for implementation, as the individuals are now experienced, and know better of the consequences and causes of possible failure, and thus integrate the adequate tolerances wherever required. Furthermore, as the team graduates in terms of strength, acquires prodigious knowledge, and begins its efficient transfer, the individuals in charge of the project along with the managers focus more on the optimized solutions rather than the traditional ones to minimize the required resources. Hence, as time progresses, the authorities prioritize attainment of the required logic, at a lower amount of dedicated resources. For empirical analysis of the stated theory, leaders and key figures in organizations are surveyed for their ideas on appropriate logic required for tackling a problem. Key-pointers spotted in successfully implemented solutions are noted from the analysis of the responses and a metric for measuring logic is developed. A graph is plotted with the quantifiable logic on the Y-axis, and the dedicated resources for the solutions to various problems on the X-axis. The dedicated resources are plotted over time, and hence the X-axis is also a measure of time. In the initial stages of the project, the graph is rather linear, as the required logic will be attained, but the consumed resources are also high. With time, the authorities begin focusing on optimized solutions, since the logic attained through them is higher, but the resources deployed are comparatively lower. Hence, the difference between consecutive plotted ‘resources’ reduces and as a result, the slope of the graph gradually increases. On an overview, the graph takes a parabolic shape (beginning on the origin), as with each resource investment, ideally, the difference keeps on decreasing, and the logic attained through the solution keeps increasing. Even if the resource investment is higher, the managers and authorities, ideally make sure that the investment is being made on a proportionally high logic for a larger problem, that is, ideally the slope of the graph increases with the plotting of each point.

Keywords: decision-making, leadership, logic, strategic management

Procedia PDF Downloads 108
29300 Analysis of Brushless DC Motor with Trapezoidal Back EMF Using Matlab

Authors: Taha Ahmed Husain

Abstract:

The dynamic characteristics such as speed and torque as well as voltages and currents of pwm brushless DC motor inverter are analyzed with a MATLAB model. The contribution of external load torque and friction torque is monitored. The switching function technique is adopted for the current control of the embedded three phase inverter that drives the brushless DC motor.In switching functions the power conversions circuits can be modeled according to their functions rather than circuit topologies. Therefore, it can achieve simplification of the overall power conversion functions. The trapezoidal type (back emf) is used in the model as ithas lower switching loss compared with sinusoidal type (back emf). Results show reliable time analysis for speed, torque, phase and line voltages and currents and the effect of current commutation is clearly observed.

Keywords: BLDC motor, brushless dc motors, pwm inverter, DC motor control, trapezoidal back emf, ripple torque in brushless DC motor

Procedia PDF Downloads 598
29299 Is the Okun's Law Valid in Tunisia?

Authors: El Andari Chifaa, Bouaziz Rached

Abstract:

The central focus of this paper was to check whether the Okun’s law in Tunisia is valid or not. For this purpose, we have used quarterly time series data during the period 1990Q1-2014Q1. Firstly, we applied the error correction model instead of the difference version of Okun's Law, the Engle-Granger and Johansen test are employed to find out long run association between unemployment, production, and how error correction mechanism (ECM) is used for short run dynamic. Secondly, we used the gap version of Okun’s law where the estimation is done from three band pass filters which are mathematical tools used in macro-economic and especially in business cycles theory. The finding of the study indicates that the inverse relationship between unemployment and output is verified in the short and long term, and the Okun's law holds for the Tunisian economy, but with an Okun’s coefficient lower than required. Therefore, our empirical results have important implications for structural and cyclical policymakers in Tunisia to promote economic growth in a context of lower unemployment growth.

Keywords: Okun’s law, validity, unit root, cointegration, error correction model, bandpass filters

Procedia PDF Downloads 317