Search results for: equivalent linear model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19014

Search results for: equivalent linear model

17844 Optimization Approach to Integrated Production-Inventory-Routing Problem for Oxygen Supply Chains

Authors: Yena Lee, Vassilis M. Charitopoulos, Karthik Thyagarajan, Ian Morris, Jose M. Pinto, Lazaros G. Papageorgiou

Abstract:

With globalisation, the need to have better coordination of production and distribution decisions has become increasingly important for industrial gas companies in order to remain competitive in the marketplace. In this work, we investigate a problem that integrates production, inventory, and routing decisions in a liquid oxygen supply chain. The oxygen supply chain consists of production facilities, external third-party suppliers, and multiple customers, including hospitals and industrial customers. The product produced by the plants or sourced from the competitors, i.e., third-party suppliers, is distributed by a fleet of heterogenous vehicles to satisfy customer demands. The objective is to minimise the total operating cost involving production, third-party, and transportation costs. The key decisions for production include production and inventory levels and product amount from third-party suppliers. In contrast, the distribution decisions involve customer allocation, delivery timing, delivery amount, and vehicle routing. The optimisation of the coordinated production, inventory, and routing decisions is a challenging problem, especially when dealing with large-size problems. Thus, we present a two-stage procedure to solve the integrated problem efficiently. First, the problem is formulated as a mixed-integer linear programming (MILP) model by simplifying the routing component. The solution from the first-stage MILP model yields the optimal customer allocation, production and inventory levels, and delivery timing and amount. Then, we fix the previous decisions and solve a detailed routing. In the second stage, we propose a column generation scheme to address the computational complexity of the resulting detailed routing problem. A case study considering a real-life oxygen supply chain in the UK is presented to illustrate the capability of the proposed models and solution method. Furthermore, a comparison of the solutions from the proposed approach with the corresponding solutions provided by existing metaheuristic techniques (e.g., guided local search and tabu search algorithms) is presented to evaluate the efficiency.

Keywords: production planning, inventory routing, column generation, mixed-integer linear programming

Procedia PDF Downloads 99
17843 Detection of High Fructose Corn Syrup in Honey by Near Infrared Spectroscopy and Chemometrics

Authors: Mercedes Bertotto, Marcelo Bello, Hector Goicoechea, Veronica Fusca

Abstract:

The National Service of Agri-Food Health and Quality (SENASA), controls honey to detect contamination by synthetic or natural chemical substances and establishes and controls the traceability of the product. The utility of near-infrared spectroscopy for the detection of adulteration of honey with high fructose corn syrup (HFCS) was investigated. First of all, a mixture of different authentic artisanal Argentinian honey was prepared to cover as much heterogeneity as possible. Then, mixtures were prepared by adding different concentrations of high fructose corn syrup (HFCS) to samples of the honey pool. 237 samples were used, 108 of them were authentic honey and 129 samples corresponded to honey adulterated with HFCS between 1 and 10%. They were stored unrefrigerated from time of production until scanning and were not filtered after receipt in the laboratory. Immediately prior to spectral collection, honey was incubated at 40°C overnight to dissolve any crystalline material, manually stirred to achieve homogeneity and adjusted to a standard solids content (70° Brix) with distilled water. Adulterant solutions were also adjusted to 70° Brix. Samples were measured by NIR spectroscopy in the range of 650 to 7000 cm⁻¹. The technique of specular reflectance was used, with a lens aperture range of 150 mm. Pretreatment of the spectra was performed by Standard Normal Variate (SNV). The ant colony optimization genetic algorithm sample selection (ACOGASS) graphical interface was used, using MATLAB version 5.3, to select the variables with the greatest discriminating power. The data set was divided into a validation set and a calibration set, using the Kennard-Stone (KS) algorithm. A combined method of Potential Functions (PF) was chosen together with Partial Least Square Linear Discriminant Analysis (PLS-DA). Different estimators of the predictive capacity of the model were compared, which were obtained using a decreasing number of groups, which implies more demanding validation conditions. The optimal number of latent variables was selected as the number associated with the minimum error and the smallest number of unassigned samples. Once the optimal number of latent variables was defined, we proceeded to apply the model to the training samples. With the calibrated model for the training samples, we proceeded to study the validation samples. The calibrated model that combines the potential function methods and PLSDA can be considered reliable and stable since its performance in future samples is expected to be comparable to that achieved for the training samples. By use of Potential Functions (PF) and Partial Least Square Linear Discriminant Analysis (PLS-DA) classification, authentic honey and honey adulterated with HFCS could be identified with a correct classification rate of 97.9%. The results showed that NIR in combination with the PT and PLS-DS methods can be a simple, fast and low-cost technique for the detection of HFCS in honey with high sensitivity and power of discrimination.

Keywords: adulteration, multivariate analysis, potential functions, regression

Procedia PDF Downloads 110
17842 Neighborhood Graph-Optimized Preserving Discriminant Analysis for Image Feature Extraction

Authors: Xiaoheng Tan, Xianfang Li, Tan Guo, Yuchuan Liu, Zhijun Yang, Hongye Li, Kai Fu, Yufang Wu, Heling Gong

Abstract:

The image data collected in reality often have high dimensions, and it contains noise and redundant information. Therefore, it is necessary to extract the compact feature expression of the original perceived image. In this process, effective use of prior knowledge such as data structure distribution and sample label is the key to enhance image feature discrimination and robustness. Based on the above considerations, this paper proposes a local preserving discriminant feature learning model based on graph optimization. The model has the following characteristics: (1) Locality preserving constraint can effectively excavate and preserve the local structural relationship between data. (2) The flexibility of graph learning can be improved by constructing a new local geometric structure graph using label information and the nearest neighbor threshold. (3) The L₂,₁ norm is used to redefine LDA, and the diagonal matrix is introduced as the scale factor of LDA, and the samples are selected, which improves the robustness of feature learning. The validity and robustness of the proposed algorithm are verified by experiments in two public image datasets.

Keywords: feature extraction, graph optimization local preserving projection, linear discriminant analysis, L₂, ₁ norm

Procedia PDF Downloads 136
17841 Globally Convergent Sequential Linear Programming for Multi-Material Topology Optimization Using Ordered Solid Isotropic Material with Penalization Interpolation

Authors: Darwin Castillo Huamaní, Francisco A. M. Gomes

Abstract:

The aim of the multi-material topology optimization (MTO) is to obtain the optimal topology of structures composed by many materials, according to a given set of constraints and cost criteria. In this work, we seek the optimal distribution of materials in a domain, such that the flexibility of the structure is minimized, under certain boundary conditions and the intervention of external forces. In the case we have only one material, each point of the discretized domain is represented by two values from a function, where the value of the function is 1 if the element belongs to the structure or 0 if the element is empty. A common way to avoid the high computational cost of solving integer variable optimization problems is to adopt the Solid Isotropic Material with Penalization (SIMP) method. This method relies on the continuous interpolation function, power function, where the base variable represents a pseudo density at each point of domain. For proper exponent values, the SIMP method reduces intermediate densities, since values other than 0 or 1 usually does not have a physical meaning for the problem. Several extension of the SIMP method were proposed for the multi-material case. The one that we explore here is the ordered SIMP method, that has the advantage of not being based on the addition of variables to represent material selection, so the computational cost is independent of the number of materials considered. Although the number of variables is not increased by this algorithm, the optimization subproblems that are generated at each iteration cannot be solved by methods that rely on second derivatives, due to the cost of calculating the second derivatives. To overcome this, we apply a globally convergent version of the sequential linear programming method, which solves a linear approximation sequence of optimization problems.

Keywords: globally convergence, multi-material design ordered simp, sequential linear programming, topology optimization

Procedia PDF Downloads 297
17840 Definition of a Computing Independent Model and Rules for Transformation Focused on the Model-View-Controller Architecture

Authors: Vanessa Matias Leite, Jandira Guenka Palma, Flávio Henrique de Oliveira

Abstract:

This paper presents a model-oriented development approach to software development in the Model-View-Controller (MVC) architectural standard. This approach aims to expose a process of extractions of information from the models, in which through rules and syntax defined in this work, assists in the design of the initial model and its future conversions. The proposed paper presents a syntax based on the natural language, according to the rules agreed in the classic grammar of the Portuguese language, added to the rules of conversions generating models that follow the norms of the Object Management Group (OMG) and the Meta-Object Facility MOF.

Keywords: BNF Syntax, model driven architecture, model-view-controller, transformation, UML

Procedia PDF Downloads 378
17839 Nonlinear Finite Element Analysis of Optimally Designed Steel Angelina™ Beams

Authors: Ferhat Erdal, Osman Tunca, Serkan Tas, Serdar Carbas

Abstract:

Web-expanded steel beams provide an easy and economical solution for the systems having longer structural members. The main goal of manufacturing these beams is to increase the moment of inertia and section modulus, which results in greater strength and rigidity. Until recently, there were two common types of open web-expanded beams: with hexagonal openings, also called castellated beams, and beams with circular openings referred to as cellular beams, until the generation of sinusoidal web-expanded beams. In the present research, the optimum design of a new generation beams, namely sinusoidal web-expanded beams, will be carried out and the design results will be compared with castellated and cellular beam solutions. Thanks to a reduced fabrication process and substantial material savings, the web-expanded beam with sinusoidal holes (Angelina™ Beam) meets the economic requirements of steel design problems while ensuring optimum safety. The objective of this research is to carry out non-linear finite element analysis (FEA) of the web-expanded beam with sinusoidal holes. The FE method has been used to predict their entire response to increasing values of external loading until they lose their load carrying capacity. FE model of each specimen that is utilized in the experimental studies is carried out. These models are used to simulate the experimental work to verify of test results and to investigate the non-linear behavior of failure modes such as web-post buckling, shear buckling and vierendeel bending of beams.

Keywords: steel structures, web-expanded beams, angelina beam, optimum design, failure modes, finite element analysis

Procedia PDF Downloads 266
17838 Investigating the Effect of Different Design Factors on the Required Length of the Ambient Air Vaporizer

Authors: F. S. Alavi

Abstract:

In this study, MATLAB engineering software was used in order to model an industrial Ambient Air Vaporizer (AAV), considering combined convection and conduction heat transfers from the fins and the tube. The developed theoretical model was then used to investigate the effects of various design factors such as gas flow rate, ambient air temperature, fin thickness and etc. on total vaporizer ‘s length required. Cryogenic liquid nitrogen was selected as an input fluid, in all cases. According to the results, increasing the inlet fluid flow rate has direct linear effect on the total required length of vaporizer. Vaporizer’s required length decreases by increasing the size of fin radius or size of fin thickness. The dependency of vaporizer’s length on fin thickness’ size reduces at higher values of thickness and gradually converge to zero. For low flow rates, internal convection heat transfer coefficient depends directly on gas flow rate but it becomes constant, independent on flow rate after a specific value. As the ambient air temperature increases, the external heat transfer coefficient also increases and the total required length of vaporizer decreases.

Keywords: heat exchanger, modeling, heat transfer, design

Procedia PDF Downloads 102
17837 Enhancing Secondary School Mathematics Retention with Blended Learning: Integrating Concepts for Improved Understanding

Authors: Felix Oromena Egara, Moeketsi Mosia

Abstract:

The study aimed to evaluate the impact of blended learning on mathematics retention among secondary school students. Conducted in the Isoko North Local Government Area of Delta State, Nigeria, the research involved 1,235 senior class one (SS 1) students. Employing a non-equivalent control group pre-test-post-test quasi-experimental design, a sample of 70 students was selected from two secondary schools with ICT facilities through purposive sampling. Random allocation of students into experimental and control groups was achieved through balloting within each selected school. The investigation included three assessment points: pre-Mathematics Achievement Test (MAT), post-MAT, and post-post-MAT (retention), administered systematically by the researchers. Data collection utilized the established MAT instrument, which demonstrated a high reliability score of 0.86. Statistical analysis was conducted using the Statistical Package for Social Sciences (SPSS) version 28, with mean and standard deviation addressing study questions and analysis of covariance scrutinizing hypotheses at a significance level of .05. Results revealed significantly greater improvements in mathematics retention scores among students exposed to blended learning compared to those instructed through conventional methods. Moreover, noticeable differences in mean retention scores were observed, with male students in the blended learning group exhibiting notably higher performance. Based on these findings, recommendations were made, advocating for mathematics educators to integrate blended learning, particularly in geometry teaching, to enhance students’ retention of mathematical concepts.

Keywords: blended learning, flipped classroom model, secondary school students, station rotation model

Procedia PDF Downloads 24
17836 1D Velocity Model for the Gobi-Altai Region from Local Earthquakes

Authors: Dolgormaa Munkhbaatar, Munkhsaikhan Adiya, Tseedulam Khuut

Abstract:

We performed an inversion method to determine the 1D-velocity model with station corrections of the Gobi-Altai area in the southern part of Mongolia using earthquake data collected in the National Data Center during the last 10 years. In this study, the concept of the new 1D model has been employed to minimize the average RMS of a set of well-located earthquakes, recorded at permanent (between 2006 and 2016) and temporary seismic stations (between 2014 and 2016), compute solutions for the coupled hypocenter and 1D velocity model. We selected 4800 events with RMS less than 0.5 seconds and with a maximum GAP of 170 degrees and determined velocity structures. Also, we relocated all possible events located in the Gobi-Altai area using the new 1D velocity model and achieved constrained hypocentral determinations for events within this area. We concluded that the estimated new 1D velocity model is a relatively low range compared to the previous velocity model in a significant improvement intend to, and the quality of the information basis for future research center locations to determine the earthquake epicenter area with this new transmission model.

Keywords: 1D velocity model, earthquake, relocation, Velest

Procedia PDF Downloads 146
17835 Investigation of the Material Behaviour of Polymeric Interlayers in Broken Laminated Glass

Authors: Martin Botz, Michael Kraus, Geralt Siebert

Abstract:

The use of laminated glass gains increasing importance in structural engineering. For safety reasons, at least two glass panes are laminated together with a polymeric interlayer. In case of breakage of one or all of the glass panes, the glass fragments are still connected to the interlayer due to adhesion forces and a certain residual load-bearing capacity is left in the system. Polymer interlayers used in the laminated glass show a viscoelastic material behavior, e.g. stresses and strains in the interlayer are dependent on load duration and temperature. In the intact stage only small strains appear in the interlayer, thus the material can be described in a linear way. In the broken stage, large strains can appear and a non-linear viscoelasticity material theory is necessary. Relaxation tests on two different types of polymeric interlayers are performed at different temperatures and strain amplitudes to determine the border to the non-linear material regime. Based on the small-scale specimen results further tests on broken laminated glass panes are conducted. So-called ‘through-crack-bending’ (TCB) tests are performed, in which the laminated glass has a defined crack pattern. The test set-up is realized in a way that one glass layer is still able to transfer compressive stresses but tensile stresses have to be transferred by the interlayer solely. The TCB-tests are also conducted under different temperatures but constant force (creep test). Aims of these experiments are to elaborate if the results of small-scale tests on the interlayer are transferable to a laminated glass system in the broken stage. In this study, limits of the applicability of linear-viscoelasticity are established in the context of two commercially available polymer-interlayers. Furthermore, it is shown that the results of small-scale tests agree to a certain degree to the results of the TCB large-scale experiments. In a future step, the results can be used to develop material models for the post breakage performance of laminated glass.

Keywords: glass breakage, laminated glass, relaxation test, viscoelasticity

Procedia PDF Downloads 112
17834 Digital Phase Shifting Holography in a Non-Linear Interferometer using Undetected Photons

Authors: Sebastian Töpfer, Marta Gilaberte Basset, Jorge Fuenzalida, Fabian Steinlechner, Juan P. Torres, Markus Gräfe

Abstract:

This work introduces a combination of digital phase-shifting holography with a non-linear interferometer using undetected photons. Non-linear interferometers can be used in combination with a measurement scheme called quantum imaging with undetected photons, which allows for the separation of the wavelengths used for sampling an object and detecting it in the imaging sensor. This method recently faced increasing attention, as it allows to use of exotic wavelengths (e.g., mid-infrared, ultraviolet) for object interaction while at the same time keeping the detection in spectral areas with highly developed, comparable low-cost imaging sensors. The object information, including its transmission and phase influence, is recorded in the form of an interferometric pattern. To collect these, this work combines the method of quantum imaging with undetected photons with digital phase-shifting holography with a minimal sampling of the interference. With this, the quantum imaging scheme gets extended in its measurement capabilities and brings it one step closer to application. Quantum imaging with undetected photons uses correlated photons generated by spontaneous parametric down-conversion in a non-linear interferometer to create indistinguishable photon pairs, which leads to an effect called induced coherence without induced emission. Placing an object inside changes the interferometric pattern depending on the object’s properties. Digital phase-shifting holography records multiple images of the interference with determined phase shifts to reconstruct the complete interference shape, which can afterward be used to analyze the changes introduced by the object and conclude its properties. An extensive characterization of this method was done using a proof-of-principle setup. The measured spatial resolution, phase accuracy, and transmission accuracy are compared for different combinations of camera exposure times and the number of interference sampling steps. The current limits of this method are shown to allow further improvements. To summarize, this work presents an alternative holographic measurement method using non-linear interferometers in combination with quantum imaging to enable new ways of measuring and motivating continuing research.

Keywords: digital holography, quantum imaging, quantum holography, quantum metrology

Procedia PDF Downloads 79
17833 Evaluation of Cyclic Thermo-Mechanical Responses of an Industrial Gas Turbine Rotor

Authors: Y. Rae, A. Benaarbia, J. Hughes, Wei Sun

Abstract:

This paper describes an elasto-visco-plastic computational modelling method which can be used to assess the cyclic plasticity responses of high temperature structures operating under thermo-mechanical loadings. The material constitutive equation used is an improved unified multi-axial Chaboche-Lemaitre model, which takes into account non-linear kinematic and isotropic hardening. The computational methodology is a three-dimensional framework following an implicit formulation and based on a radial return mapping algorithm. The associated user material (UMAT) code is developed and calibrated across isothermal hold-time low cycle fatigue tests for a typical turbine rotor steel for use in finite element (FE) implementation. The model is applied to a realistic industrial gas turbine rotor, where the study focuses its attention on the deformation heterogeneities and critical high stress areas within the rotor structure. The potential improvements of such FE visco-plastic approach are discussed. An integrated life assessment procedure based on R5 and visco-plasticity modelling, is also briefly addressed.

Keywords: unified visco-plasticity, thermo-mechanical, turbine rotor, finite element modelling

Procedia PDF Downloads 117
17832 Simulation of Wind Generator with Fixed Wind Turbine under Matlab-Simulink

Authors: Mahdi Motahari, Mojtaba Farzaneh, Armin Parsian Nejad

Abstract:

The rapidly growing wind industry is highly expressing the need for education and training worldwide, particularly on the system level. Modelling and simulating wind generator system using Matlab-Simulink provides expert help in understanding wind systems engineering and system design. Working under Matlab-Simulink we present the integration of the developed WECS model with public electrical grid. A test of the calculated power and Cp related to the experimental equivalent data, using statistical analysis is performed. The statistical indicators of accuracy show better results of the presented method with RMSE: 21%, 22%, MBE : 0.77%, 0.12 % and MAE :3%, 4%.On the other hand we study its behavior when integrated in whole power system. Three level of wind speeds have been chosen: low with 5m/s as the mean value, medium with 8m/s as the mean value and high speed with 12m/s as the mean value. These allowed predicting and supervising the active power produced by the system, characterized respectively by the middle powers of -150 kW, -250kW and -480 kW which will be injected directly into the public electrical grid and the reactive power, characterized respectively by the middle powers of 60 kW, 180 kW and 320 kW and will be consumed by the wind generator.

Keywords: modelling, simulation, wind generator, fixed speed wind turbine, Matlab-Simulink

Procedia PDF Downloads 610
17831 A New Reliability Allocation Method Based on Fuzzy Numbers

Authors: Peng Li, Chuanri Li, Tao Li

Abstract:

Reliability allocation is quite important during early design and development stages for a system to apportion its specified reliability goal to subsystems. This paper improves the reliability fuzzy allocation method and gives concrete processes on determining the factor set, the factor weight set, judgment set, and multi-grade fuzzy comprehensive evaluation. To determine the weight of factor set, the modified trapezoidal numbers are proposed to reduce errors caused by subjective factors. To decrease the fuzziness in the fuzzy division, an approximation method based on linear programming is employed. To compute the explicit values of fuzzy numbers, centroid method of defuzzification is considered. An example is provided to illustrate the application of the proposed reliability allocation method based on fuzzy arithmetic.

Keywords: reliability allocation, fuzzy arithmetic, allocation weight, linear programming

Procedia PDF Downloads 326
17830 Performance Analysis of Modified Solar Water Heating System for Climatic Condition of Allahabad, India

Authors: Kirti Tewari, Rahul Dev

Abstract:

Solar water heating is a thermodynamic process of heating water using sunlight with the help of solar water heater. Thus, solar water heater is a device used to harness solar energy. In this paper, a modified solar water heating system (MSWHS) has been proposed over flat plate collector (FPC) and Evacuated tube collector (ETC). The modifications include selection of materials other than glass, and glass wool which are conventionally used for fabricating FPC and ETC. Some modifications in design have also been proposed. Its collector is made of double layer of semi-cylindrical acrylic tubes and fibre reinforced plastic (FRP) insulation base. Water tank is made of double layer of acrylic sheet except base and north wall. FRP is used in base and north wall of the water tank. A concept of equivalent thickness has been utilised for calculating the dimensions of collector plate, acrylic tube and tank. A thermal model for the proposed design of MSWHS is developed and simulation is carried out on MATLAB for the capacity of 200L MSWHS having collector area of 1.6 m2, length of acrylic tubes of 2m at an inclination angle 25° which is taken nearly equal to the latitude of the given location. Latitude of Allahabad is 24.45° N. The results show that the maximum temperature of water in tank and tube has been found to be 71.2°C and 73.3°C at 17:00hr and 16:00hr respectively in March for the climatic data of Allahabad. Theoretical performance analysis has been carried out by varying number of tubes of collector, the tank capacity and climatic data for given months of winter and summer.

Keywords: acrylic, fibre reinforced plastic, solar water heating, thermal model, conventional water heaters

Procedia PDF Downloads 326
17829 A Bivariate Inverse Generalized Exponential Distribution and Its Applications in Dependent Competing Risks Model

Authors: Fatemah A. Alqallaf, Debasis Kundu

Abstract:

The aim of this paper is to introduce a bivariate inverse generalized exponential distribution which has a singular component. The proposed bivariate distribution can be used when the marginals have heavy-tailed distributions, and they have non-monotone hazard functions. Due to the presence of the singular component, it can be used quite effectively when there are ties in the data. Since it has four parameters, it is a very flexible bivariate distribution, and it can be used quite effectively for analyzing various bivariate data sets. Several dependency properties and dependency measures have been obtained. The maximum likelihood estimators cannot be obtained in closed form, and it involves solving a four-dimensional optimization problem. To avoid that, we have proposed to use an EM algorithm, and it involves solving only one non-linear equation at each `E'-step. Hence, the implementation of the proposed EM algorithm is very straight forward in practice. Extensive simulation experiments and the analysis of one data set have been performed. We have observed that the proposed bivariate inverse generalized exponential distribution can be used for modeling dependent competing risks data. One data set has been analyzed to show the effectiveness of the proposed model.

Keywords: Block and Basu bivariate distributions, competing risks, EM algorithm, Marshall-Olkin bivariate exponential distribution, maximum likelihood estimators

Procedia PDF Downloads 122
17828 An Elbow Biomechanical Model and Its Coefficients Adjustment

Authors: Jie Bai, Yongsheng Gao, Shengxin Wang, Jie Zhao

Abstract:

Through the establishment of the elbow biomechanical model, it can provide theoretical guide for rehabilitation therapy on the upper limb of the human body. A biomechanical model of the elbow joint can be built by the connection of muscle force model and elbow dynamics. But there are many undetermined coefficients in the model like the optimal joint angle and optimal muscle force which are usually specified as the experimental parameters of other workers. Because of the individual differences, there is a certain deviation of the final result. To this end, the RMS value of the deviation between the actual angle and calculated angle is considered. A set of coefficients which lead to the minimum RMS value will be chosen to be the optimal parameters. The direct search method and the conjugacy search method are used to get the optimal parameters, thus the model can be more accurate and mode adaptability.

Keywords: elbow biomechanical model, RMS, direct search, conjugacy search

Procedia PDF Downloads 528
17827 Forecasting for Financial Stock Returns Using a Quantile Function Model

Authors: Yuzhi Cai

Abstract:

In this paper, we introduce a newly developed quantile function model that can be used for estimating conditional distributions of financial returns and for obtaining multi-step ahead out-of-sample predictive distributions of financial returns. Since we forecast the whole conditional distributions, any predictive quantity of interest about the future financial returns can be obtained simply as a by-product of the method. We also show an application of the model to the daily closing prices of Dow Jones Industrial Average (DJIA) series over the period from 2 January 2004 - 8 October 2010. We obtained the predictive distributions up to 15 days ahead for the DJIA returns, which were further compared with the actually observed returns and those predicted from an AR-GARCH model. The results show that the new model can capture the main features of financial returns and provide a better fitted model together with improved mean forecasts compared with conventional methods. We hope this talk will help audience to see that this new model has the potential to be very useful in practice.

Keywords: DJIA, financial returns, predictive distribution, quantile function model

Procedia PDF Downloads 354
17826 Surveying Energy Dissipation in Stepped Spillway Using Finite Element Modeling

Authors: Mehdi Fuladipanah

Abstract:

Stepped spillway includes several steps from the crest to the toe. The steps of stepped spillway could cause to decrease the energy with making energy distribution in the longitude mode and also to reduce the outcome speed. The aim of this study was to stimulate the stepped spillway combined with stilling basin-step using Fluent model and the turbulent superficial flow using RNG, K-ε. The free surface of the flow was monitored by VOF model. The velocity and the depth of the flow were measured by tail water depth by the numerical model and then the dissipated energy was calculated along the spillway. The results indicated that the stilling basin-step complex may cause energy dissipation increment in the stepped spillway. Also, the numerical model was suggested as an effective method to predict the circular and complicated flows in the stepped spillways.

Keywords: stepped spillway, fluent model, VOF model, K-ε model, energy distribution

Procedia PDF Downloads 356
17825 Evaluating the Dosimetric Performance for 3D Treatment Planning System for Wedged and Off-Axis Fields

Authors: Nashaat A. Deiab, Aida Radwan, Mohamed S. Yahiya, Mohamed Elnagdy, Rasha Moustafa

Abstract:

This study is to evaluate the dosimetric performance of our institution's 3D treatment planning system for wedged and off-axis 6MV photon beams, guided by the recommended QA tests documented in the AAPM TG53; NCS report 15 test packages, IAEA TRS 430 and ESTRO booklet no.7. The study was performed for Elekta Precise linear accelerator designed for clinical range of 4, 6 and 15 MV photon beams with asymmetric jaws and fully integrated multileaf collimator that enables high conformance to target with sharp field edges. Ten tests were applied on solid water equivalent phantom along with 2D array dose detection system. The calculated doses using 3D treatment planning system PrecisePLAN were compared with measured doses to make sure that the dose calculations are accurate for simple situations such as square and elongated fields, different SSD, beam modifiers e.g. wedges, blocks, MLC-shaped fields and asymmetric collimator settings. The QA results showed dosimetric accuracy of the TPS within the specified tolerance limits. Except for large elongated wedged field, the central axis and outside central axis have errors of 0.2% and 0.5%, respectively, and off- planned and off-axis elongated fields the region outside the central axis of the beam errors are 0.2% and 1.1%, respectively. The dosimetric investigated results yielded differences within the accepted tolerance level as recommended. Differences between dose values predicted by the TPS and measured values at the same point are the result from limitations of the dose calculation, uncertainties in the measurement procedure, or fluctuations in the output of the accelerator.

Keywords: quality assurance, dose calculation, wedged fields, off-axis fields, 3D treatment planning system, photon beam

Procedia PDF Downloads 426
17824 Enhancing Transfer Path Analysis with In-Situ Component Transfer Path Analysis for Interface Forces Identification

Authors: Raef Cherif, Houssine Bakkali, Wafaa El Khatiri, Yacine Yaddaden

Abstract:

The analysis of how vibrations are transmitted between components is required in many engineering applications. Transfer path analysis (TPA) has been a valuable engineering tool for solving Noise, Vibration, and Harshness (NVH problems using sub-structuring applications. The most challenging part of a TPA analysis is estimating the equivalent forces at the contact points between the active and the passive side. Component TPA in situ Method calculates these forces by inverting the frequency response functions (FRFs) measured at the passive subsystem, relating the motion at indicator points to forces at the interface. However, matrix inversion could pose problems due to the ill-conditioning of the matrices leading to inaccurate results. This paper establishes a TPA model for an academic system consisting of two plates linked by four springs. A numerical study has been performed to improve the interface forces identification. Several parameters are studied and discussed, such as the singular value rejection and the number and position of indicator points chosen and used in the inversion matrix.

Keywords: transfer path analysis, matrix inverse method, indicator points, SVD decomposition

Procedia PDF Downloads 66
17823 Monitoring Three-Dimensional Models of Tree and Forest by Using Digital Close-Range Photogrammetry

Authors: S. Y. Cicekli

Abstract:

In this study, tree-dimensional model of tree was created by using terrestrial close range photogrammetry. For this close range photos were taken. Photomodeler Pro 5 software was used for camera calibration and create three-dimensional model of trees. In first test, three-dimensional model of a tree was created, in the second test three-dimensional model of three trees were created. This study aim is creating three-dimensional model of trees and indicate the use of close-range photogrammetry in forestry. At the end of the study, three-dimensional model of tree and three trees were created. This study showed that usability of close-range photogrammetry for monitoring tree and forests three-dimensional model.

Keywords: close- range photogrammetry, forest, tree, three-dimensional model

Procedia PDF Downloads 376
17822 A Mathematical-Based Formulation of EEG Fluctuations

Authors: Razi Khalafi

Abstract:

Brain is the information processing center of the human body. Stimuli in form of information are transferred to the brain and then brain makes the decision on how to respond to them. In this research we propose a new partial differential equation which analyses the EEG signals and make a relationship between the incoming stimuli and the brain response to them. In order to test the proposed model, a set of external stimuli applied to the model and the model’s outputs were checked versus the real EEG data. The results show that this model can model the EEG signal well. The proposed model is useful not only for modeling of the EEG signal in case external stimuli but it can be used for the modeling of brain response in case of internal stimuli.

Keywords: Brain, stimuli, partial differential equation, response, eeg signal

Procedia PDF Downloads 413
17821 Performance and Availability Analysis of 2N Redundancy Models

Authors: Yutae Lee

Abstract:

In this paper, we consider the performance and availability of a redundancy model. The redundancy model is a form of resilience that ensures service availability in the event of component failure. This paper considers a 2N redundancy model. In the model there are at most one active service unit and at most one standby service unit. The active one is providing the service while the standby is prepared to take over the active role when the active fails. We design our analysis model using Stochastic Reward Nets, and then evaluate the performance and availability of 2N redundancy model using Stochastic Petri Net Package (SPNP).

Keywords: availability, performance, stochastic reward net, 2N redundancy

Procedia PDF Downloads 402
17820 Stability of Hybrid Systems

Authors: Kreangkri Ratchagit

Abstract:

This paper is concerned with exponential stability of switched linear systems with interval time-varying delays. The time delay is any continuous function belonging to a given interval, in which the lower bound of delay is not restricted to zero. By constructing a suitable augmented Lyapunov-Krasovskii functional combined with Leibniz-Newton’s formula, a switching rule for the exponential stability of switched linear systems with interval time-varying delays and new delay-dependent sufficient conditions for the exponential stability of the systems are first established in terms of LMIs. Finally, some examples are exploited to illustrate the effectiveness of the proposed schemes.

Keywords: exponential stability, hybrid systems, timevarying delays, Lyapunov-Krasovskii functional, Leibniz-Newton’s formula

Procedia PDF Downloads 438
17819 Antioxidant Activity of the Methanolic Extract and Antimicrobial Activity of the Essential Oil of Rosmarinus officinalis L. Grown in Algeria

Authors: Nassim Belkacem, Amina Azzam, Dalila Haouchine, Kahina Bennacer, Samira Soufit

Abstract:

Objective: To evaluate the antioxidant activity of the methanolic extract along with the antimicrobial activity of the essential oil of the aerial parts of Rosmarinus officinalis L. collected in the region of Bejaia (northern center of Algeria). Materials and methods: The polyphenols and flavonoids contents of the methanolic extract were measured. The antioxidant activity was evaluated using two methods: the ABTS method and DPPH assay. The antimicrobial activity was studied by the agar diffusion method against five bacterial strains (Three Gram positive strains and two Gram negative strains) and one fungus. Results: The total polyphenol and flavonoid content was about 43.8 mg gallic acid equivalent per gram (GA Eq/g) and 7.04 mg quercetin equivalent per gram (Q Eq/g), respectively. In the ABTS assay, the rosemary extract has shown an inhibition of 98.02% at the concentration of 500ug/ml with a half maximal inhibitory concentration value (IC50) of 194.92ug/ml. The results of DPPH assay have shown that the rosemary extract has an inhibition of 94.67 % with an IC50 value of 17.87ug/ml, which is lower than that of Butylhydroxyanisol (BHA) about 6.03ug/ml and ascorbic acid about 1.24μg/ml. The yield in essential oil of rosemary obtained by hydrodistillation was 1.42%. Based on the determination of the diameter of inhibition, different antimicrobial activity of the essential oil was revealed against the six tested microbes. Escherichia coli from the University Hospital (UH), Streptococcus aureus (UH) and Pseudomonas aeruginosa ATCC have a minimum inhibitory concentration value (MIC) of 62.5µl/ml. However, Bacillus sp (UH) and Staphylococcus aureus ATCC have an MIC value of 125μl/ml. The inhibition zone against Candida sp was about 24 mm. The aromatograms showed that the essential oil of rosemary exercises an antifungal activity more important than the antibacterial one.

Keywords: Rosmarinus officinalis L., maceration, essential oil, antioxidant, antimicrobial activity

Procedia PDF Downloads 501
17818 A Mathematical Equation to Calculate Stock Price of Different Growth Model

Authors: Weiping Liu

Abstract:

This paper presents an equation to calculate stock prices of different growth model. This equation is mathematically derived by using discounted cash flow method. It has the advantages of being very easy to use and very accurate. It can still be used even when the first stage is lengthy. This equation is more generalized because it can be used for all the three popular stock price models. It can be programmed into financial calculator or electronic spreadsheets. In addition, it can be extended to a multistage model. It is more versatile and efficient than the traditional methods.

Keywords: stock price, multistage model, different growth model, discounted cash flow method

Procedia PDF Downloads 388
17817 Comparison of Different Machine Learning Algorithms for Solubility Prediction

Authors: Muhammet Baldan, Emel Timuçin

Abstract:

Molecular solubility prediction plays a crucial role in various fields, such as drug discovery, environmental science, and material science. In this study, we compare the performance of five machine learning algorithms—linear regression, support vector machines (SVM), random forests, gradient boosting machines (GBM), and neural networks—for predicting molecular solubility using the AqSolDB dataset. The dataset consists of 9981 data points with their corresponding solubility values. MACCS keys (166 bits), RDKit properties (20 properties), and structural properties(3) features are extracted for every smile representation in the dataset. A total of 189 features were used for training and testing for every molecule. Each algorithm is trained on a subset of the dataset and evaluated using metrics accuracy scores. Additionally, computational time for training and testing is recorded to assess the efficiency of each algorithm. Our results demonstrate that random forest model outperformed other algorithms in terms of predictive accuracy, achieving an 0.93 accuracy score. Gradient boosting machines and neural networks also exhibit strong performance, closely followed by support vector machines. Linear regression, while simpler in nature, demonstrates competitive performance but with slightly higher errors compared to ensemble methods. Overall, this study provides valuable insights into the performance of machine learning algorithms for molecular solubility prediction, highlighting the importance of algorithm selection in achieving accurate and efficient predictions in practical applications.

Keywords: random forest, machine learning, comparison, feature extraction

Procedia PDF Downloads 21
17816 A Continuous Boundary Value Method of Order 8 for Solving the General Second Order Multipoint Boundary Value Problems

Authors: T. A. Biala

Abstract:

This paper deals with the numerical integration of the general second order multipoint boundary value problems. This has been achieved by the development of a continuous linear multistep method (LMM). The continuous LMM is used to construct a main discrete method to be used with some initial and final methods (also obtained from the continuous LMM) so that they form a discrete analogue of the continuous second order boundary value problems. These methods are used as boundary value methods and adapted to cope with the integration of the general second order multipoint boundary value problems. The convergence, the use and the region of absolute stability of the methods are discussed. Several numerical examples are implemented to elucidate our solution process.

Keywords: linear multistep methods, boundary value methods, second order multipoint boundary value problems, convergence

Procedia PDF Downloads 361
17815 Reconstructed Phase Space Features for Estimating Post Traumatic Stress Disorder

Authors: Andre Wittenborn, Jarek Krajewski

Abstract:

Trauma-related sadness in speech can alter the voice in several ways. The generation of non-linear aerodynamic phenomena within the vocal tract is crucial when analyzing trauma-influenced speech production. They include non-laminar flow and formation of jets rather than well-behaved laminar flow aspects. Especially state-space reconstruction methods based on chaotic dynamics and fractal theory have been suggested to describe these aerodynamic turbulence-related phenomena of the speech production system. To extract the non-linear properties of the speech signal, we used the time delay embedding method to reconstruct from a scalar time series (reconstructed phase space, RPS). This approach results in the extraction of 7238 Features per .wav file (N= 47, 32 m, 15 f). The speech material was prompted by telling about autobiographical related sadness-inducing experiences (sampling rate 16 kHz, 8-bit resolution). After combining these features in a support vector machine based machine learning approach (leave-one-sample out validation), we achieved a correlation of r = .41 with the well-established, self-report ground truth measure (RATS) of post-traumatic stress disorder (PTSD).

Keywords: non-linear dynamics features, post traumatic stress disorder, reconstructed phase space, support vector machine

Procedia PDF Downloads 92