Search results for: time prediction algorithms
19148 Existence and Uniqueness of Solutions to Singular Higher Order Two-Point BVPs on Time Scales
Authors: Zhenjie Liu
Abstract:
This paper investigates the existence and uniqueness of solutions for singular higher order boundary value problems on time scales by using mixed monotone method. The theorems obtained are very general. For the different time scale, the problem may be the corresponding continuous or discrete boundary value problem.Keywords: mixed monotone operator, boundary value problem, time scale, green's function, positive solution, singularity
Procedia PDF Downloads 25719147 An Adiabatic Quantum Optimization Approach for the Mixed Integer Nonlinear Programming Problem
Authors: Maxwell Henderson, Tristan Cook, Justin Chan Jin Le, Mark Hodson, YoungJung Chang, John Novak, Daniel Padilha, Nishan Kulatilaka, Ansu Bagchi, Sanjoy Ray, John Kelly
Abstract:
We present a method of using adiabatic quantum optimization (AQO) to solve a mixed integer nonlinear programming (MINLP) problem instance. The MINLP problem is a general form of a set of NP-hard optimization problems that are critical to many business applications. It requires optimizing a set of discrete and continuous variables with nonlinear and potentially nonconvex constraints. Obtaining an exact, optimal solution for MINLP problem instances of non-trivial size using classical computation methods is currently intractable. Current leading algorithms leverage heuristic and divide-and-conquer methods to determine approximate solutions. Creating more accurate and efficient algorithms is an active area of research. Quantum computing (QC) has several theoretical benefits compared to classical computing, through which QC algorithms could obtain MINLP solutions that are superior to current algorithms. AQO is a particular form of QC that could offer more near-term benefits compared to other forms of QC, as hardware development is in a more mature state and devices are currently commercially available from D-Wave Systems Inc. It is also designed for optimization problems: it uses an effect called quantum tunneling to explore all lowest points of an energy landscape where classical approaches could become stuck in local minima. Our work used a novel algorithm formulated for AQO to solve a special type of MINLP problem. The research focused on determining: 1) if the problem is possible to solve using AQO, 2) if it can be solved by current hardware, 3) what the currently achievable performance is, 4) what the performance will be on projected future hardware, and 5) when AQO is likely to provide a benefit over classical computing methods. Two different methods, integer range and 1-hot encoding, were investigated for transforming the MINLP problem instance constraints into a mathematical structure that can be embedded directly onto the current D-Wave architecture. For testing and validation a D-Wave 2X device was used, as well as QxBranch’s QxLib software library, which includes a QC simulator based on simulated annealing. Our results indicate that it is mathematically possible to formulate the MINLP problem for AQO, but that currently available hardware is unable to solve problems of useful size. Classical general-purpose simulated annealing is currently able to solve larger problem sizes, but does not scale well and such methods would likely be outperformed in the future by improved AQO hardware with higher qubit connectivity and lower temperatures. If larger AQO devices are able to show improvements that trend in this direction, commercially viable solutions to the MINLP for particular applications could be implemented on hardware projected to be available in 5-10 years. Continued investigation into optimal AQO hardware architectures and novel methods for embedding MINLP problem constraints on to those architectures is needed to realize those commercial benefits.Keywords: adiabatic quantum optimization, mixed integer nonlinear programming, quantum computing, NP-hard
Procedia PDF Downloads 52619146 Study on Measuring Method and Experiment of Arc Fault Detection Device
Authors: Yang Jian-Hong, Zhang Ren-Cheng, Huang Li
Abstract:
Arc fault is one of the main inducements of electric fires. Arc Fault Detection Device (AFDD) can detect arc fault effectively. Arc fault detections and unhooking standards are the keys to AFDD practical application. First, an arc fault continuous production system was developed, which could count the arc half wave number. Then, Combining with the UL1699 standard, ignition probability curve of cotton and unhooking time of various currents intensity were obtained by experiments. The combustion degree of arc fault could be expressed effectively by arc area. Experiments proved that electric fires would be misjudged or missed only using arc half wave number as AFDD unhooking basis. At last, Practical tests were carried out on the self-developed AFDD system. The result showed that actual AFDD unhooking time was the sum of arc half wave cycling number, Arc wave identification time and unhooking mechanical operation time And the first two shared shorter time. Unhooking time standard depended on the shortest mechanical operation time.Keywords: arc fault detection device, arc area, arc half wave, unhooking time, arc fault
Procedia PDF Downloads 50919145 Decision Support: How Explainable A.I. Can Improve Transparency and Trust with Human Users
Authors: Devon Brown, Liu Chunmei
Abstract:
This paper will present an analysis as part of the researchers dissertation topic focusing on the intersection of affective and analytical directed acyclic graphs (DAGs) in the context of Decision Support Systems (DSS). The researcher’s work involves analyzing decision theory models like Affective and Bayesian Decision theory models and how they could be implemented under an Affective Computing Framework using Information Fusion and Human-Centered Design. Additionally, the researcher is beginning research on an Affective-Analytic Decision Framework (AADF) model for their dissertation research and are looking to merge logic and analytic models with empathetic insights into affective DAGs. Data-collection efforts begin Fall 2024 and in preparation for the efforts this paper looks to analyze previous research in this area and introduce the AADF framework and propose conceptual models for consideration. For this paper, the research emphasis is placed on analyzing Bayesian networks and Markov models which offer probabilistic techniques during uncertainty in decision-making. Ideally, including affect into analytic models will ensure algorithms can increase user trust with algorithms by including emotional states and the user’s experience with the goal of developing emotionally intelligent A.I. systems that can start to navigate the complex fabric of human emotion during decision-making.Keywords: decision support systems, explainable AI, HCAI techniques, affective-analytical decision framework
Procedia PDF Downloads 2019144 Optimal Solutions for Real-Time Scheduling of Reconfigurable Embedded Systems Based on Neural Networks with Minimization of Power Consumption
Authors: Ghofrane Rehaiem, Hamza Gharsellaoui, Samir Benahmed
Abstract:
In this study, Artificial Neural Networks (ANNs) were used for modeling the parameters that allow the real-time scheduling of embedded systems under resources constraints designed for real-time applications running. The objective of this work is to implement a neural networks based approach for real-time scheduling of embedded systems in order to handle real-time constraints in execution scenarios. In our proposed approach, many techniques have been proposed for both the planning of tasks and reducing energy consumption. In fact, a combination of Dynamic Voltage Scaling (DVS) and time feedback can be used to scale the frequency dynamically adjusting the operating voltage. Indeed, we present in this paper a hybrid contribution that handles the real-time scheduling of embedded systems, low power consumption depending on the combination of DVS and Neural Feedback Scheduling (NFS) with the energy Priority Earlier Deadline First (PEDF) algorithm. Experimental results illustrate the efficiency of our original proposed approach.Keywords: optimization, neural networks, real-time scheduling, low-power consumption
Procedia PDF Downloads 37119143 In vitro Estimation of Genotoxic Lesions in Peripheral Blood Lymphocytes of Rat Exposed to Organophosphate Pesticides
Authors: A. Ojha, Y. K. Gupta
Abstract:
Organophosphate (OP) pesticides are among the most widely used synthetic chemicals for controlling a wide variety of pests throughout the world. Chlorpyrifos (CPF), methyl parathion (MPT), and malathion (MLT) are among the most extensively used OP pesticides in India. DNA strand breaks and DNA-protein crosslinks (DPC) are toxic lesions associated with the mechanisms of toxicity of genotoxic compounds. In the present study, we have examined the potential of CPF, MPT, and MLT individually and in combination, to cause DNA strand breakage and DPC formation. Peripheral blood lymphocytes of rat were exposed to 1/4 and 1/10 LC50 dose of CPF, MPT, and MLT for 2, 4, 8, and 12h. The DNA strand break was measured by the comet assay and expressed as DNA damage index while DPC estimation was done by fluorescence emission. There was significantly marked increase in DNA damage and DNA-protein crosslink formation in time and dose dependent manner. It was also observed that MPT caused the highest level of DNA damage as compared to other studied OP compounds. Thus, from present study, we can conclude that studied pesticides have genotoxic potential. The pesticides mixture does not potentiate the toxicity of each other. Nonetheless, additional in vivo data are required before a definitive conclusion can be drawn regarding hazard prediction to humans.Keywords: organophosphate, pesticides, DNA damage, DNA protein crosslink, genotoxic
Procedia PDF Downloads 35619142 Optimal Placement and Sizing of Energy Storage System in Distribution Network with Photovoltaic Based Distributed Generation Using Improved Firefly Algorithms
Authors: Ling Ai Wong, Hussain Shareef, Azah Mohamed, Ahmad Asrul Ibrahim
Abstract:
The installation of photovoltaic based distributed generation (PVDG) in active distribution system can lead to voltage fluctuation due to the intermittent and unpredictable PVDG output power. This paper presented a method in mitigating the voltage rise by optimally locating and sizing the battery energy storage system (BESS) in PVDG integrated distribution network. The improved firefly algorithm is used to perform optimal placement and sizing. Three objective functions are presented considering the voltage deviation and BESS off-time with state of charge as the constraint. The performance of the proposed method is compared with another optimization method such as the original firefly algorithm and gravitational search algorithm. Simulation results show that the proposed optimum BESS location and size improve the voltage stability.Keywords: BESS, firefly algorithm, PVDG, voltage fluctuation
Procedia PDF Downloads 32119141 Temporal and Spatial Distribution Prediction of Patinopecten yessoensis Larvae in Northern China Yellow Sea
Authors: RuiJin Zhang, HengJiang Cai, JinSong Gui
Abstract:
It takes Patinopecten yessoensis larvae more than 20 days from spawning to settlement. Due to the natural environmental factors such as current, Patinopecten yessoensis larvae are transported to a distance more than hundreds of kilometers, leading to a high instability of their spatial and temporal distribution and great difficulties in the natural spat collection. Therefore predicting the distribution is of great significance to improve the operating efficiency of the collecting. Hydrodynamic model of Northern China Yellow Sea was established and the motions equations of physical oceanography and verified by the tidal harmonic constants and the measured data velocities of Dalian Bay. According to the passivity drift characteristics of the larvae, combined with the hydrodynamic model and the particle tracking model, the spatial and temporal distribution prediction model was established and the spatial and temporal distribution of the larvae under the influence of flow and wind were simulated. It can be concluded from the model results: ocean currents have greatest impacts on the passive drift path and diffusion of Patinopecten yessoensis larvae; the impact of wind is also important, which changed the direction and speed of the drift. Patinopecten yessoensis larvae were generated in the sea along Zhangzi Island and Guanglu-Dachangshan Island, but after two months, with the impact of wind and currents, the larvae appeared in the west of Dalian and the southern of Lvshun, and even in Bohai Bay. The model results are consistent with the relevant literature on qualitative analysis, and this conclusion explains where the larvae come from in the perspective of numerical simulation.Keywords: numerical simulation, Patinopecten yessoensis larvae, predicting model, spatial and temporal distribution
Procedia PDF Downloads 30519140 Hybrid Intelligent Optimization Methods for Optimal Design of Horizontal-Axis Wind Turbine Blades
Authors: E. Tandis, E. Assareh
Abstract:
Designing the optimal shape of MW wind turbine blades is provided in a number of cases through evolutionary algorithms associated with mathematical modeling (Blade Element Momentum Theory). Evolutionary algorithms, among the optimization methods, enjoy many advantages, particularly in stability. However, they usually need a large number of function evaluations. Since there are a large number of local extremes, the optimization method has to find the global extreme accurately. The present paper introduces a new population-based hybrid algorithm called Genetic-Based Bees Algorithm (GBBA). This algorithm is meant to design the optimal shape for MW wind turbine blades. The current method employs crossover and neighborhood searching operators taken from the respective Genetic Algorithm (GA) and Bees Algorithm (BA) to provide a method with good performance in accuracy and speed convergence. Different blade designs, twenty-one to be exact, were considered based on the chord length, twist angle and tip speed ratio using GA results. They were compared with BA and GBBA optimum design results targeting the power coefficient and solidity. The results suggest that the final shape, obtained by the proposed hybrid algorithm, performs better compared to either BA or GA. Furthermore, the accuracy and speed convergence increases when the GBBA is employedKeywords: Blade Design, Optimization, Genetic Algorithm, Bees Algorithm, Genetic-Based Bees Algorithm, Large Wind Turbine
Procedia PDF Downloads 31619139 A Three Elements Vector Valued Structure’s Ultimate Strength-Strong Motion-Intensity Measure
Authors: A. Nicknam, N. Eftekhari, A. Mazarei, M. Ganjvar
Abstract:
This article presents an alternative collapse capacity intensity measure in the three elements form which is influenced by the spectral ordinates at periods longer than that of the first mode period at near and far source sites. A parameter, denoted by β, is defined by which the spectral ordinate effects, up to the effective period (2T_1), on the intensity measure are taken into account. The methodology permits to meet the hazard-levelled target extreme event in the probabilistic and deterministic forms. A MATLAB code is developed involving OpenSees to calculate the collapse capacities of the 8 archetype RC structures having 2 to 20 stories for regression process. The incremental dynamic analysis (IDA) method is used to calculate the structure’s collapse values accounting for the element stiffness and strength deterioration. The general near field set presented by FEMA is used in a series of performing nonlinear analyses. 8 linear relationships are developed for the 8structutres leading to the correlation coefficient up to 0.93. A collapse capacity near field prediction equation is developed taking into account the results of regression processes obtained from the 8 structures. The proposed prediction equation is validated against a set of actual near field records leading to a good agreement. Implementation of the proposed equation to the four archetype RC structures demonstrated different collapse capacities at near field site compared to those of FEMA. The reasons of differences are believed to be due to accounting for the spectral shape effects.Keywords: collapse capacity, fragility analysis, spectral shape effects, IDA method
Procedia PDF Downloads 23919138 Human Immune Response to Surgery: The Surrogate Prediction of Postoperative Outcomes
Authors: Husham Bayazed
Abstract:
Immune responses following surgical trauma play a pivotal role in predicting postoperative outcomes from healing and recovery to postoperative complications. Postoperative complications, including infections and protracted recovery, occur in a significant number of about 300 million surgeries performed annually worldwide. Complications cause personal suffering along with a significant economic burden on the healthcare system in any community. The accurate prediction of postoperative complications and patient-targeted interventions for their prevention remain major clinical provocations. Recent Findings: Recent studies are focusing on immune dysregulation mechanisms that occur in response to surgical trauma as a key determinant of postoperative complications. Antecedent studies mainly were plunging into the detection of inflammatory plasma markers, which facilitate in providing important clues regarding their pathogenesis. However, recent Single-cell technologies, such as mass cytometry or single-cell RNA sequencing, have markedly enhanced our ability to understand the immunological basis of postoperative immunological trauma complications and to identify their prognostic biological signatures. Summary: The advent of proteomic technologies has significantly advanced our ability to predict the risk of postoperative complications. Multiomic modeling of patients' immune states holds promise for the discovery of preoperative predictive biomarkers and providing patients and surgeons with information to improve surgical outcomes. However, more studies are required to accurately predict the risk of postoperative complications in individual patients.Keywords: immune dysregulation, postoperative complications, surgical trauma, flow cytometry
Procedia PDF Downloads 8619137 Studying the Temperature Field of Hypersonic Vehicle Structure with Aero-Thermo-Elasticity Deformation
Authors: Geng Xiangren, Liu Lei, Gui Ye-Wei, Tang Wei, Wang An-ling
Abstract:
The malfunction of thermal protection system (TPS) caused by aerodynamic heating is a latent trouble to aircraft structure safety. Accurately predicting the structure temperature field is quite important for the TPS design of hypersonic vehicle. Since Thornton’s work in 1988, the coupled method of aerodynamic heating and heat transfer has developed rapidly. However, little attention has been paid to the influence of structural deformation on aerodynamic heating and structural temperature field. In the flight, especially the long-endurance flight, the structural deformation, caused by the aerodynamic heating and temperature rise, has a direct impact on the aerodynamic heating and structural temperature field. Thus, the coupled interaction cannot be neglected. In this paper, based on the method of static aero-thermo-elasticity, considering the influence of aero-thermo-elasticity deformation, the aerodynamic heating and heat transfer coupled results of hypersonic vehicle wing model were calculated. The results show that, for the low-curvature region, such as fuselage or center-section wing, structure deformation has little effect on temperature field. However, for the stagnation region with high curvature, the coupled effect is not negligible. Thus, it is quite important for the structure temperature prediction to take into account the effect of elastic deformation. This work has laid a solid foundation for improving the prediction accuracy of the temperature distribution of aircraft structures and the evaluation capacity of structural performance.Keywords: aerothermoelasticity, elastic deformation, structural temperature, multi-field coupling
Procedia PDF Downloads 34119136 Accidental Compartment Fire Dynamics: Experiment, Computational Fluid Dynamics Weakness and Expert Interview Analysis
Authors: Timothy Onyenobi
Abstract:
Accidental fires and its dynamic as it relates to building compartmentation and the impact of the compartment morphology, is still an on-going area of study; especially with the use of computational fluid dynamics (CFD) modeling methods. With better knowledge on this subject come better solution recommendations by fire engineers. Interviews were carried out for this study where it was identified that the response perspectives to accidental fire were different with the fire engineer providing qualitative data which is based on “what is expected in real fires” and the fire fighters provided information on “what actually obtains in real fires”. This further led to a study and analysis of two real and comprehensively instrumented fire experiments: the Open Plan Office Project by National Institute of Standard and Technology (NIST) USA (to study time to flashover) and the TF2000 project by the Building Research Establishment (BRE) UK (to test for conformity with Building Regulation requirements). The findings from the analysis of the experiments revealed the relative yet critical weakness of fire prediction using a CFD model (usually used by fire engineers) as well as explained the differences in response perspectives of the fire engineers and firefighters from the interview analysis.Keywords: CFD, compartment fire, experiment, fire fighters, fire engineers
Procedia PDF Downloads 33819135 A Low Order Thermal Envelope Model for Heat Transfer Characteristics of Low-Rise Residential Buildings
Authors: Nadish Anand, Richard D. Gould
Abstract:
A simplistic model is introduced for determining the thermal characteristics of a Low-rise Residential (LRR) building and then predicts the energy usage by its Heating Ventilation & Air Conditioning (HVAC) system according to changes in weather conditions which are reflected in the Ambient Temperature (Outside Air Temperature). The LRR buildings are treated as a simple lump for solving the heat transfer problem and the model is derived using the lumped capacitance model of transient conduction heat transfer from bodies. Since most contemporary HVAC systems have a thermostat control which will have an offset temperature and user defined set point temperatures which define when the HVAC system will switch on and off. The aim is to predict without any error the Body Temperature (i.e. the Inside Air Temperature) which will estimate the switching on and off of the HVAC system. To validate the mathematical model derived from lumped capacitance we have used EnergyPlus simulation engine, which simulates Buildings with considerable accuracy. We have predicted through the low order model the Inside Air Temperature of a single house kept in three different climate zones (Detroit, Raleigh & Austin) and different orientations for summer and winter seasons. The prediction error from the model for the same day as that of model parameter calculation has showed an error of < 10% in winter for almost all the orientations and climate zones. Whereas the prediction error is only <10% for all the orientations in the summer season for climate zone at higher latitudes (Raleigh & Detroit). Possible factors responsible for the large variations are also noted in the work, paving way for future research.Keywords: building energy, energy consumption, energy+, HVAC, low order model, lumped capacitance
Procedia PDF Downloads 26619134 Unlocking Green Hydrogen Potential: A Machine Learning-Based Assessment
Authors: Said Alshukri, Mazhar Hussain Malik
Abstract:
Green hydrogen is hydrogen produced using renewable energy sources. In the last few years, Oman aimed to reduce its dependency on fossil fuels. Recently, the hydrogen economy has become a global trend, and many countries have started to investigate the feasibility of implementing this sector. Oman created an alliance to establish the policy and rules for this sector. With motivation coming from both global and local interest in green hydrogen, this paper investigates the potential of producing hydrogen from wind and solar energies in three different locations in Oman, namely Duqm, Salalah, and Sohar. By using machine learning-based software “WEKA” and local metrological data, the project was designed to figure out which location has the highest wind and solar energy potential. First, various supervised models were tested to obtain their prediction accuracy, and it was found that the Random Forest (RF) model has the best prediction performance. The RF model was applied to 2021 metrological data for each location, and the results indicated that Duqm has the highest wind and solar energy potential. The system of one wind turbine in Duqm can produce 8335 MWh/year, which could be utilized in the water electrolysis process to produce 88847 kg of hydrogen mass, while a solar system consisting of 2820 solar cells is estimated to produce 1666.223 MWh/ year which is capable of producing 177591 kg of hydrogen mass.Keywords: green hydrogen, machine learning, wind and solar energies, WEKA, supervised models, random forest
Procedia PDF Downloads 7919133 Partial Least Square Regression for High-Dimentional and High-Correlated Data
Authors: Mohammed Abdullah Alshahrani
Abstract:
The research focuses on investigating the use of partial least squares (PLS) methodology for addressing challenges associated with high-dimensional correlated data. Recent technological advancements have led to experiments producing data characterized by a large number of variables compared to observations, with substantial inter-variable correlations. Such data patterns are common in chemometrics, where near-infrared (NIR) spectrometer calibrations record chemical absorbance levels across hundreds of wavelengths, and in genomics, where thousands of genomic regions' copy number alterations (CNA) are recorded from cancer patients. PLS serves as a widely used method for analyzing high-dimensional data, functioning as a regression tool in chemometrics and a classification method in genomics. It handles data complexity by creating latent variables (components) from original variables. However, applying PLS can present challenges. The study investigates key areas to address these challenges, including unifying interpretations across three main PLS algorithms and exploring unusual negative shrinkage factors encountered during model fitting. The research presents an alternative approach to addressing the interpretation challenge of predictor weights associated with PLS. Sparse estimation of predictor weights is employed using a penalty function combining a lasso penalty for sparsity and a Cauchy distribution-based penalty to account for variable dependencies. The results demonstrate sparse and grouped weight estimates, aiding interpretation and prediction tasks in genomic data analysis. High-dimensional data scenarios, where predictors outnumber observations, are common in regression analysis applications. Ordinary least squares regression (OLS), the standard method, performs inadequately with high-dimensional and highly correlated data. Copy number alterations (CNA) in key genes have been linked to disease phenotypes, highlighting the importance of accurate classification of gene expression data in bioinformatics and biology using regularized methods like PLS for regression and classification.Keywords: partial least square regression, genetics data, negative filter factors, high dimensional data, high correlated data
Procedia PDF Downloads 4919132 Interpretation of the Russia-Ukraine 2022 War via N-Gram Analysis
Authors: Elcin Timur Cakmak, Ayse Oguzlar
Abstract:
This study presents the results of the tweets sent by Twitter users on social media about the Russia-Ukraine war by bigram and trigram methods. On February 24, 2022, Russian President Vladimir Putin declared a military operation against Ukraine, and all eyes were turned to this war. Many people living in Russia and Ukraine reacted to this war and protested and also expressed their deep concern about this war as they felt the safety of their families and their futures were at stake. Most people, especially those living in Russia and Ukraine, express their views on the war in different ways. The most popular way to do this is through social media. Many people prefer to convey their feelings using Twitter, one of the most frequently used social media tools. Since the beginning of the war, it is seen that there have been thousands of tweets about the war from many countries of the world on Twitter. These tweets accumulated in data sources are extracted using various codes for analysis through Twitter API and analysed by Python programming language. The aim of the study is to find the word sequences in these tweets by the n-gram method, which is known for its widespread use in computational linguistics and natural language processing. The tweet language used in the study is English. The data set consists of the data obtained from Twitter between February 24, 2022, and April 24, 2022. The tweets obtained from Twitter using the #ukraine, #russia, #war, #putin, #zelensky hashtags together were captured as raw data, and the remaining tweets were included in the analysis stage after they were cleaned through the preprocessing stage. In the data analysis part, the sentiments are found to present what people send as a message about the war on Twitter. Regarding this, negative messages make up the majority of all the tweets as a ratio of %63,6. Furthermore, the most frequently used bigram and trigram word groups are found. Regarding the results, the most frequently used word groups are “he, is”, “I, do”, “I, am” for bigrams. Also, the most frequently used word groups are “I, do, not”, “I, am, not”, “I, can, not” for trigrams. In the machine learning phase, the accuracy of classifications is measured by Classification and Regression Trees (CART) and Naïve Bayes (NB) algorithms. The algorithms are used separately for bigrams and trigrams. We gained the highest accuracy and F-measure values by the NB algorithm and the highest precision and recall values by the CART algorithm for bigrams. On the other hand, the highest values for accuracy, precision, and F-measure values are achieved by the CART algorithm, and the highest value for the recall is gained by NB for trigrams.Keywords: classification algorithms, machine learning, sentiment analysis, Twitter
Procedia PDF Downloads 7319131 Comparative Study and Parallel Implementation of Stochastic Models for Pricing of European Options Portfolios using Monte Carlo Methods
Authors: Vinayak Bassi, Rajpreet Singh
Abstract:
Over the years, with the emergence of sophisticated computers and algorithms, finance has been quantified using computational prowess. Asset valuation has been one of the key components of quantitative finance. In fact, it has become one of the embryonic steps in determining risk related to a portfolio, the main goal of quantitative finance. This study comprises a drawing comparison between valuation output generated by two stochastic dynamic models, namely Black-Scholes and Dupire’s bi-dimensionality model. Both of these models are formulated for computing the valuation function for a portfolio of European options using Monte Carlo simulation methods. Although Monte Carlo algorithms have a slower convergence rate than calculus-based simulation techniques (like FDM), they work quite effectively over high-dimensional dynamic models. A fidelity gap is analyzed between the static (historical) and stochastic inputs for a sample portfolio of underlying assets. In order to enhance the performance efficiency of the model, the study emphasized the use of variable reduction methods and customizing random number generators to implement parallelization. An attempt has been made to further implement the Dupire’s model on a GPU to achieve higher computational performance. Furthermore, ideas have been discussed around the performance enhancement and bottleneck identification related to the implementation of options-pricing models on GPUs.Keywords: monte carlo, stochastic models, computational finance, parallel programming, scientific computing
Procedia PDF Downloads 16219130 Prediction and Optimization of Machining Induced Residual Stresses in End Milling of AISI 1045 Steel
Authors: Wajid Ali Khan
Abstract:
Extensive experimentation and numerical investigation are performed to predict the machining-induced residual stresses in the end milling of AISI 1045 steel, and an optimization code has been developed using the particle swarm optimization technique. Experiments were conducted using a single factor at a time and design of experiments approach. Regression analysis was done, and a mathematical model of the cutting process was developed, thus predicting the machining-induced residual stress with reasonable accuracy. The mathematical model served as the objective function to be optimized using particle swarm optimization. The relationship between the different cutting parameters and the output variables, force, and residual stresses has been studied. The combined effect of the process parameters, speed, feed, and depth of cut was examined, and it is understood that 85% of the variation of these variables can be attributed to these machining parameters under research. A 3D finite element model is developed to predict the cutting forces and the machining-induced residual stresses in end milling operation. The results were validated experimentally and against the Johnson-cook model available in the literature.Keywords: residual stresses, end milling, 1045 steel, optimization
Procedia PDF Downloads 10219129 Indexing and Incremental Approach Using Map Reduce Bipartite Graph (MRBG) for Mining Evolving Big Data
Authors: Adarsh Shroff
Abstract:
Big data is a collection of dataset so large and complex that it becomes difficult to process using data base management tools. To perform operations like search, analysis, visualization on big data by using data mining; which is the process of extraction of patterns or knowledge from large data set. In recent years, the data mining applications become stale and obsolete over time. Incremental processing is a promising approach to refreshing mining results. It utilizes previously saved states to avoid the expense of re-computation from scratch. This project uses i2MapReduce, an incremental processing extension to Map Reduce, the most widely used framework for mining big data. I2MapReduce performs key-value pair level incremental processing rather than task level re-computation, supports not only one-step computation but also more sophisticated iterative computation, which is widely used in data mining applications, and incorporates a set of novel techniques to reduce I/O overhead for accessing preserved fine-grain computation states. To optimize the mining results, evaluate i2MapReduce using a one-step algorithm and three iterative algorithms with diverse computation characteristics for efficient mining.Keywords: big data, map reduce, incremental processing, iterative computation
Procedia PDF Downloads 35119128 Land Suitability Prediction Modelling for Agricultural Crops Using Machine Learning Approach: A Case Study of Khuzestan Province, Iran
Authors: Saba Gachpaz, Hamid Reza Heidari
Abstract:
The sharp increase in population growth leads to more pressure on agricultural areas to satisfy the food supply. To achieve this, more resources should be consumed and, besides other environmental concerns, highlight sustainable agricultural development. Land-use management is a crucial factor in obtaining optimum productivity. Machine learning is a widely used technique in the agricultural sector, from yield prediction to customer behavior. This method focuses on learning and provides patterns and correlations from our data set. In this study, nine physical control factors, namely, soil classification, electrical conductivity, normalized difference water index (NDWI), groundwater level, elevation, annual precipitation, pH of water, annual mean temperature, and slope in the alluvial plain in Khuzestan (an agricultural hotspot in Iran) are used to decide the best agricultural land use for both rainfed and irrigated agriculture for ten different crops. For this purpose, each variable was imported into Arc GIS, and a raster layer was obtained. In the next level, by using training samples, all layers were imported into the python environment. A random forest model was applied, and the weight of each variable was specified. In the final step, results were visualized using a digital elevation model, and the importance of all factors for each one of the crops was obtained. Our results show that despite 62% of the study area being allocated to agricultural purposes, only 42.9% of these areas can be defined as a suitable class for cultivation purposes.Keywords: land suitability, machine learning, random forest, sustainable agriculture
Procedia PDF Downloads 8419127 Task Evoked Pupillary Response for Surgical Task Difficulty Prediction via Multitask Learning
Authors: Beilei Xu, Wencheng Wu, Lei Lin, Rachel Melnyk, Ahmed Ghazi
Abstract:
In operating rooms, excessive cognitive stress can impede the performance of a surgeon, while low engagement can lead to unavoidable mistakes due to complacency. As a consequence, there is a strong desire in the surgical community to be able to monitor and quantify the cognitive stress of a surgeon while performing surgical procedures. Quantitative cognitiveload-based feedback can also provide valuable insights during surgical training to optimize training efficiency and effectiveness. Various physiological measures have been evaluated for quantifying cognitive stress for different mental challenges. In this paper, we present a study using the cognitive stress measured by the task evoked pupillary response extracted from the time series eye-tracking measurements to predict task difficulties in a virtual reality based robotic surgery training environment. In particular, we proposed a differential-task-difficulty scale, utilized a comprehensive feature extraction approach, and implemented a multitask learning framework and compared the regression accuracy between the conventional single-task-based and three multitask approaches across subjects.Keywords: surgical metric, task evoked pupillary response, multitask learning, TSFresh
Procedia PDF Downloads 14619126 Harmonizing Cities: Integrating Land Use Diversity and Multimodal Transit for Social Equity
Authors: Zi-Yan Chao
Abstract:
With the rapid development of urbanization and increasing demand for efficient transportation systems, the interaction between land use diversity and transportation resource allocation has become a critical issue in urban planning. Achieving a balance of land use types, such as residential, commercial, and industrial areas, is crucial role in ensuring social equity and sustainable urban development. Simultaneously, optimizing multimodal transportation networks, including bus, subway, and car routes, is essential for minimizing total travel time and costs, while ensuring fairness for all social groups, particularly in meeting the transportation needs of low-income populations. This study develops a bilevel programming model to address these challenges, with land use diversity as the foundation for measuring equity. The upper-level model maximizes land use diversity for balanced land distribution across regions. The lower-level model optimizes multimodal transportation networks to minimize travel time and costs while maintaining user equilibrium. The model also incorporates constraints to ensure fair resource allocation, such as balancing transportation accessibility and cost differences across various social groups. A solution approach is developed to solve the bilevel optimization problem, ensuring efficient exploration of the solution space for land use and transportation resource allocation. This study maximizes social equity by maximizing land use diversity and achieving user equilibrium with optimal transportation resource distribution. The proposed method provides a robust framework for addressing urban planning challenges, contributing to sustainable and equitable urban development.Keywords: bilevel programming model, genetic algorithms, land use diversity, multimodal transportation optimization, social equity
Procedia PDF Downloads 2319125 Investigation of Time Pressure and Instinctive Reaction in Moral Dilemmas While Driving
Authors: Jacqueline Miller, Dongyuan Y. Wang, F. Dan Richard
Abstract:
Before trying to make an ethical machine that holds a higher ethical standard than humans, a better understanding of human moral standards that could be used as a guide is crucial. How humans make decisions in dangerous driving situations like moral dilemmas can contribute to developing acceptable ethical principles for autonomous vehicles (AVs). This study uses a driving simulator to investigate whether drivers make utilitarian choices (choices that maximize lives saved and minimize harm) in unavoidable automobile accidents (moral dilemmas) with time pressure manipulated. This study also investigates how impulsiveness influences drivers’ behavior in moral dilemmas. Manipulating time pressure results in collisions that occur at varying time intervals (4 s, 5 s, 7s). Manipulating time pressure helps investigate how time pressure may influence drivers’ response behavior. Thirty-one undergraduates participated in this study using a STISM driving simulator to respond to driving moral dilemmas. The results indicated that the percentage of utilitarian choices generally increased when given more time to respond (from 4 s to 7 s). Additionally, participants in vehicle scenarios preferred responding right over responding left. Impulsiveness did not influence utilitarian choices. However, as time pressure decreased, response time increased. Findings have potential implications and applications on the regulation of driver assistance technologies and AVs.Keywords: time pressure, automobile moral dilemmas, impulsiveness, reaction time
Procedia PDF Downloads 5419124 Implementation of CNV-CH Algorithm Using Map-Reduce Approach
Authors: Aishik Deb, Rituparna Sinha
Abstract:
We have developed an algorithm to detect the abnormal segment/"structural variation in the genome across a number of samples. We have worked on simulated as well as real data from the BAM Files and have designed a segmentation algorithm where abnormal segments are detected. This algorithm aims to improve the accuracy and performance of the existing CNV-CH algorithm. The next-generation sequencing (NGS) approach is very fast and can generate large sequences in a reasonable time. So the huge volume of sequence information gives rise to the need for Big Data and parallel approaches of segmentation. Therefore, we have designed a map-reduce approach for the existing CNV-CH algorithm where a large amount of sequence data can be segmented and structural variations in the human genome can be detected. We have compared the efficiency of the traditional and map-reduce algorithms with respect to precision, sensitivity, and F-Score. The advantages of using our algorithm are that it is fast and has better accuracy. This algorithm can be applied to detect structural variations within a genome, which in turn can be used to detect various genetic disorders such as cancer, etc. The defects may be caused by new mutations or changes to the DNA and generally result in abnormally high or low base coverage and quantification values.Keywords: cancer detection, convex hull segmentation, map reduce, next generation sequencing
Procedia PDF Downloads 13619123 A Study of Behavioral Phenomena Using an Artificial Neural Network
Authors: Yudhajit Datta
Abstract:
Will is a phenomenon that has puzzled humanity for a long time. It is a belief that Will Power of an individual affects the success achieved by an individual in life. It is thought that a person endowed with great will power can overcome even the most crippling setbacks of life while a person with a weak will cannot make the most of life even the greatest assets. Behavioral aspects of the human experience such as will are rarely subjected to quantitative study owing to the numerous uncontrollable parameters involved. This work is an attempt to subject the phenomena of will to the test of an artificial neural network. The claim being tested is that will power of an individual largely determines success achieved in life. In the study, an attempt is made to incorporate the behavioral phenomenon of will into a computational model using data pertaining to the success of individuals obtained from an experiment. A neural network is to be trained using data based upon part of the model, and subsequently used to make predictions regarding will corresponding to data points of success. If the prediction is in agreement with the model values, the model is to be retained as a candidate. Ultimately, the best-fit model from among the many different candidates is to be selected, and used for studying the correlation between success and will.Keywords: will power, will, success, apathy factor, random factor, characteristic function, life story
Procedia PDF Downloads 38019122 Adaptive Discharge Time Control for Battery Operation Time Enhancement
Authors: Jong-Bae Lee, Seongsoo Lee
Abstract:
This paper proposes an adaptive discharge time control method to balance cell voltages in alternating battery cell discharging method. In the alternating battery cell discharging method, battery cells are periodically discharged in turn. Recovery effect increases battery output voltage while the given battery cell rests without discharging, thus battery operation time of target system increases. However, voltage mismatch between cells leads two problems. First, voltage difference between cells induces inter-cell current with wasted power. Second, it degrades battery operation time, since system stops when any cell reaches to the minimum system operation voltage. To solve this problem, the proposed method adaptively controls cell discharge time to equalize both cell voltages. In the proposed method, battery operation time increases about 19%, while alternating battery cell discharging method shows about 7% improvement.Keywords: battery, recovery effect, low-power, alternating battery cell discharging, adaptive discharge time control
Procedia PDF Downloads 35219121 Stability of Hybrid Stochastic Systems
Authors: Manlika Ratchagit
Abstract:
This paper is concerned with robust mean square stability of uncertain stochastic switched discrete time-delay systems. The system to be considered is subject to interval time-varying delays, which allows the delay to be a fast time-varying function and the lower bound is not restricted to zero. Based on the discrete Lyapunov functional, a switching rule for the robust mean square stability for the uncertain stochastic discrete time-delay system is designed via linear matrix inequalities. Finally, some examples are exploited to illustrate the effectiveness of the proposed schemes.Keywords: robust mean square stability, discrete-time stochastic systems, hybrid systems, interval time-varying delays, Lyapunov functional, linear matrix inequalities
Procedia PDF Downloads 48619120 Development of a Real-Time Simulink Based Robotic System to Study Force Feedback Mechanism during Instrument-Object Interaction
Authors: Jaydip M. Desai, Antonio Valdevit, Arthur Ritter
Abstract:
Robotic surgery is used to enhance minimally invasive surgical procedure. It provides greater degree of freedom for surgical tools but lacks of haptic feedback system to provide sense of touch to the surgeon. Surgical robots work on master-slave operation, where user is a master and robotic arms are the slaves. Current, surgical robots provide precise control of the surgical tools, but heavily rely on visual feedback, which sometimes cause damage to the inner organs. The goal of this research was to design and develop a real-time simulink based robotic system to study force feedback mechanism during instrument-object interaction. Setup includes three Velmex XSlide assembly (XYZ Stage) for three dimensional movement, an end effector assembly for forceps, electronic circuit for four strain gages, two Novint Falcon 3D gaming controllers, microcontroller board with linear actuators, MATLAB and Simulink toolboxes. Strain gages were calibrated using Imada Digital Force Gauge device and tested with a hard-core wire to measure instrument-object interaction in the range of 0-35N. Designed simulink model successfully acquires 3D coordinates from two Novint Falcon controllers and transfer coordinates to the XYZ stage and forceps. Simulink model also reads strain gages signal through 10-bit analog to digital converter resolution of a microcontroller assembly in real time, converts voltage into force and feedback the output signals to the Novint Falcon controller for force feedback mechanism. Experimental setup allows user to change forward kinematics algorithms to achieve the best-desired movement of the XYZ stage and forceps. This project combines haptic technology with surgical robot to provide sense of touch to the user controlling forceps through machine-computer interface.Keywords: surgical robot, haptic feedback, MATLAB, strain gage, simulink
Procedia PDF Downloads 53419119 New Results on Stability of Hybrid Stochastic Systems
Authors: Manlika Rajchakit
Abstract:
This paper is concerned with robust mean square stability of uncertain stochastic switched discrete time-delay systems. The system to be considered is subject to interval time-varying delays, which allows the delay to be a fast time-varying function and the lower bound is not restricted to zero. Based on the discrete Lyapunov functional, a switching rule for the robust mean square stability for the uncertain stochastic discrete time-delay system is designed via linear matrix inequalities. Finally, some examples are exploited to illustrate the effectiveness of the proposed schemes.Keywords: robust mean square stability, discrete-time stochastic systems, hybrid systems, interval time-varying delays, lyapunov functional, linear matrix inequalities
Procedia PDF Downloads 430