Search results for: predictive modeling
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4730

Search results for: predictive modeling

4070 Atmospheric Transport Modeling of Radio-Xenon Detections Possibly Related to the Announced Nuclear Test in North Korea on February 12, 2013

Authors: Kobi Kutsher

Abstract:

On February 12th 2013, monitoring stations of the Preparatory Commission of the Comprehensive Nuclear Test-Ban Treaty Organization (CTBTO) detected a seismic event with explosion-like underground characteristics in the Democratic People’s Republic of Korea (DPRK). The location was found to be in the vicinity of the two previous announced nuclear tests in 2006 and 2009.The nuclear test was also announced by the government of the DPRK.After an underground nuclear explosion, radioactive fission products (mostly noble gases) can seep through layers of rock and sediment until they escape into the atmosphere. The fission products are dispersed in the atmosphere and may be detected thousands of kilometers downwind from the test site. Indeed, more than 7 weeks after the explosion, unusual detections of noble gases was reported at the radionuclide station in Takasaki, Japan. The radionuclide station is a part of the International Monitoring System, operated to verify the CTBT. This study provides an estimation of the possible source region and the total radioactivity of the release using Atmospheric Transport Modeling.

Keywords: atmospheric transport modeling, CTBTO, nuclear tests, radioactive fission products

Procedia PDF Downloads 422
4069 E-Consumers’ Attribute Non-Attendance Switching Behavior: Effect of Providing Information on Attributes

Authors: Leonard Maaya, Michel Meulders, Martina Vandebroek

Abstract:

Discrete Choice Experiments (DCE) are used to investigate how product attributes affect decision-makers’ choices. In DCEs, choice situations consisting of several alternatives are presented from which choice-makers select the preferred alternative. Standard multinomial logit models based on random utility theory can be used to estimate the utilities for the attributes. The overarching principle in these models is that respondents understand and use all the attributes when making choices. However, studies suggest that respondents sometimes ignore some attributes (commonly referred to as Attribute Non-Attendance/ANA). The choice modeling literature presents ANA as a static process, i.e., respondents’ ANA behavior does not change throughout the experiment. However, respondents may ignore attributes due to changing factors like availability of information on attributes, learning/fatigue in experiments, etc. We develop a dynamic mixture latent Markov model to model changes in ANA when information on attributes is provided. The model is illustrated on e-consumers’ webshop choices. The results indicate that the dynamic ANA model describes the behavioral changes better than modeling the impact of information using changes in parameters. Further, we find that providing information on attributes leads to an increase in the attendance probabilities for the investigated attributes.

Keywords: choice models, discrete choice experiments, dynamic models, e-commerce, statistical modeling

Procedia PDF Downloads 133
4068 Transient Modeling of Velocity Profile and Heat Transfer of Electrohydrodynamically Augmented Micro Heat Pipe

Authors: H. Shokouhmand, M. Tajerian

Abstract:

At this paper velocity profile modeling and heat transfer in the micro heat pipes by using electrohydrodynamic (EHD) field at the transient regime have been studied. In the transient flow, one dimensional and two phase fluid flow and heat transfer for micro heat pipes with square cross section, have been studied. At this model Coulomb and dielectrophoretic forces are considered. Coupled, non-linear equations governed on the model (continuity, momentum, and energy equations) have been solved simultaneously by numerical methods. Transient behavior of affecting parameters e.g. substrate temperature, velocity of coolant liquid, radius of curvature and coolant liquid pressure, has been verified. By obtaining and plotting the mentioned parameters, it has been shown that the EHD field enhances the heat transfer process. So, the time required to reach the steady state regime decreases from 16 seconds to 2.4 seconds after applying EHD field. Another result has been observed implicitly that by increasing the heat input the effect of EHD field became more significant. The numerical results of model predict the experimental results available in the literature successfully, and it has been observed there is a good agreement between them.

Keywords: micro heat pipe, transient modeling, electrohydrodynamics, capillary, meniscus

Procedia PDF Downloads 258
4067 Thermodynamic Modeling of Three Pressure Level Reheat HRSG, Parametric Analysis and Optimization Using PSO

Authors: Mahmoud Nadir, Adel Ghenaiet

Abstract:

The main purpose of this study is the thermodynamic modeling, the parametric analysis, and the optimization of three pressure level reheat HRSG (Heat Recovery Steam Generator) using PSO method (Particle Swarm Optimization). In this paper, a parametric analysis followed by a thermodynamic optimization is presented. The chosen objective function is the specific work of the steam cycle that may be, in the case of combined cycle (CC), a good criterion of thermodynamic performance analysis, contrary to the conventional steam turbines in which the thermal efficiency could be also an important criterion. The technologic constraints such as maximal steam cycle temperature, minimal steam fraction at steam turbine outlet, maximal steam pressure, minimal stack temperature, minimal pinch point, and maximal superheater effectiveness are also considered. The parametric analyses permitted to understand the effect of design parameters and the constraints on steam cycle specific work variation. PSO algorithm was used successfully in HRSG optimization, knowing that the achieved results are in accordance with those of the previous studies in which genetic algorithms were used. Moreover, this method is easy to implement comparing with the other methods.

Keywords: combined cycle, HRSG thermodynamic modeling, optimization, PSO, steam cycle specific work

Procedia PDF Downloads 376
4066 Stochastic Modeling for Parameters of Modified Car-Following Model in Area-Based Traffic Flow

Authors: N. C. Sarkar, A. Bhaskar, Z. Zheng

Abstract:

The driving behavior in area-based (i.e., non-lane based) traffic is induced by the presence of other individuals in the choice space from the driver’s visual perception area. The driving behavior of a subject vehicle is constrained by the potential leaders and leaders are frequently changed over time. This paper is to determine a stochastic model for a parameter of modified intelligent driver model (MIDM) in area-based traffic (as in developing countries). The parametric and non-parametric distributions are presented to fit the parameters of MIDM. The goodness of fit for each parameter is measured in two different ways such as graphically and statistically. The quantile-quantile (Q-Q) plot is used for a graphical representation of a theoretical distribution to model a parameter and the Kolmogorov-Smirnov (K-S) test is used for a statistical measure of fitness for a parameter with a theoretical distribution. The distributions are performed on a set of estimated parameters of MIDM. The parameters are estimated on the real vehicle trajectory data from India. The fitness of each parameter with a stochastic model is well represented. The results support the applicability of the proposed modeling for parameters of MIDM in area-based traffic flow simulation.

Keywords: area-based traffic, car-following model, micro-simulation, stochastic modeling

Procedia PDF Downloads 144
4065 Control-Oriented Enhanced Zero-Dimensional Two-Zone Combustion Modelling of Internal Combustion Engines

Authors: Razieh Arian, Hadi Adibi-Asl

Abstract:

This paper investigates an efficient combustion modeling for cycle simulation of internal combustion engine (ICE) studies. The term “efficient model” means that the models must generate desired simulation results while having fast simulation time. In other words, the efficient model is defined based on the application of the model. The objective of this study is to develop math-based models for control applications or shortly control-oriented models. This study compares different modeling approaches used to model the ICEs such as mean-value models, zero dimensional, quasi-dimensional, and multi-dimensional models for control applications. Mean-value models have been widely used for model-based control applications, but recently by developing advanced simulation tools (e.g. Maple/MapleSim) the higher order models (more complex) could be considered as control-oriented models. This paper presents the enhanced zero-dimensional cycle-by-cycle modeling and simulation of a spark ignition engine with a two-zone combustion model. The simulation results are cross-validated against the simulation results from GT-Power package and show a good agreement in terms of trends and values.

Keywords: Two-zone combustion, control-oriented model, wiebe function, internal combustion engine

Procedia PDF Downloads 334
4064 Predicting the Areal Development of the City of Mashhad with the Automaton Fuzzy Cell Method

Authors: Mehran Dizbadi, Daniyal Safarzadeh, Behrooz Arastoo, Ansgar Brunn

Abstract:

Rapid and uncontrolled expansion of cities has led to unplanned aerial development. In this way, modeling and predicting the urban growth of a city helps decision-makers. In this study, the aspect of sustainable urban development has been studied for the city of Mashhad. In general, the prediction of urban aerial development is one of the most important topics of modern town management. In this research, using the Cellular Automaton (CA) model developed for geo data of Geographic Information Systems (GIS) and presenting a simple and powerful model, a simulation of complex urban processes has been done.

Keywords: urban modeling, sustainable development, fuzzy cellular automaton, geo-information system

Procedia PDF Downloads 123
4063 The Prognostic Prediction Value of Positive Lymph Nodes Numbers for the Hypopharyngeal Squamous Cell Carcinoma

Authors: Wendu Pang, Yaxin Luo, Junhong Li, Yu Zhao, Danni Cheng, Yufang Rao, Minzi Mao, Ke Qiu, Yijun Dong, Fei Chen, Jun Liu, Jian Zou, Haiyang Wang, Wei Xu, Jianjun Ren

Abstract:

We aimed to compare the prognostic prediction value of positive lymph node number (PLNN) to the American Joint Committee on Cancer (AJCC) tumor, lymph node, and metastasis (TNM) staging system for patients with hypopharyngeal squamous cell carcinoma (HPSCC). A total of 826 patients with HPSCC from the Surveillance, Epidemiology, and End Results database (2004–2015) were identified and split into two independent cohorts: training (n=461) and validation (n=365). Univariate and multivariate Cox regression analyses were used to evaluate the prognostic effects of PLNN in patients with HPSCC. We further applied six Cox regression models to compare the survival predictive values of the PLNN and AJCC TNM staging system. PLNN showed a significant association with overall survival (OS) and cancer-specific survival (CSS) (P < 0.001) in both univariate and multivariable analyses, and was divided into three groups (PLNN 0, PLNN 1-5, and PLNN>5). In the training cohort, multivariate analysis revealed that the increased PLNN of HPSCC gave rise to significantly poor OS and CSS after adjusting for age, sex, tumor size, and cancer stage; this trend was also verified by the validation cohort. Additionally, the survival model incorporating a composite of PLNN and TNM classification (C-index, 0.705, 0.734) performed better than the PLNN and AJCC TNM models. PLNN can serve as a powerful survival predictor for patients with HPSCC and is a surrogate supplement for cancer staging systems.

Keywords: hypopharyngeal squamous cell carcinoma, positive lymph nodes number, prognosis, prediction models, survival predictive values

Procedia PDF Downloads 144
4062 Functional Connectivity Signatures of Polygenic Depression Risk in Youth

Authors: Louise Moles, Steve Riley, Sarah D. Lichenstein, Marzieh Babaeianjelodar, Robert Kohler, Annie Cheng, Corey Horien Abigail Greene, Wenjing Luo, Jonathan Ahern, Bohan Xu, Yize Zhao, Chun Chieh Fan, R. Todd Constable, Sarah W. Yip

Abstract:

Background: Risks for depression are myriad and include both genetic and brain-based factors. However, relationships between these systems are poorly understood, limiting understanding of disease etiology, particularly at the developmental level. Methods: We use a data-driven machine learning approach connectome-based predictive modeling (CPM) to identify functional connectivity signatures associated with polygenic risk scores for depression (DEP-PRS) among youth from the Adolescent Brain and Cognitive Development (ABCD) study across diverse brain states, i.e., during resting state, during affective working memory, during response inhibition, during reward processing. Results: Using 10-fold cross-validation with 100 iterations and permutation testing, CPM identified connectivity signatures of DEP-PRS across all examined brain states (rho’s=0.20-0.27, p’s<.001). Across brain states, DEP-PRS was positively predicted by increased connectivity between frontoparietal and salience networks, increased motor-sensory network connectivity, decreased salience to subcortical connectivity, and decreased subcortical to motor-sensory connectivity. Subsampling analyses demonstrated that model accuracies were robust across random subsamples of N’s=1,000, N’s=500, and N’s=250 but became unstable at N’s=100. Conclusions: These data, for the first time, identify neural networks of polygenic depression risk in a large sample of youth before the onset of significant clinical impairment. Identified networks may be considered potential treatment targets or vulnerability markers for depression risk.

Keywords: genetics, functional connectivity, pre-adolescents, depression

Procedia PDF Downloads 52
4061 Modeling and Simulation of Honeycomb Steel Sandwich Panels under Blast Loading

Authors: Sayed M. Soleimani, Nader H. Ghareeb, Nourhan H. Shaker, Muhammad B. Siddiqui

Abstract:

Honeycomb sandwich panels have been widely used as protective structural elements against blast loading. The main advantages of these panels include their light weight due to the presence of voids, as well as their energy absorption capability. Terrorist activities have imposed new challenges to structural engineers to design protective measures for vital structures. Since blast loading is not usually considered in the load combinations during the design process of a structure, researchers around the world have been motivated to study the behavior of potential elements capable of resisting sudden loads imposed by the detonation of explosive materials. One of the best candidates for this objective is the honeycomb sandwich panel. Studying the effects of explosive materials on the panels requires costly and time-consuming experiments. Moreover, these type of experiments need permission from defense organizations which can become a hurdle. As a result, modeling and simulation using an appropriate tool can be considered as a good alternative. In this research work, the finite element package ABAQUS® is used to study the behavior of hexagonal and squared honeycomb steel sandwich panels under the explosive effects of different amounts of trinitrotoluene (TNT). The results of finite element modeling of a specific honeycomb configuration are initially validated by comparing them with the experimental results from literature. Afterwards, several configurations including different geometrical properties of the honeycomb wall are investigated and the results are compared with the original model. Finally, the effectiveness of the core shape and wall thickness are discussed, and conclusions are made.

Keywords: Abaqus, blast loading, finite element modeling, steel honeycomb sandwich panel

Procedia PDF Downloads 347
4060 A Tutorial on Model Predictive Control for Spacecraft Maneuvering Problem with Theory, Experimentation and Applications

Authors: O. B. Iskender, K. V. Ling, V. Dubanchet, L. Simonini

Abstract:

This paper discusses the recent advances and future prospects of spacecraft position and attitude control using Model Predictive Control (MPC). First, the challenges of the space missions are summarized, in particular, taking into account the errors, uncertainties, and constraints imposed by the mission, spacecraft and, onboard processing capabilities. The summary of space mission errors and uncertainties provided in categories; initial condition errors, unmodeled disturbances, sensor, and actuator errors. These previous constraints are classified into two categories: physical and geometric constraints. Last, real-time implementation capability is discussed regarding the required computation time and the impact of sensor and actuator errors based on the Hardware-In-The-Loop (HIL) experiments. The rationales behind the scenarios’ are also presented in the scope of space applications as formation flying, attitude control, rendezvous and docking, rover steering, and precision landing. The objectives of these missions are explained, and the generic constrained MPC problem formulations are summarized. Three key design elements used in MPC design: the prediction model, the constraints formulation and the objective cost function are discussed. The prediction models can be linear time invariant or time varying depending on the geometry of the orbit, whether it is circular or elliptic. The constraints can be given as linear inequalities for input or output constraints, which can be written in the same form. Moreover, the recent convexification techniques for the non-convex geometrical constraints (i.e., plume impingement, Field-of-View (FOV)) are presented in detail. Next, different objectives are provided in a mathematical framework and explained accordingly. Thirdly, because MPC implementation relies on finding in real-time the solution to constrained optimization problems, computational aspects are also examined. In particular, high-speed implementation capabilities and HIL challenges are presented towards representative space avionics. This covers an analysis of future space processors as well as the requirements of sensors and actuators on the HIL experiments outputs. The HIL tests are investigated for kinematic and dynamic tests where robotic arms and floating robots are used respectively. Eventually, the proposed algorithms and experimental setups are introduced and compared with the authors' previous work and future plans. The paper concludes with a conjecture that MPC paradigm is a promising framework at the crossroads of space applications while could be further advanced based on the challenges mentioned throughout the paper and the unaddressed gap.

Keywords: convex optimization, model predictive control, rendezvous and docking, spacecraft autonomy

Procedia PDF Downloads 106
4059 Forecasting of COVID-19 Cases, Hospitalization Admissions, and Death Cases Based on Wastewater Sars-COV-2 Surveillance Using Copula Time Series Model

Authors: Hueiwang Anna Jeng, Norou Diawara, Nancy Welch, Cynthia Jackson, Rekha Singh, Kyle Curtis, Raul Gonzalez, David Jurgens, Sasanka Adikari

Abstract:

Modeling effort is needed to predict the COVID-19 trends for developing management strategies and adaptation measures. The objective of this study was to assess whether SARS-CoV-2 viral load in wastewater could serve as a predictor for forecasting COVID-19 cases, hospitalization cases, and death cases using copula-based time series modeling. SARS-CoV-2 RNA load in raw wastewater in Chesapeake VA was measured using the RT-qPCR method. Gaussian copula time series marginal regression model, incorporating an autoregressive moving average model and the copula function, served as a forecasting model. COVID-19 cases were correlated with wastewater viral load, hospitalization cases, and death cases. The forecasted trend of COVID-19 cases closely paralleled one of the reported cases, with over 90% of the forecasted COVID-19 cases falling within the 99% confidence interval of the reported cases. Wastewater SARS-CoV-2 viral load could serve as a predictor for COVID-19 cases and hospitalization cases.

Keywords: COVID-19, modeling, time series, copula function

Procedia PDF Downloads 62
4058 On the Homology Modeling, Structural Function Relationship and Binding Site Prediction of Human Alsin Protein

Authors: Y. Ruchi, A. Prerna, S. Deepshikha

Abstract:

Amyotrophic lateral sclerosis (ALS), also known as “Lou Gehrig’s disease”. It is a neurodegenerative disease associated with degeneration of motor neurons in the cerebral cortex, brain stem, and spinal cord characterized by distal muscle weakness, atrophy, normal sensation, pyramidal signs and progressive muscular paralysis reflecting. ALS2 is a juvenile autosomal recessive disorder, slowly progressive, that maps to chromosome 2q33 and is associated with mutations in the alsin gene, a putative GTPase regulator. In this paper we have done homology modeling of alsin2 protein using multiple templates (3KCI_A, 4LIM_A, 402W_A, 4D9S_A, and 4DNV_A) designed using the Prime program in Schrödinger software. Further modeled structure is used to identify effective binding sites on the basis of structural and physical properties using sitemap program in Schrödinger software, structural and function analysis is done by using Prosite and ExPASy server that gives insight into conserved domains and motifs that can be used for protein classification. This paper summarizes the structural, functional and binding site property of alsin2 protein. These binding sites can be potential drug target sites and can be used for docking studies.

Keywords: ALS, binding site, homology modeling, neuronal degeneration

Procedia PDF Downloads 381
4057 Gearbox Defect Detection in the Semi Autogenous Mills Using the Vibration Analysis Technique

Authors: Mostafa Firoozabadi, Alireza Foroughi Nematollahi

Abstract:

Semi autogenous mills are designed for grinding or primary crushed ore, and are the most widely used in concentrators globally. Any defect occurrence in semi autogenous mills can stop the production line. A Gearbox is a significant part of a rotating machine or a mill, so, the gearbox monitoring is a necessary process to prevent the unwanted defects. When a defect happens in a gearbox bearing, this defect can be transferred to the other parts of the equipment like inner ring, outer ring, balls, and the bearing cage. Vibration analysis is one of the most effective and common ways to detect the bearing defects in the mills. Vibration signal in a mill can be made by different parts of the mill including electromotor, pinion girth gear, different rolling bearings, and tire. When a vibration signal, made by the aforementioned parts, is added to the gearbox vibration spectrum, an accurate and on time defect detection in the gearbox will be difficult. In this paper, a new method is proposed to detect the gearbox bearing defects in the semi autogenous mill on time and accurately, using the vibration signal analysis method. In this method, if the vibration values are increased in the vibration curve, the probability of defect occurrence is investigated by comparing the equipment vibration values and the standard ones. Then, all vibration frequencies are extracted from the vibration signal and the equipment defect is detected using the vibration spectrum curve. This method is implemented on the semi autogenous mills in the Golgohar mining and industrial company in Iran. The results show that the proposed method can detect the bearing looseness on time and accurately. After defect detection, the bearing is opened before the equipment failure and the predictive maintenance actions are implemented on it.

Keywords: condition monitoring, gearbox defects, predictive maintenance, vibration analysis

Procedia PDF Downloads 457
4056 Disrupted or Discounted Cash Flow: Impact of Digitisation on Business Valuation

Authors: Matthias Haerri, Tobias Huettche, Clemens Kustner

Abstract:

This article discusses the impact of digitization on business valuation. In order to become and remain ‘digital’, investments are necessary whose return on investment (ROI) often remains vague. This uncertainty is contradictory for a valuation, that rely on predictable cash flows, fixed capital structures and the steady state. However digitisation does not make a company valuation impossible, but traditional approaches must be reconsidered. The authors identify four areas that are to be changing: (1) Tools instead of intuition - In the future, company valuation will neither be art nor science, but craft. This does not require intuition, but experience and good tools. Digital evaluation tools beyond Excel will therefore gain in importance. (2) Real-time instead of deadline - At present, company valuations are always carried out on a case-by-case basis and on a specific key date. This will change with the digitalization and the introduction of web-based valuation tools. Company valuations can thus not only be carried out faster and more efficiently, but can also be offered more frequently. Instead of calculating the value for a previous key date, current and real-time valuations can be carried out. (3) Predictive planning instead of analysis of the past - Past data will also be needed in the future, but its use will not be limited to monovalent time series or key figure analyses. With pictures of ‘black swans’ and the ‘turkey illusion’ it was made clear to us that we build forecasts on too few data points of the past and underestimate the power of chance. Predictive planning can help here. (4) Convergence instead of residual value - Digital transformation shortens the lifespan of viable business models. If companies want to live forever, they have to change forever. For the company valuation, this means that the business model valid on the valuation date only has a limited service life.

Keywords: business valuation, corporate finance, digitisation, disruption

Procedia PDF Downloads 128
4055 Modeling Methodologies for Optimization and Decision Support on Coastal Transport Information System (Co.Tr.I.S.)

Authors: Vassilios Moussas, Dimos N. Pantazis, Panagioths Stratakis

Abstract:

The aim of this paper is to present the optimization methodology developed in the frame of a Coastal Transport Information System. The system will be used for the effective design of coastal transportation lines and incorporates subsystems that implement models, tools and techniques that may support the design of improved networks. The role of the optimization and decision subsystem is to provide the user with better and optimal scenarios that will best fulfill any constrains, goals or requirements posed. The complexity of the problem and the large number of parameters and objectives involved led to the adoption of an evolutionary method (Genetic Algorithms). The problem model and the subsystem structure are presented in detail, and, its support for simulation is also discussed.

Keywords: coastal transport, modeling, optimization

Procedia PDF Downloads 494
4054 Impact of Transitioning to Renewable Energy Sources on Key Performance Indicators and Artificial Intelligence Modules of Data Center

Authors: Ahmed Hossam ElMolla, Mohamed Hatem Saleh, Hamza Mostafa, Lara Mamdouh, Yassin Wael

Abstract:

Artificial intelligence (AI) is reshaping industries, and its potential to revolutionize renewable energy and data center operations is immense. By harnessing AI's capabilities, we can optimize energy consumption, predict fluctuations in renewable energy generation, and improve the efficiency of data center infrastructure. This convergence of technologies promises a future where energy is managed more intelligently, sustainably, and cost-effectively. The integration of AI into renewable energy systems unlocks a wealth of opportunities. Machine learning algorithms can analyze vast amounts of data to forecast weather patterns, solar irradiance, and wind speeds, enabling more accurate energy production planning. AI-powered systems can optimize energy storage and grid management, ensuring a stable power supply even during intermittent renewable generation. Moreover, AI can identify maintenance needs for renewable energy infrastructure, preventing costly breakdowns and maximizing system lifespan. Data centers, which consume substantial amounts of energy, are prime candidates for AI-driven optimization. AI can analyze energy consumption patterns, identify inefficiencies, and recommend adjustments to cooling systems, server utilization, and power distribution. Predictive maintenance using AI can prevent equipment failures, reducing energy waste and downtime. Additionally, AI can optimize data placement and retrieval, minimizing energy consumption associated with data transfer. As AI transforms renewable energy and data center operations, modified Key Performance Indicators (KPIs) will emerge. Traditional metrics like energy efficiency and cost-per-megawatt-hour will continue to be relevant, but additional KPIs focused on AI's impact will be essential. These might include AI-driven cost savings, predictive accuracy of energy generation and consumption, and the reduction of carbon emissions attributed to AI-optimized operations. By tracking these KPIs, organizations can measure the success of their AI initiatives and identify areas for improvement. Ultimately, the synergy between AI, renewable energy, and data centers holds the potential to create a more sustainable and resilient future. By embracing these technologies, we can build smarter, greener, and more efficient systems that benefit both the environment and the economy.

Keywords: data center, artificial intelligence, renewable energy, energy efficiency, sustainability, optimization, predictive analytics, energy consumption, energy storage, grid management, data center optimization, key performance indicators, carbon emissions, resiliency

Procedia PDF Downloads 15
4053 Evaluation of Hydrogen Particle Volume on Surfaces of Selected Nanocarbons

Authors: M. Ziółkowska, J. T. Duda, J. Milewska-Duda

Abstract:

This paper describes an approach to the adsorption phenomena modeling aimed at specifying the adsorption mechanisms on localized or nonlocalized adsorbent sites, when applied to the nanocarbons. The concept comes from the fundamental thermodynamic description of adsorption equilibrium and is based on numerical calculations of the hydrogen adsorbed particles volume on the surface of selected nanocarbons: single-walled nanotube and nanocone. This approach enables to obtain information on adsorption mechanism and then as a consequence to take appropriate mathematical adsorption model, thus allowing for a more reliable identification of the material porous structure. Theoretical basis of the approach is discussed and newly derived results of the numerical calculations are presented for the selected nanocarbons.

Keywords: adsorption, mathematical modeling, nanocarbons, numerical analysis

Procedia PDF Downloads 261
4052 Application of Directed Acyclic Graphs for Threat Identification Based on Ontologies

Authors: Arun Prabhakar

Abstract:

Threat modeling is an important activity carried out in the initial stages of the development lifecycle that helps in building proactive security measures in the product. Though there are many techniques and tools available today, one of the common challenges with the traditional methods is the lack of a systematic approach in identifying security threats. The proposed solution describes an organized model by defining ontologies that help in building patterns to enumerate threats. The concepts of graph theory are applied to build the pattern for discovering threats for any given scenario. This graph-based solution also brings in other benefits, making it a customizable and scalable model.

Keywords: directed acyclic graph, ontology, patterns, threat identification, threat modeling

Procedia PDF Downloads 130
4051 Impacts of Social Support on Perceived Level of Stress and Self-Esteem among Students of Private Universities of Karachi-Pakistan

Authors: Sheeba Farhan

Abstract:

This study is conducted to explore the predictive relationship of perceived stress and self-esteem with social support of students and to explore the factors, which contribute to develop or enhance the level of stress in students of private universities in Karachi-Pakistan. After literature review following hypotheses were formulated; 1)social support would predict perceived stress of students of business administration of private organizations of Higher education, 2) social support would predict the self-esteem of students of private organizations of Higher education, 3) there will be a relationship of perceived stress and self-esteem of students of private organizations of Higher education, 4) there will be a relationship of self esteem and social support of students of private organizations of Higher education. Sample of the study is comprise of 100 students of private organizations of Higher education in Karachi- Pakistan (i.e. males= 50 & females= 50). The age range of participants is 18-26 years. The measures, used in the study are: Demographic information form, a semi structured interview form, Rosenberg self esteem scale (Rosenberg, 1965) and perceived stress scale (Cohen, Kamarck, and Mermelstein, 1983) and multidimensional scale of perceived social support (Zimet, 1988) Descriptive statistics is used for getting a better statistical view of characteristics of sample. Regression analysis is used to explore the predictive relationship of study related stress and self esteem with academic achievement of students of private organizations of Higher education. Percentages and ratios were calculated to explore the level of perceived stress with respect to Socio-demographic characteristics in students of private organizations of Higher education. Finding shows that social support is significantly associated with the higher level of self-esteem among students of graduation but insignificantly associated with stress that has been experienced by them. These results are correlated with a wide variety of studies in which social support has proposed to be a predictor of well being for the students.

Keywords: private universities of Karachi-Pakistan, Self-esteem, social support, stress

Procedia PDF Downloads 285
4050 VISSIM Modeling of Driver Behavior at Connecticut Roundabouts

Authors: F. Clara Fang, Hernan Castaneda

Abstract:

The Connecticut Department of Transportation (ConnDOT) has constructed four roundabouts in the State of Connecticut within the past ten years. VISSIM traffic simulation software was utilized to analyze these roundabouts during their design phase. The queue length and level of service observed in the field appear to be better than predicted by the VISSIM model. The objectives of this project are to: identify VISSIM input variables most critical to accurate modeling; recommend VISSIM calibration factors; and, provide other recommendations for roundabout traffic operations modeling. Traffic data were collected at these roundabouts using Miovision Technologies. Cameras were set up to capture vehicle circulating activity and entry behavior for two weekdays. A large sample size of filed data was analyzed to achieve accurate and statistically significant results. The data extracted from the videos include: vehicle circulating speed; critical gap estimated by Maximum Likelihood Method; peak hour volume; follow-up headway; travel time; and, vehicle queue length. A VISSIM simulation of existing roundabouts was built to compare both queue length and travel time predicted from simulation with measured in the field. The research investigated a variety of simulation parameters as calibration factors for describing driver behaviors at roundabouts. Among them, critical gap is the most effective calibration variable in roundabout simulation. It has a significant impact to queue length, particularly when the volume is higher. The results will improve the design of future roundabouts in Connecticut and provide decision makers with insights on the relationship between various choices and future performance.

Keywords: driver critical gap, roundabout analysis, simulation, VISSIM modeling

Procedia PDF Downloads 284
4049 Evaluation of a Data Fusion Algorithm for Detecting and Locating a Radioactive Source through Monte Carlo N-Particle Code Simulation and Experimental Measurement

Authors: Hadi Ardiny, Amir Mohammad Beigzadeh

Abstract:

Through the utilization of a combination of various sensors and data fusion methods, the detection of potential nuclear threats can be significantly enhanced by extracting more information from different data. In this research, an experimental and modeling approach was employed to track a radioactive source by combining a surveillance camera and a radiation detector (NaI). To run this experiment, three mobile robots were utilized, with one of them equipped with a radioactive source. An algorithm was developed in identifying the contaminated robot through correlation between camera images and camera data. The computer vision method extracts the movements of all robots in the XY plane coordinate system, and the detector system records the gamma-ray count. The position of the robots and the corresponding count of the moving source were modeled using the MCNPX simulation code while considering the experimental geometry. The results demonstrated a high level of accuracy in finding and locating the target in both the simulation model and experimental measurement. The modeling techniques prove to be valuable in designing different scenarios and intelligent systems before initiating any experiments.

Keywords: nuclear threats, radiation detector, MCNPX simulation, modeling techniques, intelligent systems

Procedia PDF Downloads 107
4048 Groundwater Level Modelling by ARMA and PARMA Models (Case Study: Qorveh Aquifer)

Authors: Motalleb Byzedi, Seyedeh Chaman Naderi Korvandan

Abstract:

Regarding annual statistics of groundwater level resources about current piezometers at Qorveh plains, both ARMA & PARMA modeling methods were applied in this study by the using of SAMS software. Upon performing required tests, a model was used with minimum amount of Akaike information criteria and suitable model was selected for piezometers. Then it was possible to make necessary estimations by using these models for future fluctuations in each piezometer. According to the results, ARMA model had more facilities for modeling of aquifer. Also it was cleared that eastern parts of aquifer had more failures than other parts. Therefore it is necessary to prohibit critical parts along with more supervision on taking rates of wells.

Keywords: qorveh plain, groundwater level, ARMA, PARMA

Procedia PDF Downloads 277
4047 Metabolic Pathway Analysis of Microbes using the Artificial Bee Colony Algorithm

Authors: Serena Gomez, Raeesa Tanseen, Netra Shaligram, Nithin Francis, Sandesh B. J.

Abstract:

The human gut consists of a community of microbes which has a lot of effects on human health disease. Metabolic modeling can help to predict relative populations of stable microbes and their effect on health disease. In order to study and visualize microbes in the human gut, we developed a tool that offers the following modules: Build a tool that can be used to perform Flux Balance Analysis for microbes in the human gut using the Artificial Bee Colony optimization algorithm. Run simulations for an individual microbe in different conditions, such as aerobic and anaerobic and visualize the results of these simulations.

Keywords: microbes, metabolic modeling, flux balance analysis, artificial bee colony

Procedia PDF Downloads 93
4046 Tc-99m MIBI Scintigraphy to Differentiate Malignant from Benign Lesions, Detected on Planar Bone Scan

Authors: Aniqa Jabeen

Abstract:

The aim of this study was to evaluate the effectiveness of Tc-99m MIBI (Technetium 99-methoxy-iso-butyl-isonitrile) scintigraphy to differentiate malignancies from benign lesions, which were detected on planar bone scans. Materials and Methods: 59 patients with bone lesions were enrolled in the study. The scintigraphic findings were compared with the clinical, radiological and the histological findings. Each patient initially underwent a three-phase bone scan with Tc-99m MDP (Methylene Diphosphonate) and if evidence of lesion found, the patient then underwent a dynamic and static MIBI scintigraphy after three to four days. The MDP and MIBI scans were evaluated visually and quantitatively. For quantitative analysis count ratios of lesions and contralateral normal side (L/C) were taken by region of interests drawn on scans. The Student T test was applied to assess the significant difference between benign and malignant lesions p-value < 0.05 was considered significant. Result: The MDP scans showed the increase tracer uptake, but there was no significant difference between benign and malignant uptake of the radiotracer. However significant difference (p-value 0.015), in uptake was seen in malignant (L/C = 3.51 ± 1.02) and benign lesion (L/C = 2.50±0.42) on MIBI scan. Three of thirty benign lesions did not show significant MIBI uptake. Seven malignant appeared as false negatives. Specificity of the scan was 86.66%, and its Negative Predictive Value (NPV) was 81.25% whereas the sensitivity of scan was 79.31%. In excluding the axial metastasis from the lesions, the sensitivity of MIBI scan increased to 91.66% and the NPV also increased to 92.85%. Conclusion: MIBI scintigraphy provides its usefulness by distinguishing malignant from benign lesions. MIBI also correctly identifies metastatic lesions. The negative predictive value of the scan points towards its ability to accurately diagnose the normal (benign) cases. However, biopsy remains the gold standard and a definitive diagnostic modality in musculoskeletal tumors. MIBI scan provides useful information in preoperative assessment and in distinguishing between malignant and benign lesions.

Keywords: benign, malignancies, MDP bone scan, MIBI scintigraphy

Procedia PDF Downloads 397
4045 An Optimal Approach for Full-Detailed Friction Model Identification of Reaction Wheel

Authors: Ghasem Sharifi, Hamed Shahmohamadi Ousaloo, Milad Azimi, Mehran Mirshams

Abstract:

The ever-increasing use of satellites demands a search for increasingly accurate and reliable pointing systems. Reaction wheels are rotating devices used commonly for the attitude control of the spacecraft since provide a wide range of torque magnitude and high reliability. The numerical modeling of this device can significantly enhance the accuracy of the satellite control in space. Modeling the wheel rotation in the presence of the various frictions is one of the critical parts of this approach. This paper presents a Dynamic Model Control of a Reaction Wheel (DMCR) in the current control mode. In current-mode, the required current is delivered to the coils in order to achieve the desired torque. During this research, all the friction parameters as viscous and coulomb, motor coefficient, resistance and voltage constant are identified. In order to model identification of a reaction wheel, numerous varying current commands apply on the particular wheel to verify the estimated model. All the parameters of DMCR are identified by classical Levenberg-Marquardt (CLM) optimization method. The experimental results demonstrate that the developed model has an appropriate precise and can be used in the satellite control simulation.

Keywords: experimental modeling, friction parameters, model identification, reaction wheel

Procedia PDF Downloads 228
4044 Quantifying Wave Attenuation over an Eroding Marsh through Numerical Modeling

Authors: Donald G. Danmeier, Gian Marco Pizzo, Matthew Brennan

Abstract:

Although wetlands have been proposed as a green alternative to manage coastal flood hazards because of their capacity to adapt to sea level rise and provision of multiple ecological and social co-benefits, they are often overlooked due to challenges in quantifying the uncertainty and naturally, variability of these systems. This objective of this study was to quantify wave attenuation provided by a natural marsh surrounding a large oil refinery along the US Gulf Coast that has experienced steady erosion along the shoreward edge. The vegetation module of the SWAN was activated and coupled with a hydrodynamic model (DELFT3D) to capture two-way interactions between the changing water level and wavefield over the course of a storm event. Since the marsh response to relative sea level rise is difficult to predict, a range of future marsh morphologies is explored. Numerical results were examined to determine the amount of wave attenuation as a function of marsh extent and the relative contributions from white-capping, depth-limited wave breaking, bottom friction, and flexing of vegetation. In addition to the coupled DELFT3D-SWAN modeling of a storm event, an uncoupled SWAN-VEG model was applied to a simplified bathymetry to explore a larger experimental design space. The wave modeling revealed that the rate of wave attenuation reduces for higher surge but was still significant over a wide range of water levels and outboard wave heights. The results also provide insights to the minimum marsh extent required to fully realize the potential wave attenuation so the changing coastal hazards can be managed.

Keywords: green infrastructure, wave attenuation, wave modeling, wetland

Procedia PDF Downloads 128
4043 Modeling of Thermally Induced Acoustic Emission Memory Effects in Heterogeneous Rocks with Consideration for Fracture Develo

Authors: Vladimir A. Vinnikov

Abstract:

The paper proposes a model of an inhomogeneous rock mass with initially random distribution of microcracks on mineral grain boundaries. It describes the behavior of cracks in a medium under the effect of thermal field, the medium heated instantaneously to a predetermined temperature. Crack growth occurs according to the concept of fracture mechanics provided that the stress intensity factor K exceeds the critical value of Kc. The modeling of thermally induced acoustic emission memory effects is based on the assumption that every event of crack nucleation or crack growth caused by heating is accompanied by a single acoustic emission event. Parameters of the thermally induced acoustic emission memory effect produced by cyclic heating and cooling (with the temperature amplitude increasing from cycle to cycle) were calculated for several rock texture types (massive, banded, and disseminated). The study substantiates the adaptation of the proposed model to humidity interference with the thermally induced acoustic emission memory effect. The influence of humidity on the thermally induced acoustic emission memory effect in quasi-homogeneous and banded rocks is estimated. It is shown that such modeling allows the structure and texture of rocks to be taken into account and the influence of interference factors on the distinctness of the thermally induced acoustic emission memory effect to be estimated. The numerical modeling can be used to obtain information about the thermal impacts on rocks in the past and determine the degree of rock disturbance by means of non-destructive testing.

Keywords: degree of rock disturbance, non-destructive testing, thermally induced acoustic emission memory effects, structure and texture of rocks

Procedia PDF Downloads 259
4042 The Convection Heater Numerical Simulation

Authors: Cristian Patrascioiu, Loredana Negoita

Abstract:

This paper is focused on modeling and simulation of the tubular heaters. The paper is structured in four parts: the structure of the tubular convection section, the heat transfer model, the adaptation of the mathematical model and the solving model. The main hypothesis of the heat transfer modeling is that the heat exchanger of the convective tubular heater is a lumped system. In the same time, the model uses the heat balance relations, Newton’s law and criteria relations. The numerical program achieved allows for the estimation of the burn gases outlet temperature and the heated flow outlet temperature.

Keywords: heat exchanger, mathematical modelling, nonlinear equation system, Newton-Raphson algorithm

Procedia PDF Downloads 287
4041 Modeling of Bed Level Changes in Larak Island

Authors: Saeed Zeinali, Nasser Talebbeydokhti, Mehdi Saeidian, Shahrad Vosough

Abstract:

In this article, bathymetry changes have been studied as a case study for Larak Island, located in The South of Iran. The advanced 2D model of Mike21 has been used for this purpose. A simple procedure has been utilized in this model. First, the hydrodynamic (HD) module of Mike21 has been used to obtain the required output for sediment transport model (ST module). The ST module modeled the area for tidal currents only. Bed level changes are resulted by series of modeling for both HD and ST module in 3 months time step. The final bathymetry in each time step is used as the primary bathymetry for next time step. This consecutive procedure been continued until bathymetry for the year 2020 is obtained.

Keywords: bed level changes, Larak Island, hydrodynamic, sediment transport

Procedia PDF Downloads 256