Search results for: equivalent linear model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19028

Search results for: equivalent linear model

16568 Integration GIS–SCADA Power Systems to Enclosure Air Dispersion Model

Authors: Ibrahim Shaker, Amr El Hossany, Moustafa Osman, Mohamed El Raey

Abstract:

This paper will explore integration model between GIS–SCADA system and enclosure quantification model to approach the impact of failure-safe event. There are real demands to identify spatial objects and improve control system performance. Nevertheless, the employed methodology is predicting electro-mechanic operations and corresponding time to environmental incident variations. Open processing, as object systems technology, is presented for integration enclosure database with minimal memory size and computation time via connectivity drivers such as ODBC:JDBC during main stages of GIS–SCADA connection. The function of Geographic Information System is manipulating power distribution in contrast to developing issues. In other ward, GIS-SCADA systems integration will require numerical objects of process to enable system model calibration and estimation demands, determine of past events for analysis and prediction of emergency situations for response training.

Keywords: air dispersion model, environmental management, SCADA systems, GIS system, integration power system

Procedia PDF Downloads 350
16567 Human Action Recognition Using Variational Bayesian HMM with Dirichlet Process Mixture of Gaussian Wishart Emission Model

Authors: Wanhyun Cho, Soonja Kang, Sangkyoon Kim, Soonyoung Park

Abstract:

In this paper, we present the human action recognition method using the variational Bayesian HMM with the Dirichlet process mixture (DPM) of the Gaussian-Wishart emission model (GWEM). First, we define the Bayesian HMM based on the Dirichlet process, which allows an infinite number of Gaussian-Wishart components to support continuous emission observations. Second, we have considered an efficient variational Bayesian inference method that can be applied to drive the posterior distribution of hidden variables and model parameters for the proposed model based on training data. And then we have derived the predictive distribution that may be used to classify new action. Third, the paper proposes a process of extracting appropriate spatial-temporal feature vectors that can be used to recognize a wide range of human behaviors from input video image. Finally, we have conducted experiments that can evaluate the performance of the proposed method. The experimental results show that the method presented is more efficient with human action recognition than existing methods.

Keywords: human action recognition, Bayesian HMM, Dirichlet process mixture model, Gaussian-Wishart emission model, Variational Bayesian inference, prior distribution and approximate posterior distribution, KTH dataset

Procedia PDF Downloads 339
16566 Understanding Hydrodynamic in Lake Victoria Basin in a Catchment Scale: A Literature Review

Authors: Seema Paul, John Mango Magero, Prosun Bhattacharya, Zahra Kalantari, Steve W. Lyon

Abstract:

The purpose of this review paper is to develop an understanding of lake hydrodynamics and the potential climate impact on the Lake Victoria (LV) catchment scale. This paper briefly discusses the main problems of lake hydrodynamics and its’ solutions that are related to quality assessment and climate effect. An empirical methodology in modeling and mapping have considered for understanding lake hydrodynamic and visualizing the long-term observational daily, monthly, and yearly mean dataset results by using geographical information system (GIS) and Comsol techniques. Data were obtained for the whole lake and five different meteorological stations, and several geoprocessing tools with spatial analysis are considered to produce results. The linear regression analyses were developed to build climate scenarios and a linear trend on lake rainfall data for a long period. A potential evapotranspiration rate has been described by the MODIS and the Thornthwaite method. The rainfall effect on lake water level observed by Partial Differential Equations (PDE), and water quality has manifested by a few nutrients parameters. The study revealed monthly and yearly rainfall varies with monthly and yearly maximum and minimum temperatures, and the rainfall is high during cool years and the temperature is high associated with below and average rainfall patterns. Rising temperatures are likely to accelerate evapotranspiration rates and more evapotranspiration is likely to lead to more rainfall, drought is more correlated with temperature and cloud is more correlated with rainfall. There is a trend in lake rainfall and long-time rainfall on the lake water surface has affected the lake level. The onshore and offshore have been concentrated by initial literature nutrients data. The study recommended that further studies should consider fully lake bathymetry development with flow analysis and its’ water balance, hydro-meteorological processes, solute transport, wind hydrodynamics, pollution and eutrophication these are crucial for lake water quality, climate impact assessment, and water sustainability.

Keywords: climograph, climate scenarios, evapotranspiration, linear trend flow, rainfall event on LV, concentration

Procedia PDF Downloads 77
16565 Simulation Research of the Aerodynamic Drag of 3D Structures for Individual Transport Vehicle

Authors: Pawel Magryta, Mateusz Paszko

Abstract:

In today's world, a big problem of individual mobility, especially in large urban areas, occurs. Commonly used grand way of transport such as buses, trains or cars do not fulfill their tasks, i.e. they are not able to meet the increasing mobility needs of the growing urban population. Additional to that, the limitations of civil infrastructure construction in the cities exist. Nowadays the most common idea is to transfer the part of urban transport on the level of air transport. However to do this, there is a need to develop an individual flying transport vehicle. The biggest problem occurring in this concept is the type of the propulsion system from which the vehicle will obtain a lifting force. Standard propeller drives appear to be too noisy. One of the ideas is to provide the required take-off and flight power by the machine using the innovative ejector system. This kind of the system will be designed through a suitable choice of the three-dimensional geometric structure with special shape of nozzle in order to generate overpressure. The authors idea is to make a device that would allow to cumulate the overpressure using the a five-sided geometrical structure that will be limited on the one side by the blowing flow of air jet. In order to test this hypothesis a computer simulation study of aerodynamic drag of such 3D structures have been made. Based on the results of these studies, the tests on real model were also performed. The final stage of work was a comparative analysis of the results of simulation and real tests. The CFD simulation studies of air flow was conducted using the Star CD - Star Pro 3.2 software. The design of virtual model was made using the Catia v5 software. Apart from the objective to obtain advanced aviation propulsion system, all of the tests and modifications of 3D structures were also aimed at achieving high efficiency of this device while maintaining the ability to generate high value of overpressures. This was possible only in case of a large mass flow rate of air. All these aspects have been possible to verify using CFD methods for observing the flow of the working medium in the tested model. During the simulation tests, the distribution and size of pressure and velocity vectors were analyzed. Simulations were made with different boundary conditions (supply air pressure), but with a fixed external conditions (ambient temp., ambient pressure, etc.). The maximum value of obtained overpressure is 2 kPa. This value is too low to exploit the power of this device for the individual transport vehicle. Both the simulation model and real object shows a linear dependence of the overpressure values obtained from the different geometrical parameters of three-dimensional structures. Application of computational software greatly simplifies and streamlines the design and simulation capabilities. This work has been financed by the Polish Ministry of Science and Higher Education.

Keywords: aviation propulsion, CFD, 3d structure, aerodynamic drag

Procedia PDF Downloads 292
16564 Reliability and Probability Weighted Moment Estimation for Three Parameter Mukherjee-Islam Failure Model

Authors: Ariful Islam, Showkat Ahmad Lone

Abstract:

The Mukherjee-Islam Model is commonly used as a simple life time distribution to assess system reliability. The model exhibits a better fit for failure information and provides more appropriate information about hazard rate and other reliability measures as shown by various authors. It is possible to introduce a location parameter at a time (i.e., a time before which failure cannot occur) which makes it a more useful failure distribution than the existing ones. Even after shifting the location of the distribution, it represents a decreasing, constant and increasing failure rate. It has been shown to represent the appropriate lower tail of the distribution of random variables having fixed lower bound. This study presents the reliability computations and probability weighted moment estimation of three parameter model. A comparative analysis is carried out between three parameters finite range model and some existing bathtub shaped curve fitting models. Since probability weighted moment method is used, the results obtained can also be applied on small sample cases. Maximum likelihood estimation method is also applied in this study.

Keywords: comparative analysis, maximum likelihood estimation, Mukherjee-Islam failure model, probability weighted moment estimation, reliability

Procedia PDF Downloads 262
16563 A Resource Optimization Strategy for CPU (Central Processing Unit) Intensive Applications

Authors: Junjie Peng, Jinbao Chen, Shuai Kong, Danxu Liu

Abstract:

On the basis of traditional resource allocation strategies, the usage of resources on physical servers in cloud data center is great uncertain. It will cause waste of resources if the assignment of tasks is not enough. On the contrary, it will cause overload if the assignment of tasks is too much. This is especially obvious when the applications are the same type because of its resource preferences. Considering CPU intensive application is one of the most common types of application in the cloud, we studied the optimization strategy for CPU intensive applications on the same server. We used resource preferences to analyze the case that multiple CPU intensive applications run simultaneously, and put forward a model which can predict the execution time for CPU intensive applications which run simultaneously. Based on the prediction model, we proposed the method to select the appropriate number of applications for a machine. Experiments show that the model can predict the execution time accurately for CPU intensive applications. To improve the execution efficiency of applications, we propose a scheduling model based on priority for CPU intensive applications. Extensive experiments verify the validity of the scheduling model.

Keywords: cloud computing, CPU intensive applications, resource optimization, strategy

Procedia PDF Downloads 265
16562 The Mediation Role of Loneliness in the Relationship between Interpersonal Trust and Empathy

Authors: Ghazal Doostmohammadi, Susan Rahimzadeh

Abstract:

Aim: This research aimed to investigate the relationship between empathy and interpersonal trust and recognize the mediating role of loneliness between them in both genders. Methods: With a correlational descriptive design, 192 university students (130 female and 62 male) responded to the questionnaires on “empathy quotient,” “loneliness,” and “interpersonal trust” tests. These tests were designed and validated by experts in the field. Data were analysed using Pearson correlation and path analysis, which is a statistical technique that uses standard linear regression equations to determine the degree of conformity of a theoretical causal model with reality. Results: The data analysis showed that there was no significant correlation between interpersonal trust, both with loneliness (t=0.169) and empathy (t=0.186), while there was a significant negative correlation (t=0.359) between empathy and loneliness. This means that there is an inverse correlation between empathy and loneliness. The path analysis confirmed the hypothesis of the research about the mediating role of loneliness between empathy and interpersonal trust. But gender did not play a role in this relationship. Conclusion: As an outcome, clinical professionals and education trainers should pay more attention to interpersonal trust as a basic need and try to recreate and shape it to prevent people's social breakdown, and on the other hand, self-disclosure training (especially in Men), expression of feelings and courage should be given double importance to prevent the consequences of loneliness.

Keywords: empathy, loneliness, interpersonal trust, gender

Procedia PDF Downloads 62
16561 Simulation Model of Biosensor Based on Gold Nanoparticles

Authors: Kholod Hajo

Abstract:

In this study COMSOL Multiphysics was used to design lateral flow biosensors (LFBs) which provide advantages in low cost, simplicity, rapidity, stability and portability thus making LFBs popular in biomedical, agriculture, food and environmental sciences. This study was focused on simulation model of biosensor based on gold nanoparticles (GNPs) designed using software package (COMSOL Multiphysics), the magnitude of the laminar velocity field in the flow cell, concentration distribution in the analyte stream and surface coverage of adsorbed species and average fractional surface coverage of adsorbed analyte were discussed from the model and couples of suggestion was given in order to functionalize GNPs and to increase the accuracy of the biosensor design, all above were obtained acceptable results.

Keywords: model, gold nanoparticles, biosensor, COMSOL Multiphysics

Procedia PDF Downloads 241
16560 Demonstration of Powering up Low Power Wireless Sensor Network by RF Energy Harvesting System

Authors: Lim Teck Beng, Thiha Kyaw, Poh Boon Kiat, Lee Ngai Meng

Abstract:

This work presents discussion on the possibility of merging two emerging technologies in microwave; wireless power transfer (WPT) and RF energy harvesting. The current state of art of the two technologies is discussed and the strength and weakness of the two technologies is also presented. The equivalent circuit of wireless power transfer is modeled and explained as how the range and efficiency can be further increased by controlling certain parameters in the receiver. The different techniques of harvesting the RF energy from the ambient are also extensive study. Last but not least, we demonstrate that a low power wireless sensor network (WSN) can be power up by RF energy harvesting. The WSN is designed to transmit every 3 minutes of information containing the temperature of the environment and also the voltage of the node. One thing worth mention is both the sensors that are used for measurement are also powering up by the RF energy harvesting system.

Keywords: energy harvesting, wireless power transfer, wireless sensor network and magnetic coupled resonator

Procedia PDF Downloads 504
16559 The Effect of Normal Cervical Sagittal Configuration in the Management of Cervicogenic Dizziness: A 1-Year Randomized Controlled Study

Authors: Moustafa Ibrahim Moustafa

Abstract:

The purpose of this study was to determine the immediate and long term effects of a multimodal program, with the addition of cervical sagittal curve restoration and forward head correction, on severity of dizziness, disability, frequency of dizziness, and severity of cervical pain. 72 patients with cervicogenic dizziness, definite hypolordotic cervical spine, and forward head posture were randomized to experimental or a control group. Both groups received the multimodal program, additionally, the study group received the Denneroll cervical traction. All outcome measures were measured at three intervals. The general linear model indicated a significant group × time effects in favor of experimental group on measures of anterior head translation (F=329.4 P < .0005), cervical lordosis (F=293.7 P < .0005), severity of dizziness (F=262.1 P < .0005), disability (F=248.9 P < .0005), frequency of dizziness (F=53.9 P < .0005), and severity of cervical pain (F=350.1 P < .0005). The addition of Dennroll cervical traction to a multimodal program can positively affect dizziness management outcomes.

Keywords: randomized controlled trial, traction, dizziness, cervical

Procedia PDF Downloads 290
16558 Constructing a Co-Working Innovation Model for Multiple Art Integration: A Case Study of Children's Musical

Authors: Nai-Chia Chao, Meng-Chi Shih

Abstract:

Under today’s fast technology and massive data era, the working method start to change. In this study, based under literature meaning of “Co-working” we had implemented the new “Co-working innovation model”. Research concluded that co-working innovation model shall not be limited in co-working space but use under different field when applying multiple art integration stragies. Research show co-working should not be limited in special field or group, should be use or adapt whenever different though or ideas where found, it should be use under different field and plans.

Keywords: arts integration, co-working, children's musical

Procedia PDF Downloads 283
16557 Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs

Authors: Taysir Soliman

Abstract:

One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms.

Keywords: spectral clustering, power iteration clustering, deflation-based power iteration clustering, Apache spark, large graph

Procedia PDF Downloads 175
16556 Classification of Barley Varieties by Artificial Neural Networks

Authors: Alper Taner, Yesim Benal Oztekin, Huseyin Duran

Abstract:

In this study, an Artificial Neural Network (ANN) was developed in order to classify barley varieties. For this purpose, physical properties of barley varieties were determined and ANN techniques were used. The physical properties of 8 barley varieties grown in Turkey, namely thousand kernel weight, geometric mean diameter, sphericity, kernel volume, surface area, bulk density, true density, porosity and colour parameters of grain, were determined and it was found that these properties were statistically significant with respect to varieties. As ANN model, three models, N-l, N-2 and N-3 were constructed. The performances of these models were compared. It was determined that the best-fit model was N-1. In the N-1 model, the structure of the model was designed to be 11 input layers, 2 hidden layers and 1 output layer. Thousand kernel weight, geometric mean diameter, sphericity, kernel volume, surface area, bulk density, true density, porosity and colour parameters of grain were used as input parameter; and varieties as output parameter. R2, Root Mean Square Error and Mean Error for the N-l model were found as 99.99%, 0.00074 and 0.009%, respectively. All results obtained by the N-l model were observed to have been quite consistent with real data. By this model, it would be possible to construct automation systems for classification and cleaning in flourmills.

Keywords: physical properties, artificial neural networks, barley, classification

Procedia PDF Downloads 163
16555 Analyzing Large Scale Recurrent Event Data with a Divide-And-Conquer Approach

Authors: Jerry Q. Cheng

Abstract:

Currently, in analyzing large-scale recurrent event data, there are many challenges such as memory limitations, unscalable computing time, etc. In this research, a divide-and-conquer method is proposed using parametric frailty models. Specifically, the data is randomly divided into many subsets, and the maximum likelihood estimator from each individual data set is obtained. Then a weighted method is proposed to combine these individual estimators as the final estimator. It is shown that this divide-and-conquer estimator is asymptotically equivalent to the estimator based on the full data. Simulation studies are conducted to demonstrate the performance of this proposed method. This approach is applied to a large real dataset of repeated heart failure hospitalizations.

Keywords: big data analytics, divide-and-conquer, recurrent event data, statistical computing

Procedia PDF Downloads 148
16554 Tackling the Value-Action-Gap: Improving Civic Participation Using a Holistic Behavioral Model Approach

Authors: Long Pham, Julia Blanke

Abstract:

An increasingly popular way of establishing citizen engagement within communities is through ‘city apps’. Currently, most of these mobile applications seem to be extensions of the existing communication media, sometimes merely replicating the information available on the classical city web sites, and therefore provide minimal additional impact on citizen behavior and engagement. In order to overcome this challenge, we propose to use a holistic behavioral model to generate dynamic and contextualized app content based on optimizing well defined city-related performance goals constrained by the proposed behavioral model. In this paper, we will show how the data collected by the CorkCitiEngage project in the Irish city of Cork can be utilized to calibrate aspects of the proposed model enabling the design of a personalized citizen engagement app aiming at positively influencing people’s behavior towards more active participation in their communities. We will focus on the important aspect of intentions to act, which is essential for understanding the reasons behind the common value-action-gap being responsible for the mismatch between good intentions and actual observable behavior, and will discuss how customized app design can be based on a rigorous model of behavior optimized towards maximizing well defined city-related performance goals.

Keywords: city apps, holistic behaviour model, intention to act, value-action-gap, citizen engagement

Procedia PDF Downloads 212
16553 Variations of Total Electron Content over High Latitude Region during the 24th Solar Cycle

Authors: Arun Kumar Singh, Rupesh M. Das, Shailendra Saini

Abstract:

The effect of solar cycle and seasons on the total electron content has been investigated over high latitude region during 24th solar cycle (2010-2014). The total electron content data has been observed with the help of Global Ionospheric Scintillation and TEC monitoring (GISTM) system installed at Indian permanent scientific 'Maitri station' [70˚46’00”S 11˚43’56” E]. The dependence of TEC over a solar cycle has been examined by the performing linear regression analysis between the vertical total electron content (VTEC) and daily total sunspot numbers (SSN). It has been found that the season and level of geomagnetic activity has a considerable effect on the VTEC. It is observed that the VTEC and SSN follow better agreement during summer seasons as compared to winter and equinox seasons and extraordinary agreement during minimum phase (during the year 2010) of the solar cycle. There is a significant correlation between VTEC and SSN during quiet days of the years as compared to overall days of the years (2010-2014). Further, saturation effect has been observed during maximum phase (during the year 2014) of the 24th solar cycle. It is also found that Ap index and SSN has a linear correlation (R=0.37) and the most of the geomagnetic activity occurs during the declining phase of the solar cycle.

Keywords: high latitude ionosphere, sunspot number, correlation, vertical total electron content

Procedia PDF Downloads 182
16552 Numerical Analysis of 3D Electromagnetic Fields in Annular Induction Plasma

Authors: Abderazak Guettaf

Abstract:

The mathematical models of the physical phenomena interacting in inductive plasma were described by the physics equations of the continuous mediums. A 3D model based on magnetic potential vector and electric scalar potential (A, V) formulation is used. The finished volume method is applied to electromagnetic equation, to obtain the field distribution inside the plasma. The numerical results of the method developed on a basic model designed starting from a real three-dimensional model were exposed. From the mathematical model 3D spreading assumptions and boundary conditions, we evaluated the electric field in the load and we have developed a numerical code made under the MATLAB environment, all verifying the effectiveness and validity of this code.

Keywords: electric field, 3D magnetic potential vector and electric scalar potential (A, V) formulation, finished volumes, annular plasma

Procedia PDF Downloads 478
16551 Dissipation of Tebuconazole in Cropland Soils as Affected by Soil Factors

Authors: Bipul Behari Saha, Sunil Kumar Singh, P. Padmaja, Kamlesh Vishwakarma

Abstract:

Dissipation study of tebuconazole in alluvial, black and deep-black clayey soils collected from paddy, mango and peanut cropland of tropical agro-climatic zone of India at three concentration levels were carried out for monitoring the water contamination through persisted residual toxicity. The soil-slurry samples were analyzed by capillary GC-NPD methods followed by ultrasound-assisted extraction (UAE) technique and cleanup process. An excellent linear relationship between peak area and concentration obtained in the range 1 to 50 μgkg-1. The detection (S/N, 3 ± 0.5) and quantification (S/N, 7.5 ± 2.5) limits were 3 and 10 μgkg-1 respectively. Well spiked recoveries were achieved from 96.28 to 99.33 % at levels 5 and 20 μgkg-1 and method precision (% RSD) was ≤ 5%. The soils dissipation of tebuconazole was fitted in first order kinetic-model with half-life between 34.48 to 48.13 days. The soil organic-carbon (SOC) content correlated well with the dissipation rate constants (DRC) of the fungicide Tebuconazole. An increase in the SOC content resulted in faster dissipation. The results indicate that the soil organic carbon and tebuconazole concentrations plays dominant role in dissipation processes. The initial concentration illustrated that the degradation rate of tebuconazole in soils was concentration dependent.

Keywords: cropland soil, dissipation, laboratory incubation, tebuconazole

Procedia PDF Downloads 241
16550 Optimization Process for Ride Quality of a Nonlinear Suspension Model Based on Newton-Euler’ Augmented Formulation

Authors: Mohamed Belhorma, Aboubakar S. Bouchikhi, Belkacem Bounab

Abstract:

This paper addresses modeling a Double A-Arm suspension, a three-dimensional nonlinear model has been developed using the multibody systems formalism. Dynamical study of the different components responses was done, particularly for the wheel assembly. To validate those results, the system was constructed and simulated by RecurDyn, a professional multibody dynamics simulation software. The model has been used as the Objectif function in an optimization algorithm for ride quality improvement.

Keywords: double A-Arm suspension, multibody systems, ride quality optimization, dynamic simulation

Procedia PDF Downloads 122
16549 An Advanced Exponential Model for Seismic Isolators Having Hardening or Softening Behavior at Large Displacements

Authors: Nicolò Vaiana, Giorgio Serino

Abstract:

In this paper, an advanced Nonlinear Exponential Model (NEM), able to simulate the uniaxial dynamic behavior of seismic isolators having a continuously decreasing tangent stiffness with increasing displacement in the relatively large displacements range and a hardening or softening behavior at large displacements, is presented. The mathematical model is validated by comparing the experimental force-displacement hysteresis loops obtained during cyclic tests, conducted on a helical wire rope isolator and a recycled rubber-fiber reinforced bearing, with those predicted analytically. Good agreement between the experimental and simulated results shows that the proposed model can be an effective numerical tool to predict the force-displacement relationship of seismic isolation devices within the large displacements range. Compared to the widely used Bouc-Wen model, unable to simulate the response of seismic isolators at large displacements, the proposed one allows to avoid the numerical solution of a first order nonlinear ordinary differential equation for each time step of a nonlinear time history analysis, thus reducing the computation effort. Furthermore, the proposed model can simulate the smooth transition of the hysteresis loops from small to large displacements by adopting only one set of five parameters determined from the experimental hysteresis loops having the largest amplitude.

Keywords: base isolation, hardening behavior, nonlinear exponential model, seismic isolators, softening behavior

Procedia PDF Downloads 312
16548 The Effectiveness of Environmental Policy Instruments for Promoting Renewable Energy Consumption: Command-and-Control Policies versus Market-Based Policies

Authors: Mahmoud Hassan

Abstract:

Understanding the impact of market- and non-market-based environmental policy instruments on renewable energy consumption (REC) is crucial for the design and choice of policy packages. This study aims to empirically investigate the effect of environmental policy stringency index (EPS) and its components on REC in 27 OECD countries over the period from 1990 to 2015, and then use the results to identify what the appropriate environmental policy mix should look like. By relying on the two-step system GMM estimator, we provide evidence that increasing environmental policy stringency as a whole promotes renewable energy consumption in these 27 developed economies. Moreover, policymakers are able, through the market- and non-market-based environmental policy instruments, to increase the use of renewable energy. However, not all of these instruments are effective for achieving this goal. The results indicate that R&D subsidies and trading schemes have a positive and significant impact on REC, while taxes, feed-in tariff and emission standards have not a significant effect. Furthermore, R&D subsidies are more effective than trading schemes for stimulating the use of clean energy. These findings proved to be robust across the three alternative panel techniques used.

Keywords: environmental policy stringency, renewable energy consumption, two-step system-GMM estimation, linear dynamic panel data model

Procedia PDF Downloads 172
16547 Optimizing Mechanical Behavior of Middle Ear Prosthesis Using Finite Element Method with Material Degradation Functionally Graded Materials in Three Functions

Authors: Khatir Omar, Fekih Sidi Mohamed, Sahli Abderahmene, Benkhettou Abdelkader, Boudjemaa Ismail

Abstract:

Advancements in technology have revolutionized healthcare, with notable impacts on auditory health. This study introduces an approach aimed at optimizing materials for middle ear prostheses to enhance auditory performance. We have developed a finite element (FE) model of the ear incorporating a pure titanium TORP prosthesis, validated against experimental data. Subsequently, we applied the Functionally Graded Materials (FGM) methodology, utilizing linear, exponential, and logarithmic degradation functions to modify prosthesis materials. Biocompatible materials suitable for auditory prostheses, including Stainless Steel, titanium, and Hydroxyapatite, were investigated. The findings indicate that combinations such as Stainless Steel with titanium and Hydroxyapatite offer improved outcomes compared to pure titanium and Hydroxyapatite ceramic in terms of both displacement and stress. Additionally, personalized prostheses tailored to individual patient needs are feasible, underscoring the potential for further advancements in auditory healthcare.

Keywords: middle ear, prosthesis, ossicles, FGM, vibration analysis, finite-element method

Procedia PDF Downloads 54
16546 A Scalable Model of Fair Socioeconomic Relations Based on Blockchain and Machine Learning Algorithms-1: On Hyperinteraction and Intuition

Authors: Merey M. Sarsengeldin, Alexandr S. Kolokhmatov, Galiya Seidaliyeva, Alexandr Ozerov, Sanim T. Imatayeva

Abstract:

This series of interdisciplinary studies is an attempt to investigate and develop a scalable model of fair socioeconomic relations on the base of blockchain using positive psychology techniques and Machine Learning algorithms for data analytics. In this particular study, we use hyperinteraction approach and intuition to investigate their influence on 'wisdom of crowds' via created mobile application which was created for the purpose of this research. Along with the public blockchain and private Decentralized Autonomous Organization (DAO) which were elaborated by us on the base of Ethereum blockchain, a model of fair financial relations of members of DAO was developed. We developed a smart contract, so-called, Fair Price Protocol and use it for implementation of model. The data obtained from mobile application was analyzed by ML algorithms. A model was tested on football matches.

Keywords: blockchain, Naïve Bayes algorithm, hyperinteraction, intuition, wisdom of crowd, decentralized autonomous organization

Procedia PDF Downloads 157
16545 Using Simulation Modeling Approach to Predict USMLE Steps 1 and 2 Performances

Authors: Chau-Kuang Chen, John Hughes, Jr., A. Dexter Samuels

Abstract:

The prediction models for the United States Medical Licensure Examination (USMLE) Steps 1 and 2 performances were constructed by the Monte Carlo simulation modeling approach via linear regression. The purpose of this study was to build robust simulation models to accurately identify the most important predictors and yield the valid range estimations of the Steps 1 and 2 scores. The application of simulation modeling approach was deemed an effective way in predicting student performances on licensure examinations. Also, sensitivity analysis (a/k/a what-if analysis) in the simulation models was used to predict the magnitudes of Steps 1 and 2 affected by changes in the National Board of Medical Examiners (NBME) Basic Science Subject Board scores. In addition, the study results indicated that the Medical College Admission Test (MCAT) Verbal Reasoning score and Step 1 score were significant predictors of the Step 2 performance. Hence, institutions could screen qualified student applicants for interviews and document the effectiveness of basic science education program based on the simulation results.

Keywords: prediction model, sensitivity analysis, simulation method, USMLE

Procedia PDF Downloads 325
16544 Reduced Model Investigations Supported by Fuzzy Cognitive Map to Foster Circular Economy

Authors: A. Buruzs, M. F. Hatwágner, L. T. Kóczy

Abstract:

The aim of the present paper is to develop an integrated method that may provide assistance to decision makers during system planning, design, operation and evaluation. In order to support the realization of Circular Economy (CE), it is essential to evaluate local needs and conditions which help to select the most appropriate system components and resource needs. Each of these activities requires careful planning, however, the model of CE offers a comprehensive interdisciplinary framework. The aim of this research was to develop and to introduce a practical methodology for evaluation of local and regional opportunities to promote CE.

Keywords: circular economy, factors, fuzzy cognitive map, model reduction, sustainability

Procedia PDF Downloads 226
16543 The Use of Geographically Weighted Regression for Deforestation Analysis: Case Study in Brazilian Cerrado

Authors: Ana Paula Camelo, Keila Sanches

Abstract:

The Geographically Weighted Regression (GWR) was proposed in geography literature to allow relationship in a regression model to vary over space. In Brazil, the agricultural exploitation of the Cerrado Biome is the main cause of deforestation. In this study, we propose a methodology using geostatistical methods to characterize the spatial dependence of deforestation in the Cerrado based on agricultural production indicators. Therefore, it was used the set of exploratory spatial data analysis tools (ESDA) and confirmatory analysis using GWR. It was made the calibration a non-spatial model, evaluation the nature of the regression curve, election of the variables by stepwise process and multicollinearity analysis. After the evaluation of the non-spatial model was processed the spatial-regression model, statistic evaluation of the intercept and verification of its effect on calibration. In an analysis of Spearman’s correlation the results between deforestation and livestock was +0.783 and with soybeans +0.405. The model presented R²=0.936 and showed a strong spatial dependence of agricultural activity of soybeans associated to maize and cotton crops. The GWR is a very effective tool presenting results closer to the reality of deforestation in the Cerrado when compared with other analysis.

Keywords: deforestation, geographically weighted regression, land use, spatial analysis

Procedia PDF Downloads 345
16542 Measurement of 238U, 232Th and 40K in Soil Samples Collected from Coal City Dhanbad, India

Authors: Zubair Ahmad

Abstract:

Specific activities of the natural radionuclides 238U, 232Th and 40K were measured by using γ - ray spectrometric technique in soil samples collected from the city of Dhanbad, which is located near coal mines. Mean activity values for 238U, 232Th and 40K were found to be 60.29 Bq/kg, 64.50 Bq/kg and 481.0 Bq/kg, respectively. Mean radium equivalent activity, absorbed dose rate, outdoor dose, external hazard index, internal hazard index, for the area under study were determined as 189.53 Bq/kg, 87.21 nGy/h, 0.37 mSv/y, 0.52 and 0.64, respectively. The annual effective dose to the general public was found 0.44 mSv/y. This value lies well below the limit of 1 mSv/y as recommended by International Commission on Radiological Protection. Measured values were found safe for environment and public health.

Keywords: coal city Dhanbad, gamma-ray spectroscopy, natural radioactivity, soil samples

Procedia PDF Downloads 251
16541 Dislocation Density-Based Modeling of the Grain Refinement in Surface Mechanical Attrition Treatment

Authors: Reza Miresmaeili, Asghar Heydari Astaraee, Fereshteh Dolati

Abstract:

In the present study, an analytical model based on dislocation density model was developed to simulate grain refinement in surface mechanical attrition treatment (SMAT). The correlation between SMAT time and development in plastic strain on one hand, and dislocation density evolution, on the other hand, was established to simulate the grain refinement in SMAT. A dislocation density-based constitutive material law was implemented using VUHARD subroutine. A random sequence of shots is taken into consideration for multiple impacts model using Python programming language by utilizing a random function. The simulation technique was to model each impact in a separate run and then transferring the results of each run as initial conditions for the next run (impact). The developed Finite Element (FE) model of multiple impacts describes the coverage evolution in SMAT. Simulations were run to coverage levels as high as 4500%. It is shown that the coverage implemented in the FE model is equal to the experimental coverage. It is depicted that numerical SMAT coverage parameter is adequately conforming to the well-known Avrami model. Comparison between numerical results and experimental measurements for residual stresses and depth of deformation layers confirms the performance of the established FE model for surface engineering evaluations in SMA treatment. X-ray diffraction (XRD) studies of grain refinement, including resultant grain size and dislocation density, were conducted to validate the established model. The full width at half-maximum in XRD profiles can be used to measure the grain size. Numerical results and experimental measurements of grain refinement illustrate good agreement and show the capability of established FE model to predict the gradient microstructure in SMA treatment.

Keywords: dislocation density, grain refinement, severe plastic deformation, simulation, surface mechanical attrition treatment

Procedia PDF Downloads 123
16540 Generating High-Frequency Risk Factor Collections with Transformer

Authors: Wenyan Xu, Rundong Wang, Chen Li, Yonghong Hu, Zhonghua Lu

Abstract:

In the field of quantitative trading, it is common to find patterns in short-term volatile trends of the market. These patterns are known as High-Frequency (HF) risk factors, serving as effective indicators of future stock price volatility. However, in the past, these risk factors were usually generated by traditional financial models, and the validity of these risk factors is heavily based on domain-specific knowledge manually added instead of extensive market data. Inspired by symbolic regression (SR), the task of inferring mathematical laws from existing data, we take the extraction of formulaic risk factors from high-frequency trading (HFT) market data as an SR task. In this paper, we challenge the procedure of manually constructing risk factors and propose an end-to-end methodology, Intraday Risk Factor Transformer (IRFT) to directly predict the full formulaic factors, constants included. Specifically, we utilize a hybrid symbolic-numeric vocabulary where symbolic tokens denote operators/stock features and numeric tokens denote constants. Then, we train a Transformer model on the HFT dataset to directly generate complete formulaic HF risk factors without relying on the skeleton, which is a parametric function using a pre-defined list of operators – typically, the math operations (+, ×, /) and functions(√x, log x, cos x). It determines the general shape of the stock volatility law up to a choice of constants, e.g., f(x) = tan(ax+b) (x is the stock price). We further refine predicted constants(a,b) using the Broyden–Fletcher–Goldfarb–Shanno algorithm (BFGS) as informed guesses to mitigate non-linear issues. Compared to the 10 approaches in SRBench, which is a living benchmark for SR, IRFT gains a 30% excess investment return on the HS300 and S&P500 datasets, with inference times orders of magnitude faster than theirs in HF risk factor mining tasks.

Keywords: transformer, factor-mining language model, highfrequency risk factor collections

Procedia PDF Downloads 12
16539 The Grand Unified Theory of Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow Model

Authors: Tory Erickson

Abstract:

The "Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow" (BST-SCWPDF) Model introduces a framework aimed at unifying general relativity (GR) and quantum mechanics (QM). By proposing a concept of bidirectional spacetime, this model suggests that time can flow in more than one direction, thus offering a perspective on temporal dynamics. Integrated with spatial covariance and wave-particle duality in spacetime flow, the BST-SCWPDF Model resolves long-standing discrepancies between GR and QM. This unified theory has profound implications for quantum gravity, potentially offering insights into quantum entanglement, the collapse of the wave function, and the fabric of spacetime itself. The Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow" (BST-SCWPDF) Model offers researchers a framework for a better understanding of theoretical physics.

Keywords: astrophysics, quantum mechanics, general relativity, unification theory, theoretical physics

Procedia PDF Downloads 62