Search results for: estimations of probability distributions
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1900

Search results for: estimations of probability distributions

1750 Urban-Rural Inequality in Mexico after Nafta: A Quantile Regression Analysis

Authors: Rene Valdiviezo-Issa

Abstract:

In this paper, we use Mexico’s Households Income and Expenditures (ENIGH) survey to explain the behaviour that the urban-rural expenditure gap has had since Mexico’s incorporation to the North American Free Trade Agreement (NAFTA) in 1994 and we compare it with the latest available survey, which took place in 2014. We use real trimestral expenditure per capita (RTEPC) as the measure of welfare. We use quantile regressions and a quantile regression decomposition to describe the gap between urban and rural distributions of log RTEPC. We discover that the decrease in the difference between the urban and rural distributions of log RTEPC, or inequality, is motivated because of a deprivation of the urban areas, in very specific characteristics, rather than an improvement of the urban areas. When using the decomposition we observe that the gap is primarily brought about because differences in returns to covariates between the urban and rural areas.

Keywords: quantile regression, urban-rural inequality, inequality in Mexico, income decompositon

Procedia PDF Downloads 256
1749 Removing Maturational Influences from Female Youth Swimming: The Application of Corrective Adjustment Procedures

Authors: Clorinda Hogan, Shaun Abbott, Mark Halaki, Marcela Torres Catiglioni, Goshi Yamauchi, Lachlan Mitchell, James Salter, Michael Romann, Stephen Cobley

Abstract:

Introduction: Common annual age-group competition structures unintentionally introduce participation inequalities, performance (dis)advantages and selection biases due to the effect of maturational variation between youth swimmers. On this basis, there are implications for improving performance evaluation strategies. Therefore the aim was to: (1) To determine maturity timing distributions in female youth swimming; (2) quantify the relationship between maturation status and 100-m FC performance; (3) apply Maturational-based Corrective Adjustment Procedures (Mat-CAPs) for removal of maturational status performance influences. Methods: (1) Cross-sectional analysis of 663 female (10-15 years) swimmers who underwent assessment of anthropometrics (mass, height and sitting height) and estimations of maturity timing and offset. (2) 100-m front-crawl performance (seconds) was assessed at Australian regional, state, and national-level competitions between 2016-2020. To determine the relationship between maturation status and 100-m front-crawl performance, MO was plotted against 100-m FC performance time. The expected maturity status - performance relationship for females aged 10-15 years of age was obtained through a quadratic function (y = ax2 + bx + c) from unstandardized coefficients. The regression equation was subsequently used for Mat-CAPs. (3) Participants aged 10-13 years were categorised into maturity-offset categories. Maturity offset distributions for Raw (‘All’, ‘Top 50%’ & ‘Top 25%’) and Correctively Adjusted swim times were examined. Chi-square, Cramer’s V and ORs determined the occurrence of maturation biases for each age group and selection level. Results—: (1) Maturity timing distributions illustrated overrepresentation of ‘normative’ maturing swimmers (11.82 ± 0.40 years), with a descriptive shift toward the early maturing relative to the normative population. (2) A curvilinear relationship between maturity-offset and swim performance was identified (R2 = 0.53, P < 0.001) and subsequently utilised for Mat-CAPs. (3) Raw maturity offset categories identified partial maturation status skewing towards biologically older swimmers at 10/11 and 12 years, with effect magnitudes increasing in the ‘Top 50%’ and ‘25%’ of performance times. Following Mat-CAPs application, maturity offset biases were removed in similar age groups and selection levels. When adjusting performance times for maturity offset, Mat-CAPs was successful in mitigating against maturational biases until approximately 1-year post Peak Height Velocity. The overrepresentation of ‘normative’ maturing female swimmers contrasted with the substantial overrepresentation of ‘early’ maturing male swimmers found previously in 100-m front-crawl. These findings suggest early maturational timing is not advantageous in females, but findings associated with Aim 2, highlight how advanced maturational status remained beneficial to performance. Observed differences between female and male maturational biases may relate to the differential impact of physiological development during pubertal years. Females experience greater increases of fat mass and potentially differing changes in body shape which can negatively affect swim performance. Conclusions: Transient maturation status-based participation and performance advantages were apparent within a large sample of Australian female youth 100-m FC swimmers. By removing maturity status performance biases within female youth swimming, Mat-CAPs could help improve participation experiences and the accuracy of identifying genuinely skilled female youth swimmers.

Keywords: athlete development, long-term sport participation, performance evaluation, talent identification, youth competition

Procedia PDF Downloads 155
1748 Prioritized Processor-Sharing with a Maximum Permissible Sojourn Time

Authors: Yoshiaki Shikata

Abstract:

A prioritized processor-sharing (PS) system with a maximum permissible sojourn time (MPST) is proposed. In this PS system, a higher-priority request is allocated a larger service ratio than a lower-priority request. Moreover, each request receiving service is guaranteed the maximum permissible sojourn time determined by each priority class, regardless of its service time. Arriving requests that cannot receive service due to this guarantee are rejected. We further propose a guarantee method for implementing such a system, and discuss performance evaluation procedures for the resulting system. Practical performance measures, such as the relationships between the loss probability or mean sojourn time of each class request and the maximum permissible sojourn time are evaluated via simulation. At the arrival of each class request, its acceptance or rejection is judged using extended sojourn times of all requests receiving service in the server. As the MPST increases, the mean sojourn time increases almost linearly. However, the logarithm of the loss probability decreases almost linearly. Moreover with an MPST, the difference in the mean sojourn time for different MPSTs increases with the traffic rate. Conversely, the difference in the loss probability for different MPSTs decreases as the traffic rate increases.

Keywords: prioritized processor sharing, priority ratio, permissible sojourn time, loss probability, mean sojourn time, simulation

Procedia PDF Downloads 164
1747 The Effect of Non-Normality on CB-SEM and PLS-SEM Path Estimates

Authors: Z. Jannoo, B. W. Yap, N. Auchoybur, M. A. Lazim

Abstract:

The two common approaches to Structural Equation Modeling (SEM) are the Covariance-Based SEM (CB-SEM) and Partial Least Squares SEM (PLS-SEM). There is much debate on the performance of CB-SEM and PLS-SEM for small sample size and when distributions are non-normal. This study evaluates the performance of CB-SEM and PLS-SEM under normality and non-normality conditions via a simulation. Monte Carlo Simulation in R programming language was employed to generate data based on the theoretical model with one endogenous and four exogenous variables. Each latent variable has three indicators. For normal distributions, CB-SEM estimates were found to be inaccurate for small sample size while PLS-SEM could produce the path estimates. Meanwhile, for a larger sample size, CB-SEM estimates have lower variability compared to PLS-SEM. Under non-normality, CB-SEM path estimates were inaccurate for small sample size. However, CB-SEM estimates are more accurate than those of PLS-SEM for sample size of 50 and above. The PLS-SEM estimates are not accurate unless sample size is very large.

Keywords: CB-SEM, Monte Carlo simulation, normality conditions, non-normality, PLS-SEM

Procedia PDF Downloads 371
1746 Micromechanical Modeling of Fiber-Matrix Debonding in Unidirectional Composites

Authors: M. Palizvan, M. T. Abadi, M. H. Sadr

Abstract:

Due to variations in damage mechanisms in the microscale, the behavior of fiber-reinforced composites is nonlinear and difficult to model. To make use of computational advantages, homogenization method is applied to the micro-scale model in order to minimize the cost at the expense of detail of local microscale phenomena. In this paper, the effective stiffness is calculated using the homogenization of nonlinear behavior of a composite representative volume element (RVE) containing fiber-matrix debonding. The damage modes for the RVE are considered by using cohesive elements and contacts for the cohesive behavior of the interface between fiber and matrix. To predict more realistic responses of composite materials, different random distributions of fibers are proposed besides square and hexagonal arrays. It was shown that in some cases, there is quite different damage behavior in different fiber distributions. A comprehensive comparison has been made between different graphs.

Keywords: homogenization, cohesive zone model, fiber-matrix debonding, RVE

Procedia PDF Downloads 144
1745 Improving Flash Flood Forecasting with a Bayesian Probabilistic Approach: A Case Study on the Posina Basin in Italy

Authors: Zviad Ghadua, Biswa Bhattacharya

Abstract:

The Flash Flood Guidance (FFG) provides the rainfall amount of a given duration necessary to cause flooding. The approach is based on the development of rainfall-runoff curves, which helps us to find out the rainfall amount that would cause flooding. An alternative approach, mostly experimented with Italian Alpine catchments, is based on determining threshold discharges from past events and on finding whether or not an oncoming flood has its magnitude more than some critical discharge thresholds found beforehand. Both approaches suffer from large uncertainties in forecasting flash floods as, due to the simplistic approach followed, the same rainfall amount may or may not cause flooding. This uncertainty leads to the question whether a probabilistic model is preferable over a deterministic one in forecasting flash floods. We propose the use of a Bayesian probabilistic approach in flash flood forecasting. A prior probability of flooding is derived based on historical data. Additional information, such as antecedent moisture condition (AMC) and rainfall amount over any rainfall thresholds are used in computing the likelihood of observing these conditions given a flash flood has occurred. Finally, the posterior probability of flooding is computed using the prior probability and the likelihood. The variation of the computed posterior probability with rainfall amount and AMC presents the suitability of the approach in decision making in an uncertain environment. The methodology has been applied to the Posina basin in Italy. From the promising results obtained, we can conclude that the Bayesian approach in flash flood forecasting provides more realistic forecasting over the FFG.

Keywords: flash flood, Bayesian, flash flood guidance, FFG, forecasting, Posina

Procedia PDF Downloads 106
1744 Determining the Effects of Wind-Aided Midge Movement on the Probability of Coexistence of Multiple Bluetongue Virus Serotypes in Patchy Environments

Authors: Francis Mugabi, Kevin Duffy, Joseph J. Y. T Mugisha, Obiora Collins

Abstract:

Bluetongue virus (BTV) has 27 serotypes, with some of them coexisting in patchy (different) environments, which make its control difficult. Wind-aided midge movement is a known mechanism in the spread of BTV. However, its effects on the probability of coexistence of multiple BTV serotypes are not clear. Deterministic and stochastic models for r BTV serotypes in n discrete patches connected by midge and/or cattle movement are formulated and analyzed. For the deterministic model without midge and cattle movement, using the comparison principle, it is shown that if the patch reproduction number R0 < 1, i=1,2,...,n, j=1,2,...,r, all serotypes go extinct. If R^j_i0>1, competitive exclusion takes place. Using numerical simulations, it is shown that when the n patches are connected by midge movement, coexistence takes place. To account for demographic and movement variability, the deterministic model is transformed into a continuous-time Markov chain stochastic model. Utilizing a multitype branching process, it is shown that the midge movement can have a large effect on the probability of coexistence of multiple BTV serotypes. The probability of coexistence can be brought to zero when the control interventions that directly kill the adult midges are applied. These results indicate the significance of wind-aided midge movement and vector control interventions on the coexistence and control of multiple BTV serotypes in patchy environments.

Keywords: bluetongue virus, coexistence, multiple serotypes, midge movement, branching process

Procedia PDF Downloads 118
1743 Moderating Effect of Owner's Influence on the Relationship between the Probability of Client Failure and Going Concern Opinion Issuance

Authors: Mohammad Noor Hisham Osman, Ahmed Razman Abdul Latiff, Zaidi Mat Daud, Zulkarnain Muhamad Sori

Abstract:

The problem that Malaysian auditors do not issue going concern opinion (GC opinion) to seriously financially distressed companies is still a pressing issue. Policy makers, particularly the Financial Statement Review Committee (FSRC) of Malaysian Institute of Accountant, have raised this issue as early as in 2009. Similar problem happened in the US, UK, and many developing countries. It is important for auditors to issue GC opinion properly because such opinion is one signal about the viability of a company much needed by stakeholders. There are at least two unanswered questions or research gaps in the literature on determinants of GC opinion. Firstly, is client’s probability of failure associated with GC opinion issuance? Secondly, to what extent influential owners (management, family, and institution) moderate the association between client probability of failure and GC opinion issuance. The objective of this study is, therefore, twofold; (1) To examine the extent of the relationship between the probability of client failure and the issuance of GC opinion and (2) To examine the level of management, family, and institutional ownerships moderate the association between client probability of failure and the issuance of GC opinion. This study is quantitative in nature, and the sources of data are secondary (mainly company’s annual reports). A total of four hypotheses have been developed and tested on data accumulated from annual reports of seriously financially distressed Malaysian public listed companies. Data from 2006 to 2012 on a sample of 644 observations have been analyzed using panel logistic regression. It is found that certainty (rather than probability) of client failure affects the issuance of GC opinion. In addition, it is found that only the level of family ownership does positively moderate the relationship between client probability of failure and GC opinion issuance. This study is a contribution to auditing literature as its findings can enhance our understanding about audit quality; particularly on the variables that are associated with the issuance of GC opinion. The findings of this study shed light on the roles family owners in GC opinion issuance process, and this would open ways for the researcher to suggest measures that can be used to tackle the problem of auditors do not want to issue GC opinion to financially distressed clients. The measures to be suggested can be useful to policy makers in formulating future promulgations.

Keywords: audit quality, auditing, auditor characteristics, going concern opinion, Malaysia

Procedia PDF Downloads 228
1742 Applied Bayesian Regularized Artificial Neural Network for Up-Scaling Wind Speed Profile and Distribution

Authors: Aghbalou Nihad, Charki Abderafi, Saida Rahali, Reklaoui Kamal

Abstract:

Maximize the benefit from the wind energy potential is the most interest of the wind power stakeholders. As a result, the wind tower size is radically increasing. Nevertheless, choosing an appropriate wind turbine for a selected site require an accurate estimate of vertical wind profile. It is also imperative from cost and maintenance strategy point of view. Then, installing tall towers or even more expensive devices such as LIDAR or SODAR raises the costs of a wind power project. Various models were developed coming within this framework. However, they suffer from complexity, generalization and lacks accuracy. In this work, we aim to investigate the ability of neural network trained using the Bayesian Regularization technique to estimate wind speed profile up to height of 100 m based on knowledge of wind speed lower heights. Results show that the proposed approach can achieve satisfactory predictions and proof the suitability of the proposed method for generating wind speed profile and probability distributions based on knowledge of wind speed at lower heights.

Keywords: bayesian regularization, neural network, wind shear, accuracy

Procedia PDF Downloads 462
1741 Cognitive Relaying in Interference Limited Spectrum Sharing Environment: Outage Probability and Outage Capacity

Authors: Md Fazlul Kader, Soo Young Shin

Abstract:

In this paper, we consider a cognitive relay network (CRN) in which the primary receiver (PR) is protected by peak transmit power $\bar{P}_{ST}$ and/or peak interference power Q constraints. In addition, the interference effect from the primary transmitter (PT) is considered to show its impact on the performance of the CRN. We investigate the outage probability (OP) and outage capacity (OC) of the CRN by deriving closed-form expressions over Rayleigh fading channel. Results show that both the OP and OC improve by increasing the cooperative relay nodes as well as when the PT is far away from the SR.

Keywords: cognitive relay, outage, interference limited, decode-and-forward (DF)

Procedia PDF Downloads 481
1740 Modeling of System Availability and Bayesian Analysis of Bivariate Distribution

Authors: Muhammad Farooq, Ahtasham Gul

Abstract:

To meet the desired standard, it is important to monitor and analyze different engineering processes to get desired output. The bivariate distributions got a lot of attention in recent years to describe the randomness of natural as well as artificial mechanisms. In this article, a bivariate model is constructed using two independent models developed by the nesting approach to study the effect of each component on reliability for better understanding. Further, the Bayes analysis of system availability is studied by considering prior parametric variations in the failure time and repair time distributions. Basic statistical characteristics of marginal distribution, like mean median and quantile function, are discussed. We use inverse Gamma prior to study its frequentist properties by conducting Monte Carlo Markov Chain (MCMC) sampling scheme.

Keywords: reliability, system availability Weibull, inverse Lomax, Monte Carlo Markov Chain, Bayesian

Procedia PDF Downloads 48
1739 Development of Visual Working Memory Precision: A Cross-Sectional Study of Simultaneously Delayed Responses Paradigm

Authors: Yao Fu, Xingli Zhang, Jiannong Shi

Abstract:

Visual working memory (VWM) capacity is the ability to maintain and manipulate short-term information which is not currently available. It is well known for its significance to form the basis of numerous cognitive abilities and its limitation in holding information. VWM span, the most popular measurable indicator, is found to reach the adult level (3-4 items) around 12-13 years’ old, while less is known about the precision development of the VWM capacity. By using simultaneously delayed responses paradigm, the present study investigates the development of VWM precision among 6-18-year-old children and young adults, besides its possible relationships with fluid intelligence and span. Results showed that precision and span both increased with age, and precision reached the maximum in 16-17 age-range. Moreover, when remembering 3 simultaneously presented items, the probability of remembering target item correlated with fluid intelligence and the probability of wrap errors (misbinding target and non-target items) correlated with age. When remembering more items, children had worse performance than adults due to their wrap errors. Compared to span, VWM precision was effective predictor of intelligence even after controlling for age. These results suggest that unlike VWM span, precision developed in a slow, yet longer fashion. Moreover, decreasing probability of wrap errors might be the main reason for the development of precision. Last, precision correlated more closely with intelligence than span in childhood and adolescence, which might be caused by the probability of remembering target item.

Keywords: fluid intelligence, precision, visual working memory, wrap errors

Procedia PDF Downloads 247
1738 Application of Heuristic Integration Ant Colony Optimization in Path Planning

Authors: Zeyu Zhang, Guisheng Yin, Ziying Zhang, Liguo Zhang

Abstract:

This paper mainly studies the path planning method based on ant colony optimization (ACO), and proposes heuristic integration ant colony optimization (HIACO). This paper not only analyzes and optimizes the principle, but also simulates and analyzes the parameters related to the application of HIACO in path planning. Compared with the original algorithm, the improved algorithm optimizes probability formula, tabu table mechanism and updating mechanism, and introduces more reasonable heuristic factors. The optimized HIACO not only draws on the excellent ideas of the original algorithm, but also solves the problems of premature convergence, convergence to the sub optimal solution and improper exploration to some extent. HIACO can be used to achieve better simulation results and achieve the desired optimization. Combined with the probability formula and update formula, several parameters of HIACO are tested. This paper proves the principle of the HIACO and gives the best parameter range in the research of path planning.

Keywords: ant colony optimization, heuristic integration, path planning, probability formula

Procedia PDF Downloads 214
1737 Mixtures of Length-Biased Weibull Distributions for Loss Severity Modelling

Authors: Taehan Bae

Abstract:

In this paper, a class of length-biased Weibull mixtures is presented to model loss severity data. The proposed model generalizes the Erlang mixtures with the common scale parameter, and it shares many important modelling features, such as flexibility to fit various data distribution shapes and weak-denseness in the class of positive continuous distributions, with the Erlang mixtures. We show that the asymptotic tail estimate of the length-biased Weibull mixture is Weibull-type, which makes the model effective to fit loss severity data with heavy-tailed observations. A method of statistical estimation is discussed with applications on real catastrophic loss data sets.

Keywords: Erlang mixture, length-biased distribution, transformed gamma distribution, asymptotic tail estimate, EM algorithm, expectation-maximization algorithm

Procedia PDF Downloads 196
1736 Assessing the Resilience of the Insurance Industry under Solvency II

Authors: Vincenzo Russo, Rosella Giacometti

Abstract:

The paper aims to assess the insurance industry's resilience under Solvency II against adverse scenarios. Starting from the economic balance sheet available under Solvency II for insurance and reinsurance undertakings, we assume that assets and liabilities follow a bivariate geometric Brownian motion (GBM). Then, using the results available under Margrabe's formula, we establish an analytical solution to calibrate the volatility of the asset-liability ratio. In such a way, we can estimate the probability of default and the probability of breaching the undertaking's Solvency Capital Requirement (SCR). Furthermore, since estimating the volatility of the Solvency Ratio became crucial for insurers in light of the financial crises featured in the last decades, we introduce a novel measure that we call Resiliency Ratio. The Resiliency Ratio can be used, in addition to the Solvency Ratio, to evaluate the insurance industry's resilience in case of adverse scenarios. Finally, we introduce a simplified stress test tool to evaluate the economic balance sheet under stressed conditions. The model we propose is featured by analytical tractability and fast calibration procedure where only the disclosed data available under the Solvency II public reporting are needed for the calibration. Using the data published regularly by the European Insurance and Occupational Pensions Authority (EIOPA) in an aggregated form by country, an empirical analysis has been performed to calibrate the model and provide the related results at the country level.

Keywords: Solvency II, solvency ratio, volatility of the asset-liability ratio, probability of default, probability to breach the SCR, resilience ratio, stress test

Procedia PDF Downloads 54
1735 Occupational Attainment of Second Generation of Ethnic Minority Immigrants in the UK

Authors: Rukhsana Kausar, Issam Malki

Abstract:

The integration and assimilation of ethnic minority immigrants (EMIs) and their subsequent generations remains a serious unsettled issue in most of the host countries. This study conducts the labour market gender analysis to investigate specifically whether second generation of ethnic minority immigrants in the UK is gaining access to professional and managerial employment and advantaged occupational positions on par with their native counterparts. The data used to examine the labour market achievements of EMIs is taken from Labour Force Survey (LFS) for the period 2014-2018. We apply a multivalued treatment under ignorability as proposed by Cattaneo (2010), which refers to treatment effects under the assumptions of (i) selection – on – observables and (ii) common support. We report estimates of Average Treatment Effect (ATE), Average Treatment Effect on the Treated (ATET), and Potential Outcomes Means (POM) using three estimators, including the Regression Adjustment (RA), Augmented Inverse Probability Weighting (AIPW) and Inverse Probability Weighting- Regression Adjustment (IPWRA). We consider two cases: the case with four categories where the first-generation natives are the base category, the second case combine all natives as a base group. Our findings suggest the following. Under Case 1, the estimated probabilities and differences across groups are consistently similar and highly significant. As expected, first generation natives have the highest probability for higher career attainment among both men and women. The findings also suggest that first generation immigrants perform better than the remaining two groups, including the second-generation natives and immigrants. Furthermore, second generation immigrants have higher probability to attain higher professional career, while this is lower for a managerial career. Similar conclusions are reached under Case 2. That is to say that both first – generation and second – generation immigrants have a lower probability for higher career and managerial attainment. First – generation immigrants are found to perform better than second – generation immigrants.

Keywords: immigrnats, second generation, occupational attainment, ethnicity

Procedia PDF Downloads 82
1734 Time Series Simulation by Conditional Generative Adversarial Net

Authors: Rao Fu, Jie Chen, Shutian Zeng, Yiping Zhuang, Agus Sudjianto

Abstract:

Generative Adversarial Net (GAN) has proved to be a powerful machine learning tool in image data analysis and generation. In this paper, we propose to use Conditional Generative Adversarial Net (CGAN) to learn and simulate time series data. The conditions include both categorical and continuous variables with different auxiliary information. Our simulation studies show that CGAN has the capability to learn different types of normal and heavy-tailed distributions, as well as dependent structures of different time series. It also has the capability to generate conditional predictive distributions consistent with training data distributions. We also provide an in-depth discussion on the rationale behind GAN and the neural networks as hierarchical splines to establish a clear connection with existing statistical methods of distribution generation. In practice, CGAN has a wide range of applications in market risk and counterparty risk analysis: it can be applied to learn historical data and generate scenarios for the calculation of Value-at-Risk (VaR) and Expected Shortfall (ES), and it can also predict the movement of the market risk factors. We present a real data analysis including a backtesting to demonstrate that CGAN can outperform Historical Simulation (HS), a popular method in market risk analysis to calculate VaR. CGAN can also be applied in economic time series modeling and forecasting. In this regard, we have included an example of hypothetical shock analysis for economic models and the generation of potential CCAR scenarios by CGAN at the end of the paper.

Keywords: conditional generative adversarial net, market and credit risk management, neural network, time series

Procedia PDF Downloads 110
1733 A Molecular Dynamics Study on Intermittent Plasticity and Dislocation Avalanche Emissions in FCC and BCC Crystals

Authors: Javier Varillas, Jorge Alcalá

Abstract:

We investigate dislocation avalanche phenomena in face-centered cubic (FCC) and body-centered cubic (BCC) crystals using massive, large-scale molecular dynamics (MD) simulations. The analysis is focused on the intermittent development of dense dislocation arrangements subjected to uniaxial tensile straining under displacement control. We employ a novel computational scheme that allows us to inject an entangled dislocation structure in periodic MD domains. We assess the emission of plastic bursts (or dislocation avalanches) in terms of the sharp stress drops detected in the stress-strain curve. The plastic activity corresponds to the sporadic operation of specific dislocation glide processes exhibiting quiescent periods between successive avalanche events. We find that the plastic intermittences in our simulations do not overlap in time under sufficiently low strain rates as dissipation operates faster than driving, where the dense dislocation networks evolve through the emission of dislocation avalanche events whose carried slip adheres to self-organized power-law distributions. These findings enable the extension of the slip distributions obtained from strict displacement-controlled micropillar compression experiments towards smaller values of slip size. Our results furnish further understanding upon the development of entangled dislocation networks in metal plasticity, including specific mechanisms of dislocation propagation and annihilation, along with the evolution of specific dislocation populations through dislocation density analyses.

Keywords: dislocations, intermittent plasticity, molecular dynamics, slip distributions

Procedia PDF Downloads 111
1732 Accurate Binding Energy of Ytterbium Dimer from Ab Initio Calculations and Ultracold Photoassociation Spectroscopy

Authors: Giorgio Visentin, Alexei A. Buchachenko

Abstract:

Recent proposals to use Yb dimer as an optical clock and as a sensor for non-Newtonian gravity imply the knowledge of its interaction potential. Here, the ground-state Born-Oppenheimer Yb₂ potential energy curve is represented by a semi-analytical function, consisting of short- and long-range contributions. For the former, the systematic ab initio all-electron exact 2-component scalar-relativistic CCSD(T) calculations are carried out. Special care is taken to saturate diffuse basis set component with the atom- and bond-centered primitives and reach the complete basis set limit through n = D, T, Q sequence of the correlation-consistent polarized n-zeta basis sets. Similar approaches are used to the long-range dipole and quadrupole dispersion terms by implementing the CCSD(3) polarization propagator method for dynamic polarizabilities. Dispersion coefficients are then computed through Casimir-Polder integration. The semiclassical constraint on the number of the bound vibrational levels known for the ¹⁷⁴Yb isotope is used to scale the potential function. The scaling, based on the most accurate ab initio results, bounds the interaction energy of two Yb atoms within the narrow 734 ± 4 cm⁻¹ range, in reasonable agreement with the previous ab initio-based estimations. The resulting potentials can be used as the reference for more sophisticated models that go beyond the Born-Oppenheimer approximation and provide the means of their uncertainty estimations. The work is supported by Russian Science Foundation grant # 17-13-01466.

Keywords: ab initio coupled cluster methods, interaction potential, semi-analytical function, ytterbium dimer

Procedia PDF Downloads 126
1731 Stock Market Developments, Income Inequality, Wealth Inequality

Authors: Quang Dong Dang

Abstract:

This paper examines the possible effects of stock market developments by channels on income and wealth inequality. We use the Bayesian Multilevel Model with the explanatory variables of the market’s channels, such as accessibility, efficiency, and market health in six selected countries: the US, UK, Japan, Vietnam, Thailand, and Malaysia. We found that generally, the improvements in the stock market alleviate income inequality. However, stock market expansions in higher-income countries are likely to trigger income inequality. We also found that while enhancing the quality of channels of the stock market has counter-effects on wealth equality distributions, open accessibilities help reduce wealth inequality distributions within the scope of the study. In addition, the inverted U-shaped hypothesis seems not to be valid in six selected countries between the period from 2006 to 2020.

Keywords: Bayesian multilevel model, income inequality, inverted u-shaped hypothesis, stock market development, wealth inequality

Procedia PDF Downloads 68
1730 Evidence Theory Based Emergency Multi-Attribute Group Decision-Making: Application in Facility Location Problem

Authors: Bidzina Matsaberidze

Abstract:

It is known that, in emergency situations, multi-attribute group decision-making (MAGDM) models are characterized by insufficient objective data and a lack of time to respond to the task. Evidence theory is an effective tool for describing such incomplete information in decision-making models when the expert and his knowledge are involved in the estimations of the MAGDM parameters. We consider an emergency decision-making model, where expert assessments on humanitarian aid from distribution centers (HADC) are represented in q-rung ortho-pair fuzzy numbers, and the data structure is described within the data body theory. Based on focal probability construction and experts’ evaluations, an objective function-distribution centers’ selection ranking index is constructed. Our approach for solving the constructed bicriteria partitioning problem consists of two phases. In the first phase, based on the covering’s matrix, we generate a matrix, the columns of which allow us to find all possible partitionings of the HADCs with the service centers. Some constraints are also taken into consideration while generating the matrix. In the second phase, based on the matrix and using our exact algorithm, we find the partitionings -allocations of the HADCs to the centers- which correspond to the Pareto-optimal solutions. For an illustration of the obtained results, a numerical example is given for the facility location-selection problem.

Keywords: emergency MAGDM, q-rung orthopair fuzzy sets, evidence theory, HADC, facility location problem, multi-objective combinatorial optimization problem, Pareto-optimal solutions

Procedia PDF Downloads 62
1729 Forecast Based on an Empirical Probability Function with an Adjusted Error Using Propagation of Error

Authors: Oscar Javier Herrera, Manuel Angel Camacho

Abstract:

This paper addresses a cutting edge method of business demand forecasting, based on an empirical probability function when the historical behavior of the data is random. Additionally, it presents error determination based on the numerical method technique ‘propagation of errors’. The methodology was conducted characterization and process diagnostics demand planning as part of the production management, then new ways to predict its value through techniques of probability and to calculate their mistake investigated, it was tools used numerical methods. All this based on the behavior of the data. This analysis was determined considering the specific business circumstances of a company in the sector of communications, located in the city of Bogota, Colombia. In conclusion, using this application it was possible to obtain the adequate stock of the products required by the company to provide its services, helping the company reduce its service time, increase the client satisfaction rate, reduce stock which has not been in rotation for a long time, code its inventory, and plan reorder points for the replenishment of stock.

Keywords: demand forecasting, empirical distribution, propagation of error, Bogota

Procedia PDF Downloads 589
1728 Association between Healthy Eating Index-2015 Scores and the Probability of Sarcopenia in Community-Dwelling Iranian Elderly

Authors: Zahra Esmaeily, Zahra Tajari, Shahrzad Daei, Mahshid Rezaei, Atefeh Eyvazkhani, Marjan Mansouri Dara, Ahmad Reza Dorosty Motlagh, Andriko Palmowski

Abstract:

Objective: Sarcopenia (SPA) is associated with frailty and disability in the elderly. Adherence to current dietary guidelines in addition to physical activity could play a role in the prevention of muscle wasting and weakness. The Healthy Eating Index-2015 (HEI) is a tool to assess diet quality as recommended in the U.S. Dietary Guidelines for Americans. This study aimed to investigate whether there is a relationship between HEI scores and the probability of SPA (PS) among the Tehran elderly. Method: A previously validated semi-quantitative food frequency questionnaire was used to assess HEI and the dietary intake of randomly selected elderly people living in Tehran, Iran. Handgrip strength (HGS) was measured to evaluate the PS. Statistical evaluation included descriptive analysis and standard test procedures. Result: 201 subjects were included. Those probably suffering from SPA (as determined by HGS) had significantly lower HEI scores (p = 0.02). After adjusting for confounders, HEI scores and HGS were still significantly associated (adjusted R2 = 0.56, slope β = 0.03, P = 0.09). Elderly people with a low probability of SPA consumed more monounsaturated and polyunsaturated fatty acids (P = 0.06) and ingested less added sugars and saturated fats (P = 0.01 and P = 0.02, respectively). Conclusion: In this cross-sectional study, HEI scores are associated with the probability of SPA. Adhering to current dietary guidelines might contribute to ameliorating muscle strength and mass in aging individuals.

Keywords: aging, HEI-2015, Iranian, sarcopenic

Procedia PDF Downloads 170
1727 The Influence of Design Complexity of a Building Structure on the Expected Performance

Authors: Ormal Lishi

Abstract:

This research presents a computationally efficient probabilistic method to assess the performance of compartmentation walls with similar Fire Resistance Levels (FRL) but varying complexity. Specifically, a masonry brick wall and a light-steel framed (LSF) wall with comparable insulation performance are analyzed. A Monte Carlo technique, employing Latin Hypercube Sampling (LHS), is utilized to quantify uncertainties and determine the probability of failure for both walls exposed to standard and parametric fires, following ISO 834 and Eurocodes guidelines. Results show that the probability of failure for the brick masonry wall under standard fire exposure is estimated at 4.8%, while the LSF wall is 7.6%. These probabilities decrease to 0.4% and 4.8%, respectively, when subjected to parametric fires. Notably, the complex LSF wall exhibits higher variability in predicting time to failure for specific criteria compared to the less complex brick wall, especially at higher temperatures. The proposed approach highlights the need for Probabilistic Risk Assessment (PRA) to accurately evaluate the reliability and safety levels of complex designs.

Keywords: design complexity, probability of failure, monte carlo analysis, compartmentation walls, insulation

Procedia PDF Downloads 33
1726 Household Wealth and Portfolio Choice When Tail Events Are Salient

Authors: Carlson Murray, Ali Lazrak

Abstract:

Robust experimental evidence of systematic violations of expected utility (EU) establishes that individuals facing risk overweight utility from low probability gains and losses when making choices. These findings motivated development of models of preferences with probability weighting functions, such as rank dependent utility (RDU). We solve for the optimal investing strategy of an RDU investor in a dynamic binomial setting from which we derive implications for investing behavior. We show that relative to EU investors with constant relative risk aversion, commonly measured probability weighting functions produce optimal RDU terminal wealth with significant downside protection and upside exposure. We additionally find that in contrast to EU investors, RDU investors optimally choose a portfolio that contains fair bets that provide payo↵s that can be interpreted as lottery outcomes or exposure to idiosyncratic returns. In a calibrated version of the model, we calculate that RDU investors would be willing to pay 5% of their initial wealth for the freedom to trade away from an optimal EU wealth allocation. The dynamic trading strategy that supports the optimal wealth allocation implies portfolio weights that are independent of initial wealth but requires higher risky share after good stock return histories. Optimal trading also implies the possibility of non-participation when historical returns are poor. Our model fills a gap in the literature by providing new quantitative and qualitative predictions that can be tested experimentally or using data on household wealth and portfolio choice.

Keywords: behavioral finance, probability weighting, portfolio choice

Procedia PDF Downloads 399
1725 Social Capital and Adoption of Sustainable Management Practices of Non Timber Forest Product in Cameroon

Authors: Eke Bala Sophie Michelle

Abstract:

The renewable resource character of NTFPs is an opportunity to its sustainability, this study analyzed the role of social capital in the adoption of sustainable management practices of NTFPs by households in the community forest (CF) Morikouali-ye. The analysis shows that 67% of households surveyed perceive the level of degradation of NTFPs in their CF as time passes and are close to 74% for adoption of sustainable management practices of NTFPs that are domestication, sustainable management of the CF, the logging ban trees and uprooting plants, etc. 26% refused to adopt these practices estimate that, at 39% it is better to promote logging in the CF. The estimated probit model shows that social capital through trust, solidarity and social inclusion significantly influences the probability of households to adopt sustainable NTFP management practices. In addition, age, education level and income from the sale of NTFPs have a significant impact on the probability of adoption. The probability of adoption increases with the level of education and confidence among households. So should they be animated by a spirit of solidarity and trust and not let a game of competition for sustainable management of NTFPs in their CF.

Keywords: community forest, social capital, NTFP, trust, solidarity, social inclusion, sustainable management

Procedia PDF Downloads 334
1724 A Deterministic Large Deviation Model Based on Complex N-Body Systems

Authors: David C. Ni

Abstract:

In the previous efforts, we constructed N-Body Systems by an extended Blaschke product (EBP), which represents a non-temporal and nonlinear extension of Lorentz transformation. In this construction, we rely only on two parameters, nonlinear degree, and relative momentum to characterize the systems. We further explored root computation via iteration with an algorithm extended from Jenkins-Traub method. The solution sets demonstrate a form of σ+ i [-t, t], where σ and t are the real numbers, and the [-t, t] shows various canonical distributions. In this paper, we correlate the convergent sets in the original domain with solution sets, which demonstrating large-deviation distributions in the codomain. We proceed to compare our approach with the formula or principles, such as Donsker-Varadhan and Wentzell-Freidlin theories. The deterministic model based on this construction allows us to explore applications in the areas of finance and statistical mechanics.

Keywords: nonlinear Lorentz transformation, Blaschke equation, iteration solutions, root computation, large deviation distribution, deterministic model

Procedia PDF Downloads 365
1723 Numerical Simulation of the Coal Spontaneous Combustion Dangerous Area in Composite Long-Wall Gobs

Authors: Changshan Zhang, Zhijin Yu, Shixing Fan

Abstract:

A comprehensive hazard evaluation for coal self-heating in composite long-wall gobs is heavily dependent on computational simulation. In this study, the spatial distributions of cracks which caused significant air leakage were simulated by universal distinct element code (UDEC) simulation. Based on the main routes of air leakage and characteristics of coal self-heating, a computational fluid dynamics (CFD) modeling was conducted to model the coal spontaneous combustion dangerous area in composite long-wall gobs. The results included the oxygen concentration distributions and temperature profiles showed that the numerical approach is validated by comparison with the test data. Furthermore, under the conditions of specific engineering, the major locations where some techniques for extinguishing and preventing long-wall gob fires need to be put into practice were also examined.

Keywords: computational simulation, UDEC simulation, coal self-heating, CFD modeling, long-wall gobs

Procedia PDF Downloads 278
1722 A Copula-Based Approach for the Assessment of Severity of Illness and Probability of Mortality: An Exploratory Study Applied to Intensive Care Patients

Authors: Ainura Tursunalieva, Irene Hudson

Abstract:

Continuous improvement of both the quality and safety of health care is an important goal in Australia and internationally. The intensive care unit (ICU) receives patients with a wide variety of and severity of illnesses. Accurately identifying patients at risk of developing complications or dying is crucial to increasing healthcare efficiency. Thus, it is essential for clinicians and researchers to have a robust framework capable of evaluating the risk profile of a patient. ICU scoring systems provide such a framework. The Acute Physiology and Chronic Health Evaluation III and the Simplified Acute Physiology Score II are ICU scoring systems frequently used for assessing the severity of acute illness. These scoring systems collect multiple risk factors for each patient including physiological measurements then render the assessment outcomes of individual risk factors into a single numerical value. A higher score is related to a more severe patient condition. Furthermore, the Mortality Probability Model II uses logistic regression based on independent risk factors to predict a patient’s probability of mortality. An important overlooked limitation of SAPS II and MPM II is that they do not, to date, include interaction terms between a patient’s vital signs. This is a prominent oversight as it is likely there is an interplay among vital signs. The co-existence of certain conditions may pose a greater health risk than when these conditions exist independently. One barrier to including such interaction terms in predictive models is the dimensionality issue as it becomes difficult to use variable selection. We propose an innovative scoring system which takes into account a dependence structure among patient’s vital signs, such as systolic and diastolic blood pressures, heart rate, pulse interval, and peripheral oxygen saturation. Copulas will capture the dependence among normally distributed and skewed variables as some of the vital sign distributions are skewed. The estimated dependence parameter will then be incorporated into the traditional scoring systems to adjust the points allocated for the individual vital sign measurements. The same dependence parameter will also be used to create an alternative copula-based model for predicting a patient’s probability of mortality. The new copula-based approach will accommodate not only a patient’s trajectories of vital signs but also the joint dependence probabilities among the vital signs. We hypothesise that this approach will produce more stable assessments and lead to more time efficient and accurate predictions. We will use two data sets: (1) 250 ICU patients admitted once to the Chui Regional Hospital (Kyrgyzstan) and (2) 37 ICU patients’ agitation-sedation profiles collected by the Hunter Medical Research Institute (Australia). Both the traditional scoring approach and our copula-based approach will be evaluated using the Brier score to indicate overall model performance, the concordance (or c) statistic to indicate the discriminative ability (or area under the receiver operating characteristic (ROC) curve), and goodness-of-fit statistics for calibration. We will also report discrimination and calibration values and establish visualization of the copulas and high dimensional regions of risk interrelating two or three vital signs in so-called higher dimensional ROCs.

Keywords: copula, intensive unit scoring system, ROC curves, vital sign dependence

Procedia PDF Downloads 127
1721 Modeling Depth Averaged Velocity and Boundary Shear Stress Distributions

Authors: Ebissa Gadissa Kedir, C. S. P. Ojha, K. S. Hari Prasad

Abstract:

In the present study, the depth-averaged velocity and boundary shear stress in non-prismatic compound channels with three different converging floodplain angles ranging from 1.43ᶱ to 7.59ᶱ have been studied. The analytical solutions were derived by considering acting forces on the channel beds and walls. In the present study, five key parameters, i.e., non-dimensional coefficient, secondary flow term, secondary flow coefficient, friction factor, and dimensionless eddy viscosity, were considered and discussed. An expression for non-dimensional coefficient and integration constants was derived based on the boundary conditions. The model was applied to different data sets of the present experiments and experiments from other sources, respectively, to examine and analyse the influence of floodplain converging angles on depth-averaged velocity and boundary shear stress distributions. The results show that the non-dimensional parameter plays important in portraying the variation of depth-averaged velocity and boundary shear stress distributions with different floodplain converging angles. Thus, the variation of the non-dimensional coefficient needs attention since it affects the secondary flow term and secondary flow coefficient in both the main channel and floodplains. The analysis shows that the depth-averaged velocities are sensitive to a shear stress-dependent model parameter non-dimensional coefficient, and the analytical solutions are well agreed with experimental data when five parameters are included. It is inferred that the developed model may facilitate the interest of others in complex flow modeling.

Keywords: depth-average velocity, converging floodplain angles, non-dimensional coefficient, non-prismatic compound channels

Procedia PDF Downloads 50