Search results for: joint probability distribution function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11002

Search results for: joint probability distribution function

10912 Quantum Mechanics Approach for Ruin Probability

Authors: Ahmet Kaya

Abstract:

Incoming cash flows and outgoing claims play an important role to determine how is companies’ profit or loss. In this matter, ruin probability provides to describe vulnerability of the companies against ruin. Quantum mechanism is one of the significant approaches to model ruin probability as stochastically. Using the Hamiltonian method, we have performed formalisation of quantum mechanics < x|e-ᵗᴴ|x' > and obtained the transition probability of 2x2 and 3x3 matrix as traditional and eigenvector basis where A is a ruin operator and H|x' > is a Schroedinger equation. This operator A and Schroedinger equation are defined by a Hamiltonian matrix H. As a result, probability of not to be in ruin can be simulated and calculated as stochastically.

Keywords: ruin probability, quantum mechanics, Hamiltonian technique, operator approach

Procedia PDF Downloads 308
10911 Survival and Hazard Maximum Likelihood Estimator with Covariate Based on Right Censored Data of Weibull Distribution

Authors: Al Omari Mohammed Ahmed

Abstract:

This paper focuses on Maximum Likelihood Estimator with Covariate. Covariates are incorporated into the Weibull model. Under this regression model with regards to maximum likelihood estimator, the parameters of the covariate, shape parameter, survival function and hazard rate of the Weibull regression distribution with right censored data are estimated. The mean square error (MSE) and absolute bias are used to compare the performance of Weibull regression distribution. For the simulation comparison, the study used various sample sizes and several specific values of the Weibull shape parameter.

Keywords: weibull regression distribution, maximum likelihood estimator, survival function, hazard rate, right censoring

Procedia PDF Downloads 417
10910 Effect of Laser Input Energy on the Laser Joining of Polyethylene Terephthalate to Titanium

Authors: Y. J. Chen, T. M. Yue, Z. N. Guo

Abstract:

This paper reports the effects of laser energy on the characteristics of bubbles generated in the weld zone and the formation of new chemical bonds at the Polyethylene Terephthalate (PET)/Ti joint interface in laser joining of PET to Ti. The samples were produced by using different laser energies ranging from 1.5 J – 6 J in steps of 1.5 J, while all other joining parameters remained unchanged. The types of chemical bonding at the joint interface were analysed by the x-ray photoelectron spectroscopy (XPS) depth-profiling method. The results show that the characteristics of the bubbles and the thickness of the chemically bonded interface, which contains the laser generated bonds of Ti–C and Ti–O, increase markedly with increasing laser energy input. The tensile failure load of the joint depends on the combined effect of the amount and distribution of the bubbles formed and the chemical bonding intensity of the joint interface.

Keywords: laser direct joining, Ti/PET interface, laser energy, XPS depth profiling, chemical bond, tensile failure load

Procedia PDF Downloads 185
10909 Seismic Response Analysis of Frame Structures Based on Super Joint Element Model

Authors: Li Xu, Yang Hong, T. Zhao Wen

Abstract:

Experimental results of many RC beam-column subassemblage indicate that slippage of longitudinal beam rebar within the joint and the shear deformation of joint core have significant influence on seismic behavior of the subassemblage. However, rigid joint assumption has been generally used in the seismic response analysis of RC frames, in which two kinds of inelastic deformation of joint have been ignored. Based on OpenSees platform, ‘Super Joint Element Model’ with more detailed inelastic mechanism is used to simulate the inelastic response of joints. Two finite element models of typical RC plane frame, namely considering or ignoring the inelastic deformation of joint respectively, were established and analyzed under seven strong earthquake waves. The simulated global and local inelastic deformations of the RC plane frame is shown and discussed. The analyses also confirm the security of the earthquake-resistant frame designed according to Chinese codes.

Keywords: frame structure, beam-column joint, longitudinal bar slippage, shear deformation, nonlinear analysis

Procedia PDF Downloads 386
10908 Daily Probability Model of Storm Events in Peninsular Malaysia

Authors: Mohd Aftar Abu Bakar, Noratiqah Mohd Ariff, Abdul Aziz Jemain

Abstract:

Storm Event Analysis (SEA) provides a method to define rainfalls events as storms where each storm has its own amount and duration. By modelling daily probability of different types of storms, the onset, offset and cycle of rainfall seasons can be determined and investigated. Furthermore, researchers from the field of meteorology will be able to study the dynamical characteristics of rainfalls and make predictions for future reference. In this study, four categories of storms; short, intermediate, long and very long storms; are introduced based on the length of storm duration. Daily probability models of storms are built for these four categories of storms in Peninsular Malaysia. The models are constructed by using Bernoulli distribution and by applying linear regression on the first Fourier harmonic equation. From the models obtained, it is found that daily probability of storms at the Eastern part of Peninsular Malaysia shows a unimodal pattern with high probability of rain beginning at the end of the year and lasting until early the next year. This is very likely due to the Northeast monsoon season which occurs from November to March every year. Meanwhile, short and intermediate storms at other regions of Peninsular Malaysia experience a bimodal cycle due to the two inter-monsoon seasons. Overall, these models indicate that Peninsular Malaysia can be divided into four distinct regions based on the daily pattern for the probability of various storm events.

Keywords: daily probability model, monsoon seasons, regions, storm events

Procedia PDF Downloads 321
10907 Modeling Binomial Dependent Distribution of the Values: Synthesis Tables of Probabilities of Errors of the First and Second Kind of Biometrics-Neural Network Authentication System

Authors: B. S.Akhmetov, S. T. Akhmetova, D. N. Nadeyev, V. Yu. Yegorov, V. V. Smogoonov

Abstract:

Estimated probabilities of errors of the first and second kind for nonideal biometrics-neural transducers 256 outputs, the construction of nomograms based error probability of 'own' and 'alien' from the mathematical expectation and standard deviation of the normalized measures Hamming.

Keywords: modeling, errors, probability, biometrics, neural network, authentication

Procedia PDF Downloads 464
10906 Parameter Estimation for the Mixture of Generalized Gamma Model

Authors: Wikanda Phaphan

Abstract:

Mixture generalized gamma distribution is a combination of two distributions: generalized gamma distribution and length biased generalized gamma distribution. These two distributions were presented by Suksaengrakcharoen and Bodhisuwan in 2014. The findings showed that probability density function (pdf) had fairly complexities, so it made problems in estimating parameters. The problem occurred in parameter estimation was that we were unable to calculate estimators in the form of critical expression. Thus, we will use numerical estimation to find the estimators. In this study, we presented a new method of the parameter estimation by using the expectation – maximization algorithm (EM), the conjugate gradient method, and the quasi-Newton method. The data was generated by acceptance-rejection method which is used for estimating α, β, λ and p. λ is the scale parameter, p is the weight parameter, α and β are the shape parameters. We will use Monte Carlo technique to find the estimator's performance. Determining the size of sample equals 10, 30, 100; the simulations were repeated 20 times in each case. We evaluated the effectiveness of the estimators which was introduced by considering values of the mean squared errors and the bias. The findings revealed that the EM-algorithm had proximity to the actual values determined. Also, the maximum likelihood estimators via the conjugate gradient and the quasi-Newton method are less precision than the maximum likelihood estimators via the EM-algorithm.

Keywords: conjugate gradient method, quasi-Newton method, EM-algorithm, generalized gamma distribution, length biased generalized gamma distribution, maximum likelihood method

Procedia PDF Downloads 200
10905 Rational Probabilistic Method for Calculating Thermal Cracking Risk of Mass Concrete Structures

Authors: Naoyuki Sugihashi, Toshiharu Kishi

Abstract:

The probability of occurrence of thermal cracks in mass concrete in Japan is evaluated by the cracking probability diagram that represents the relationship between the thermal cracking index and the probability of occurrence of cracks in the actual structure. In this paper, we propose a method to directly calculate the cracking probability, following a probabilistic theory by modeling the variance of tensile stress and tensile strength. In this method, the relationship between the variance of tensile stress and tensile strength, the thermal cracking index, and the cracking probability are formulated and presented. In addition, standard deviation of tensile stress and tensile strength was identified, and the method of calculating cracking probability in a general construction controlled environment was also demonstrated.

Keywords: thermal crack control, mass concrete, thermal cracking probability, durability of concrete, calculating method of cracking probability

Procedia PDF Downloads 306
10904 Assessing Effects of an Intervention on Bottle-Weaning and Reducing Daily Milk Intake from Bottles in Toddlers Using Two-Part Random Effects Models

Authors: Yungtai Lo

Abstract:

Two-part random effects models have been used to fit semi-continuous longitudinal data where the response variable has a point mass at 0 and a continuous right-skewed distribution for positive values. We review methods proposed in the literature for analyzing data with excess zeros. A two-part logit-log-normal random effects model, a two-part logit-truncated normal random effects model, a two-part logit-gamma random effects model, and a two-part logit-skew normal random effects model were used to examine effects of a bottle-weaning intervention on reducing bottle use and daily milk intake from bottles in toddlers aged 11 to 13 months in a randomized controlled trial. We show in all four two-part models that the intervention promoted bottle-weaning and reduced daily milk intake from bottles in toddlers drinking from a bottle. We also show that there are no differences in model fit using either the logit link function or the probit link function for modeling the probability of bottle-weaning in all four models. Furthermore, prediction accuracy of the logit or probit link function is not sensitive to the distribution assumption on daily milk intake from bottles in toddlers not off bottles.

Keywords: two-part model, semi-continuous variable, truncated normal, gamma regression, skew normal, Pearson residual, receiver operating characteristic curve

Procedia PDF Downloads 326
10903 Conflation Methodology Applied to Flood Recovery

Authors: Eva L. Suarez, Daniel E. Meeroff, Yan Yong

Abstract:

Current flooding risk modeling focuses on resilience, defined as the probability of recovery from a severe flooding event. However, the long-term damage to property and well-being by nuisance flooding and its long-term effects on communities are not typically included in risk assessments. An approach was developed to address the probability of recovering from a severe flooding event combined with the probability of community performance during a nuisance event. A consolidated model, namely the conflation flooding recovery (&FR) model, evaluates risk-coping mitigation strategies for communities based on the recovery time from catastrophic events, such as hurricanes or extreme surges, and from everyday nuisance flooding events. The &FR model assesses the variation contribution of each independent input and generates a weighted output that favors the distribution with minimum variation. This approach is especially useful if the input distributions have dissimilar variances. The &FR is defined as a single distribution resulting from the product of the individual probability density functions. The resulting conflated distribution resides between the parent distributions, and it infers the recovery time required by a community to return to basic functions, such as power, utilities, transportation, and civil order, after a flooding event. The &FR model is more accurate than averaging individual observations before calculating the mean and variance or averaging the probabilities evaluated at the input values, which assigns the same weighted variation to each input distribution. The main disadvantage of these traditional methods is that the resulting measure of central tendency is exactly equal to the average of the input distribution’s means without the additional information provided by each individual distribution variance. When dealing with exponential distributions, such as resilience from severe flooding events and from nuisance flooding events, conflation results are equivalent to the weighted least squares method or best linear unbiased estimation. The combination of severe flooding risk with nuisance flooding improves flood risk management for highly populated coastal communities, such as in South Florida, USA, and provides a method to estimate community flood recovery time more accurately from two different sources, severe flooding events and nuisance flooding events.

Keywords: community resilience, conflation, flood risk, nuisance flooding

Procedia PDF Downloads 67
10902 Analytical Downlink Effective SINR Evaluation in LTE Networks

Authors: Marwane Ben Hcine, Ridha Bouallegue

Abstract:

The aim of this work is to provide an original analytical framework for downlink effective SINR evaluation in LTE networks. The classical single carrier SINR performance evaluation is extended to multi-carrier systems operating over frequency selective channels. Extension is achieved by expressing the link outage probability in terms of the statistics of the effective SINR. For effective SINR computation, the exponential effective SINR mapping (EESM) method is used on this work. Closed-form expression for the link outage probability is achieved assuming a log skew normal approximation for single carrier case. Then we rely on the lognormal approximation to express the exponential effective SINR distribution as a function of the mean and standard deviation of the SINR of a generic subcarrier. Achieved formulas is easily computable and can be obtained for a user equipment (UE) located at any distance from its serving eNodeB. Simulations show that the proposed framework provides results with accuracy within 0.5 dB.

Keywords: LTE, OFDMA, effective SINR, log skew normal approximation

Procedia PDF Downloads 338
10901 A Data Driven Approach for the Degradation of a Lithium-Ion Battery Based on Accelerated Life Test

Authors: Alyaa M. Younes, Nermine Harraz, Mohammad H. Elwany

Abstract:

Lithium ion batteries are currently used for many applications including satellites, electric vehicles and mobile electronics. Their ability to store relatively large amount of energy in a limited space make them most appropriate for critical applications. Evaluation of the life of these batteries and their reliability becomes crucial to the systems they support. Reliability of Li-Ion batteries has been mainly considered based on its lifetime. However, another important factor that can be considered critical in many applications such as in electric vehicles is the cycle duration. The present work presents the results of an experimental investigation on the degradation behavior of a Laptop Li-ion battery (type TKV2V) and the effect of applied load on the battery cycle time. The reliability was evaluated using an accelerated life test. Least squares linear regression with median rank estimation was used to estimate the Weibull distribution parameters needed for the reliability functions estimation. The probability density function, failure rate and reliability function under each of the applied loads were evaluated and compared. An inverse power model is introduced that can predict cycle time at any stress level given.

Keywords: accelerated life test, inverse power law, lithium-ion battery, reliability evaluation, Weibull distribution

Procedia PDF Downloads 147
10900 Co-Movement between Financial Assets: An Empirical Study on Effects of the Depreciation of Yen on Asia Markets

Authors: Yih-Wenn Laih

Abstract:

In recent times, the dependence and co-movement among international financial markets have become stronger than in the past, as evidenced by commentaries in the news media and the financial sections of newspapers. Studying the co-movement between returns in financial markets is an important issue for portfolio management and risk management. The realization of co-movement helps investors to identify the opportunities for international portfolio management in terms of asset allocation and pricing. Since the election of the new Prime Minister, Shinzo Abe, in November 2012, the yen has weakened against the US dollar from the 80 to the 120 level. The policies, known as “Abenomics,” are to encourage private investment through a more aggressive mix of monetary and fiscal policy. Given the close economic relations and competitions among Asia markets, it is interesting to discover the co-movement relations, affected by the depreciation of yen, between stock market of Japan and 5 major Asia stock markets, including China, Hong Kong, Korea, Singapore, and Taiwan. Specifically, we devote ourselves to measure the co-movement of stock markets between Japan and each one of the 5 Asia stock markets in terms of rank correlation coefficients. To compute the coefficients, return series of each stock market is first fitted by a skewed-t GARCH (generalized autoregressive conditional heteroscedasticity) model. Secondly, to measure the dependence structure between matched stock markets, we employ the symmetrized Joe-Clayton (SJC) copula to calculate the probability density function of paired skewed-t distributions. The joint probability density function is then utilized as the scoring scheme to optimize the sequence alignment by dynamic programming method. Finally, we compute the rank correlation coefficients (Kendall's  and Spearman's ) between matched stock markets based on their aligned sequences. We collect empirical data of 6 stock indexes from Taiwan Economic Journal. The data is sampled at a daily frequency covering the period from January 1, 2013 to July 31, 2015. The empirical distributions of returns indicate fatter tails than the normal distribution. Therefore, the skewed-t distribution and SJC copula are appropriate for characterizing the data. According to the computed Kendall’s τ, Korea has the strongest co-movement relation with Japan, followed by Taiwan, China, and Singapore; the weakest is Hong Kong. On the other hand, the Spearman’s ρ reveals that the strength of co-movement between markets with Japan in decreasing order are Korea, China, Taiwan, Singapore, and Hong Kong. We explore the effects of “Abenomics” on Asia stock markets by measuring the co-movement relation between Japan and five major Asia stock markets in terms of rank correlation coefficients. The matched markets are aligned by a hybrid method consisting of GARCH, copula and sequence alignment. Empirical experiments indicate that Korea has the strongest co-movement relation with Japan. The strength of China and Taiwan are better than Singapore. The Hong Kong market has the weakest co-movement relation with Japan.

Keywords: co-movement, depreciation of Yen, rank correlation, stock market

Procedia PDF Downloads 212
10899 Joint Physical Custody after Divorce and Child Well-Being

Authors: Katarzyna Kamińska

Abstract:

Joint physical custody means that both parents after divorce or separation have the right and responsibility to take care of the child on the daily basis. In a joint physical custody arrangement, the child spends substantial, but not necessarily equal, time with both parents. Joint physical custody can be symmetric care arrangement or not. However, it is accepted in the jurisprudence that the best interests of the child is served when the child spends at least 35% of the time during a two-week period with each parent. Joint physical custody, also known as joint, dual, or shared residence, is a challenge in contemporary family law. It has its supporters and opponents. On the one hand, joint physical custody is beneficial because it provides children with frequent and continuous contact with a mother and father after their divorce or separation. On the other hand, it isn’t good for children to be shuttled back and forth between two residences. Children need a home base. The conclusion is therefore that joint physical custody can’t be seen as a panacea for all post-divorce or post-separation parenting cases and the court shouldn’t automatically make such a determination. The possibility to award this arrangement requires the court to carefully weigh the pros and cons of each individual case. It is difficult to say that joint physical custody is better than single physical custody in any case. It depends on the circumstances and needs of each family. It appears that an individual approach is going to be much better as opposed to a one-size-fits-all idea.

Keywords: joint physical custody, shared residence, dual residence, the best interests of the child

Procedia PDF Downloads 61
10898 Modeling Default Probabilities of the Chosen Czech Banks in the Time of the Financial Crisis

Authors: Petr Gurný

Abstract:

One of the most important tasks in the risk management is the correct determination of probability of default (PD) of particular financial subjects. In this paper a possibility of determination of financial institution’s PD according to the credit-scoring models is discussed. The paper is divided into the two parts. The first part is devoted to the estimation of the three different models (based on the linear discriminant analysis, logit regression and probit regression) from the sample of almost three hundred US commercial banks. Afterwards these models are compared and verified on the control sample with the view to choose the best one. The second part of the paper is aimed at the application of the chosen model on the portfolio of three key Czech banks to estimate their present financial stability. However, it is not less important to be able to estimate the evolution of PD in the future. For this reason, the second task in this paper is to estimate the probability distribution of the future PD for the Czech banks. So, there are sampled randomly the values of particular indicators and estimated the PDs’ distribution, while it’s assumed that the indicators are distributed according to the multidimensional subordinated Lévy model (Variance Gamma model and Normal Inverse Gaussian model, particularly). Although the obtained results show that all banks are relatively healthy, there is still high chance that “a financial crisis” will occur, at least in terms of probability. This is indicated by estimation of the various quantiles in the estimated distributions. Finally, it should be noted that the applicability of the estimated model (with respect to the used data) is limited to the recessionary phase of the financial market.

Keywords: credit-scoring models, multidimensional subordinated Lévy model, probability of default

Procedia PDF Downloads 429
10897 Competing Risk Analyses in Survival Trials During COVID-19 Pandemic

Authors: Ping Xu, Gregory T. Golm, Guanghan (Frank) Liu

Abstract:

In the presence of competing events, traditional survival analysis may not be appropriate and can result in biased estimates, as it assumes independence between competing events and the event of interest. Instead, competing risk analysis should be considered to correctly estimate the survival probability of the event of interest and the hazard ratio between treatment groups. The COVID-19 pandemic has provided a potential source of competing risks in clinical trials, as participants in trials may experienceCOVID-related competing events before the occurrence of the event of interest, for instance, death due to COVID-19, which can affect the incidence rate of the event of interest. We have performed simulation studies to compare multiple competing risk analysis models, including the cumulative incidence function, the sub-distribution hazard function, and the cause-specific hazard function, to the traditional survival analysis model under various scenarios. We also provide a general recommendation on conducting competing risk analysis in randomized clinical trials during the era of the COVID-19 pandemic based on the extensive simulation results.

Keywords: competing risk, survival analysis, simulations, randomized clinical trial, COVID-19 pandemic

Procedia PDF Downloads 164
10896 Studying the Load Sharing and Failure Mechanism of Hybrid Composite Joints Using Experiment and Finite Element Modeling

Authors: Seyyed Mohammad Hasheminia, Heoung Jae Chun, Jong Chan Park, Hong Suk Chang

Abstract:

Composite joints have been getting attention recently due to their high specific mechanical strength to weight ratio that is crucial for structures such as aircrafts and automobiles. In this study on hybrid joints, quasi-static experiments and finite element analysis were performed to investigate the failure mechanism of hybrid composite joint with respect to the joint properties such as the adhesive material, clamping force, and joint geometry. The outcomes demonstrated that the stiffness of the adhesive is the most imperative design parameter. In this investigation, two adhesives with various stiffness values were utilized. Regarding the joints utilizing the adhesive with the lower stiffness modulus, it was observed that the load was exchanged promptly through the adhesive since it was shared more proficiently between the bolt and adhesive. This phenomenon permitted the hybrid joints with low-modulus adhesive to support more prominent loads before failure when contrasted with the joints that utilize the stiffer adhesive. In the next step, the stress share between the bond and bolt as a function of various design parameters was studied using a finite element model in which it was understood that the geometrical parameters such as joint overlap and width have a significant influence on the load sharing between the bolt and the adhesive.

Keywords: composite joints, composite materials, hybrid joints, single-lap joint

Procedia PDF Downloads 380
10895 Adaptive CFAR Analysis for Non-Gaussian Distribution

Authors: Bouchemha Amel, Chachoui Takieddine, H. Maalem

Abstract:

Automatic detection of targets in a modern communication system RADAR is based primarily on the concept of adaptive CFAR detector. To have an effective detection, we must minimize the influence of disturbances due to the clutter. The detection algorithm adapts the CFAR detection threshold which is proportional to the average power of the clutter, maintaining a constant probability of false alarm. In this article, we analyze the performance of two variants of adaptive algorithms CA-CFAR and OS-CFAR and we compare the thresholds of these detectors in the marine environment (no-Gaussian) with a Weibull distribution.

Keywords: CFAR, threshold, clutter, distribution, Weibull, detection

Procedia PDF Downloads 553
10894 The Nonlinear Research on Rotational Stiffness of Cuplock Joint

Authors: Liuyu Zhang, Di Mo, Qiang Yan, Min Liu

Abstract:

As the important equipment in the construction field, cuplock scaffold plays an important role in the construction process. As a scaffold connecting member, cuplock joint is of great importance. In order to explore the rotational stiffness nonlinear characteristics changing features of different structural forms of cuplock joint in different tightening torque condition under different conditions of load, ANSYS is used to establish four kinds of cuplock joint models with different forces to simulate the real force situation. By setting the different load conditions which means the cuplock is loaded at a certain distance from the cuplock joint in a certain direction until the cuplock is damaged and considering the gap between the cross bar joint and the vertical bar, the differences in the influence of the structural form and tightening torque on the rotation stiffness of the cuplock under different load conditions are compared. It is significantly important to improve the accuracy of calculating bearing capacity and stability of the cuplock steel pipe scaffold.

Keywords: cuplock joint, highway tunnel, non-linear characteristics, rotational stiffness, scaffold stability, theoretical analysis

Procedia PDF Downloads 97
10893 Stability Bound of Ruin Probability in a Reduced Two-Dimensional Risk Model

Authors: Zina Benouaret, Djamil Aissani

Abstract:

In this work, we introduce the qualitative and quantitative concept of the strong stability method in the risk process modeling two lines of business of the same insurance company or an insurance and re-insurance companies that divide between them both claims and premiums with a certain proportion. The approach proposed is based on the identification of the ruin probability associate to the model considered, with a stationary distribution of a Markov random process called a reversed process. Our objective, after clarifying the condition and the perturbation domain of parameters, is to obtain the stability inequality of the ruin probability which is applied to estimate the approximation error of a model with disturbance parameters by the considered model. In the stability bound obtained, all constants are explicitly written.

Keywords: Markov chain, risk models, ruin probabilities, strong stability analysis

Procedia PDF Downloads 227
10892 A Stochastic Approach to Extreme Wind Speeds Conditions on a Small Axial Wind Turbine

Authors: Nkongho Ayuketang Arreyndip, Ebobenow Joseph

Abstract:

In this paper, to model a real life wind turbine, a probabilistic approach is proposed to model the dynamics of the blade elements of a small axial wind turbine under extreme stochastic wind speeds conditions. It was found that the power and the torque probability density functions even though decreases at these extreme wind speeds but are not infinite. Moreover, we also found that it is possible to stabilize the power coefficient (stabilizing the output power) above rated wind speeds by turning some control parameters. This method helps to explain the effect of turbulence on the quality and quantity of the harness power and aerodynamic torque.

Keywords: probability, probability density function, stochastic, turbulence

Procedia PDF Downloads 551
10891 Reliability Analysis of Construction Schedule Plan Based on Building Information Modelling

Authors: Lu Ren, You-Liang Fang, Yan-Gang Zhao

Abstract:

In recent years, the application of BIM (Building Information Modelling) to construction schedule plan has been the focus of more and more researchers. In order to assess the reasonable level of the BIM-based construction schedule plan, that is whether the schedule can be completed on time, some researchers have introduced reliability theory to evaluate. In the process of evaluation, the uncertain factors affecting the construction schedule plan are regarded as random variables, and probability distributions of the random variables are assumed to be normal distribution, which is determined using two parameters evaluated from the mean and standard deviation of statistical data. However, in practical engineering, most of the uncertain influence factors are not normal random variables. So the evaluation results of the construction schedule plan will be unreasonable under the assumption that probability distributions of random variables submitted to the normal distribution. Therefore, in order to get a more reasonable evaluation result, it is necessary to describe the distribution of random variables more comprehensively. For this purpose, cubic normal distribution is introduced in this paper to describe the distribution of arbitrary random variables, which is determined by the first four moments (mean, standard deviation, skewness and kurtosis). In this paper, building the BIM model firstly according to the design messages of the structure and making the construction schedule plan based on BIM, then the cubic normal distribution is used to describe the distribution of the random variables due to the collecting statistical data of the random factors influencing construction schedule plan. Next the reliability analysis of the construction schedule plan based on BIM can be carried out more reasonably. Finally, the more accurate evaluation results can be given providing reference for the implementation of the actual construction schedule plan. In the last part of this paper, the more efficiency and accuracy of the proposed methodology for the reliability analysis of the construction schedule plan based on BIM are conducted through practical engineering case.

Keywords: BIM, construction schedule plan, cubic normal distribution, reliability analysis

Procedia PDF Downloads 112
10890 Application Reliability Method for Concrete Dams

Authors: Mustapha Kamel Mihoubi, Mohamed Essadik Kerkar

Abstract:

Probabilistic risk analysis models are used to provide a better understanding of the reliability and structural failure of works, including when calculating the stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of reliability analysis methods including the methods used in engineering. It is in our case, the use of level 2 methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type first order risk method (FORM) and the second order risk method (SORM). By way of comparison, a level three method was used which generates a full analysis of the problem and involves an integration of the probability density function of random variables extended to the field of security using the Monte Carlo simulation method. Taking into account the change in stress following load combinations: normal, exceptional and extreme acting on the dam, calculation of the results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities, thus causing a significant decrease in strength, shear forces then induce a shift that threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case the increase of uplift in a hypothetical default of the drainage system.

Keywords: dam, failure, limit-state, monte-carlo, reliability, probability, simulation, sliding, taylor

Procedia PDF Downloads 302
10889 The Problem of Legal Regulation of Joint Physical Custody: The Polish Perspective

Authors: Katarzyna Kamińska

Abstract:

The main purpose of the work is to present the results of the studies regarding joint physical custody in the Polish legal system. The issues addressed fit into the ongoing process of modernising family law regulations and their adaptation to changing social reality in Poland. The Polish legislator now faces a dilemma: whether to introduce into Polish law a developed substantive or procedural regulation of joint physical custody and then whether it should be considered a legal presumption. Joint physical custody after divorce or separation is theoretically possible in Poland. It can either follow from the court’s independent proposal based on the assessment of the circumstances or from the parenting plan submitted by parents wishing to jointly retain full parental authority. However, joint physical custody does not result directly from the Polish Family and Guardianship Code. Therefore, there is real legal uncertainty in this matter, which leads to different treatment of citizens by the public authorities and courts. Another problem is that joint physical custody is misunderstood by the Polish courts. The main thesis of the work is that joint physical custody does not only mean the system of symmetrical child care (50/50), and the possibility to award joint physical custody will require the courts to carefully weigh the pros and cons of such an arrangement in each individual case.

Keywords: joint physical custody, shared parenting, divorce, separation, parental authority

Procedia PDF Downloads 46
10888 Applied Complement of Probability and Information Entropy for Prediction in Student Learning

Authors: Kennedy Efosa Ehimwenma, Sujatha Krishnamoorthy, Safiya Al‑Sharji

Abstract:

The probability computation of events is in the interval of [0, 1], which are values that are determined by the number of outcomes of events in a sample space S. The probability Pr(A) that an event A will never occur is 0. The probability Pr(B) that event B will certainly occur is 1. This makes both events A and B a certainty. Furthermore, the sum of probabilities Pr(E₁) + Pr(E₂) + … + Pr(Eₙ) of a finite set of events in a given sample space S equals 1. Conversely, the difference of the sum of two probabilities that will certainly occur is 0. This paper first discusses Bayes, the complement of probability, and the difference of probability for occurrences of learning-events before applying them in the prediction of learning objects in student learning. Given the sum of 1; to make a recommendation for student learning, this paper proposes that the difference of argMaxPr(S) and the probability of student-performance quantifies the weight of learning objects for students. Using a dataset of skill-set, the computational procedure demonstrates i) the probability of skill-set events that have occurred that would lead to higher-level learning; ii) the probability of the events that have not occurred that requires subject-matter relearning; iii) accuracy of the decision tree in the prediction of student performance into class labels and iv) information entropy about skill-set data and its implication on student cognitive performance and recommendation of learning.

Keywords: complement of probability, Bayes’ rule, prediction, pre-assessments, computational education, information theory

Procedia PDF Downloads 134
10887 First Order Moment Bounds on DMRL and IMRL Classes of Life Distributions

Authors: Debasis Sengupta, Sudipta Das

Abstract:

The class of life distributions with decreasing mean residual life (DMRL) is well known in the field of reliability modeling. It contains the IFR class of distributions and is contained in the NBUE class of distributions. While upper and lower bounds of the reliability distribution function of aging classes such as IFR, IFRA, NBU, NBUE, and HNBUE have discussed in the literature for a long time, there is no analogous result available for the DMRL class. We obtain the upper and lower bounds for the reliability function of the DMRL class in terms of first order finite moment. The lower bound is obtained by showing that for any fixed time, the minimization of the reliability function over the class of all DMRL distributions with a fixed mean is equivalent to its minimization over a smaller class of distribution with a special form. Optimization over this restricted set can be made algebraically. Likewise, the maximization of the reliability function over the class of all DMRL distributions with a fixed mean turns out to be a parametric optimization problem over the class of DMRL distributions of a special form. The constructive proofs also establish that both the upper and lower bounds are sharp. Further, the DMRL upper bound coincides with the HNBUE upper bound and the lower bound coincides with the IFR lower bound. We also prove that a pair of sharp upper and lower bounds for the reliability function when the distribution is increasing mean residual life (IMRL) with a fixed mean. This result is proved in a similar way. These inequalities fill a long-standing void in the literature of the life distribution modeling.

Keywords: DMRL, IMRL, reliability bounds, hazard functions

Procedia PDF Downloads 372
10886 Modelling Hydrological Time Series Using Wakeby Distribution

Authors: Ilaria Lucrezia Amerise

Abstract:

The statistical modelling of precipitation data for a given portion of territory is fundamental for the monitoring of climatic conditions and for Hydrogeological Management Plans (HMP). This modelling is rendered particularly complex by the changes taking place in the frequency and intensity of precipitation, presumably to be attributed to the global climate change. This paper applies the Wakeby distribution (with 5 parameters) as a theoretical reference model. The number and the quality of the parameters indicate that this distribution may be the appropriate choice for the interpolations of the hydrological variables and, moreover, the Wakeby is particularly suitable for describing phenomena producing heavy tails. The proposed estimation methods for determining the value of the Wakeby parameters are the same as those used for density functions with heavy tails. The commonly used procedure is the classic method of moments weighed with probabilities (probability weighted moments, PWM) although this has often shown difficulty of convergence, or rather, convergence to a configuration of inappropriate parameters. In this paper, we analyze the problem of the likelihood estimation of a random variable expressed through its quantile function. The method of maximum likelihood, in this case, is more demanding than in the situations of more usual estimation. The reasons for this lie, in the sampling and asymptotic properties of the estimators of maximum likelihood which improve the estimates obtained with indications of their variability and, therefore, their accuracy and reliability. These features are highly appreciated in contexts where poor decisions, attributable to an inefficient or incomplete information base, can cause serious damages.

Keywords: generalized extreme values, likelihood estimation, precipitation data, Wakeby distribution

Procedia PDF Downloads 114
10885 VaR or TCE: Explaining the Preferences of Regulators

Authors: Silvia Faroni, Olivier Le Courtois, Krzysztof Ostaszewski

Abstract:

While a lot of research concentrates on the merits of VaR and TCE, which are the two most classic risk indicators used by financial institutions, little has been written on explaining why regulators favor the choice of VaR or TCE in their set of rules. In this paper, we investigate the preferences of regulators with the aim of understanding why, for instance, a VaR with a given confidence level is ultimately retained. Further, this paper provides equivalence rules that explain how a given choice of VaR can be equivalent to a given choice of TCE. Then, we introduce a new risk indicator that extends TCE by providing a more versatile weighting of the constituents of probability distribution tails. All of our results are illustrated using the generalized Pareto distribution.

Keywords: generalized pareto distribution, generalized tail conditional expectation, regulator preferences, risk measure

Procedia PDF Downloads 138
10884 Localisation of Fluorescently Labelled Drug-Free Phospholipid Vesicles to the Cartilage Surface of Rat Synovial Joints

Authors: Sam Yurdakul, Nick Baverstock, Jim Mills

Abstract:

TDT 064 (FLEXISEQ®) is a drug-free gel used to treat osteoarthritis (OA)-associated pain and joint stiffness. It contains ultra-deformable phospholipid Sequessome™ vesicles, which can pass through the skin barrier intact. In six randomized OA studies, topical TDT 064 was well tolerated and improved joint pain, physical function and stiffness. In the largest study, these TDT 064-mediated effects were statistically significantly greater than oral placebo and equivalent to celecoxib. To understand the therapeutic effects of TDT 064, we investigated the localisation of the drug-free vesicles within rat synovial joints. TDT 064 containing DiO-labelled Sequessome™ vesicles was applied to the knees of four 6-week-old CD® hairless rats (10 mg/kg/ joint), 2–3 times/day, for 3 days (representing the recommended clinical dose). Eighteen hours later, the animals and one untreated control were sacrificed, and the knee joints isolated, flash frozen and embedded in Acrytol Mounting Media™. Approximately 15 sections (10 µm) from each joint were analysed by fluorescence microscopy. To investigate whether the localisation of DiO fluorescence was associated with intact vesicles, an anti-PEG monoclonal antibody (mAb) was used to detect Tween, a constituent of Sequessome™ vesicles. Sections were visualized at 484 nm (DiO) and 647 nm (anti-PEG mAb) and analysed using inForm 1.4 (Perkin Elmer, Inc.). Significant fluorescence was observed at 484 nm in sections from TDT 064-treated animals. No non-specific fluorescence was observed in control sections. Fluorescence was detected as discrete vesicles on the cartilage surfaces, inside the cartilaginous matrix and within the synovial space. The number of DiO-labelled vesicles in multiple fields of view was consistent and >100 in sections from four different treated knees. DiO and anti-PEG mAb co-localised within the collagenous tissues in four different joint sections. Under higher magnification (40x), vesicles were seen in the intercellular spaces of the synovial joint tissue, but no fluorescence was seen inside cells. These data suggest that the phospholipid vesicles in TDT 064 localize at the surface of the joint cartilage; these vesicles may therefore be supplementing the phospholipid deficiency reported in OA and acting as a biolubricant within the synovial joint.

Keywords: joint pain, osteoarthritis, phospholipid vesicles, TDT 064

Procedia PDF Downloads 422
10883 Study of Human Upper Arm Girth during Elbow Isokinetic Contractions Based on a Smart Circumferential Measuring System

Authors: Xi Wang, Xiaoming Tao, Raymond C. H. So

Abstract:

As one of the convenient and noninvasive sensing approaches, the automatic limb girth measurement has been applied to detect intention behind human motion from muscle deformation. The sensing validity has been elaborated by preliminary researches but still need more fundamental study, especially on kinetic contraction modes. Based on the novel fabric strain sensors, a soft and smart limb girth measurement system was developed by the authors’ group, which can measure the limb girth in-motion. Experiments were carried out on elbow isometric flexion and elbow isokinetic flexion (biceps’ isokinetic contractions) of 90°/s, 60°/s, and 120°/s for 10 subjects (2 canoeists and 8 ordinary people). After removal of natural circumferential increments due to elbow position, the joint torque is found not uniformly sensitive to the limb circumferential strains, but declining as elbow joint angle rises, regardless of the angular speed. Moreover, the maximum joint torque was found as an exponential function of the joint’s angular speed. This research highly contributes to the application of the automatic limb girth measuring during kinetic contractions, and it is useful to predict the contraction level of voluntary skeletal muscles.

Keywords: fabric strain sensor, muscle deformation, isokinetic contraction, joint torque, limb girth strain

Procedia PDF Downloads 313