Search results for: background probability
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5484

Search results for: background probability

5304 Paternity Index Analysis on Disputed Paternity Cases at Sardjito Hospital Yogyakarta, Indonesia

Authors: Taufik Hidayat, Yudha Nurhantari, Bambang U. D. Rianto

Abstract:

Introduction: The examination of the Short Tandem Repeats (STR) locus on nuclear DNA is very useful in solving the paternity cases. The purpose of this study is to know the description of paternity cases and paternity index/probability of paternity analysis based on Indonesian allele frequency at Sardjito Hospital Yogyakarta. Method: This was an observational study with cross-sectional analytic method. Population and sample were all cases of disputed paternity from January 2011 to June 2015 that fulfill the inclusion and exclusion criteria and were examined at Forensic Medicine Unit of Sardjito Hospital, Medical Faculty of Gadjah Mada University. The paternity index was calculated with EasyDNA Program by Fung (2013). Analysis of the study was conducted by comparing the results through unpaired categorical test using Kolmogorov-Smirnov test. This study was designed with 95% confidence interval (CI) with α = 5% and significance level is p < 0,05. Results: From 42 disputed paternity cases we obtained trio paternity cases were 32 cases (76.2%) and duo without a mother was 10 cases (23.8%). The majority of the fathers' estimated ages were 21-30 years (33.3%) and the mother's age was 31-40 years (38.1%). The majority of the ages of children examined for paternity were under 12 months (47.6%). The majority of ethnic clients are Javanese. Conclusion of inclusion was 57.1%, and exclusion was 42.9%. The Kolmogorov-Smirnov test obtained p-value = 0.673. Conclusion: There is no significant difference between paternity index/probability of paternity based on Indonesian allele frequency between trio and duo of paternity.

Keywords: disputed paternity, paternity index, probability of paternity, short tandem

Procedia PDF Downloads 144
5303 Externalizing Behavior Problems Influencing Social Behavior in Early Adolescence

Authors: Zhidong Zhang, Zhi-Chao Zhang

Abstract:

This study focuses on early adolescent externalizing behavioral problems which specifically concentrate on rule breaking behavior and aggressive behavior using the instrument of Achenbach System of Empirically Based Assessment (ASEBA). The purpose was to analyze the relationships between the externalizing behavioral problems and relevant background variables such as sports activities, hobbies, chores and the number of close friends. The stratified sampling method was used to collect data from 1975 participants. The results indicated that several background variables as predictors could significantly predict rule breaking behavior and aggressive behavior. Further, a hierarchical modeling method was used to explore the causal relations among background variables, breaking behavior variables and aggressive behavior variables.

Keywords: aggressive behavior, breaking behavior, early adolescence, externalizing problem

Procedia PDF Downloads 469
5302 Investigation on Optical Performance of Operational Shutter Panels for Transparent Displays

Authors: Jaehong Kim, Sunhee Park, HongSeop Shin, Kyongho Lim, Suhyun Kwon, Don-Gyou Lee, Pureum Kim, Moojong Lim, JongSang Baek

Abstract:

Transparent displays with OLEDs are the most commonly produced forms of see-through displays on the market or in development. In order to block the visual interruption caused by the light coming from the background, the special panel is combined with transparent displays with OLEDs. There is, however, few studies optical performance of operational shutter panel for transparent displays until now. This paper, therefore, describes the optical performance of operational shutter panels. The novel evaluation method was developed by measuring the amount of light which can form a transmitted background image. The new proposed method could tell how recognize transmitted background images cannot be seen, and is consistent with viewer’s perception.

Keywords: transparent display, operational shutter panel, optical performance, OLEDs

Procedia PDF Downloads 415
5301 The Impact of Feuerstein Enhancement of Learning Potential to the Integration of Children from Socially Disadvantaged Backgrounds into Society

Authors: Michal Kozubík, Svetlana Síthová

Abstract:

Aim: Aim of this study is to introduce the method of instrumental enrichment to people who works in the helping professions, and show further possibilities of its realization with children from socially disadvantaged backgrounds into society. Methods: We focused on Feuerstein’s Instrumental Enrichment method, its theoretical grounds and practical implementation. We carried out questionnaires and directly observed children from the disadvantaged background in Partizánske district. Results: We outlined the issues of children from disadvantaged social environment and their opportunity of social integration using the method. The findings showed the utility of Feuerstein method. Conclusions: We conclude that Feuerstein methods are very suitable for children from socially disadvantaged background and importance of social workers and special educator co-operation.

Keywords: Feuerstein, inclusion, education, socially disadvantaged background

Procedia PDF Downloads 289
5300 Rupture Probability of Type of Coarse Aggregate on Fracture Surface of Concrete

Authors: B. Ramakrishna, S. Sivamurthy Reddy

Abstract:

The various types of aggregates such as granite, dolerite, Quartzite, dolomitic limestone, limestone and river gravel were used to produce the concrete with 28-day target compressive strength of 35, 60, and 80 Mpa. The compressive strength of concrete, as well as aggregates, was measured to study the effect of rupture probability of aggregate on the fracture surface of the concrete. Also, the petrographic studies were carried out to study the texture, type of minerals present and their relative proportions in various types of aggregates. The concrete of various grades produced with the same aggregate has shown a rise in RPCA with strength. However, the above relationship has ceased to exist in the concretes of the same grade, made of different types of aggregates. The carbonate aggregates namely Limestone and Dolomitic limestone have produced concrete with higher RPCA irrespective of the strength of concrete. The mode of origin, texture and mineralogical composition of aggregates have a significant impact on their pulse velocity and thereby the pulse velocity of concrete.

Keywords: RPCA, DL, G, LS, RG

Procedia PDF Downloads 260
5299 The Causal Relationships between Educational Environments and Rule-Breaking Behavior Issues in Early Adolescence

Authors: Zhidong Zhang, Zhi-Chao Zhang

Abstract:

This study focused on early adolescent rule-breaking behavioral problems using the instrument of Achenbach System of Empirically Based Assessment (ASEBA). The purpose was to analyze the relationships between the rule-breaking behavioral problems and relevant background variables such as sports activities, hobbies, chores and the number of close friends. The stratified sampling method was used to collect data from 2532 participants. The results indicated that several background variables as predictors could significantly predict rule breaking behavior and aggressive behavior. Further, a path analysis method was used to explore the correlational and causal relationships among background variables and breaking behavior variables.

Keywords: ASEBA, rule-breaking, path analysis, early adolescent

Procedia PDF Downloads 338
5298 The Falling Point of Lubricant

Authors: Arafat Husain

Abstract:

The lubricants are one of the most used resource in today’s world. Lot of the superpowers are dependent on the lubricant resource for their country to function. To see that the lubricants are not adulterated we need to develop some efficient ways and to see which fluid has been added to the lubricant. So to observe the these malpractices in the lubricant we need to develop a method. We take a elastic ball and through it at probability circle in the submerged in the lubricant at a fixed force and see the distance of pitching and the point of fall. Then we the ratio of distance of falling to the distance of pitching and if the measured ratio is greater than one the fluid is less viscous and if the ratio is lesser than the lubricant is viscous. We will check the falling point of pure lubricant at fixed force and every pure lubricant would have a fixed falling point. After that we would adulterate the lubricant and note the falling point and if the falling point is less than the standard value then adulterate is solid and if the adulterate is liquid the falling point will be more than the standard value. Hence the comparison with the standard falling point will give the efficiency of the lubricant.

Keywords: falling point of lubricant, falling point ratios, probability circle, octane number

Procedia PDF Downloads 463
5297 Probability-Based Damage Detection of Structures Using Model Updating with Enhanced Ideal Gas Molecular Movement Algorithm

Authors: M. R. Ghasemi, R. Ghiasi, H. Varaee

Abstract:

Model updating method has received increasing attention in damage detection structures based on measured modal parameters. Therefore, a probability-based damage detection (PBDD) procedure based on a model updating procedure is presented in this paper, in which a one-stage model-based damage identification technique based on the dynamic features of a structure is investigated. The presented framework uses a finite element updating method with a Monte Carlo simulation that considers the uncertainty caused by measurement noise. Enhanced ideal gas molecular movement (EIGMM) is used as the main algorithm for model updating. Ideal gas molecular movement (IGMM) is a multiagent algorithm based on the ideal gas molecular movement. Ideal gas molecules disperse rapidly in different directions and cover all the space inside. This is embedded in the high speed of molecules, collisions between them and with the surrounding barriers. In IGMM algorithm to accomplish the optimal solutions, the initial population of gas molecules is randomly generated and the governing equations related to the velocity of gas molecules and collisions between those are utilized. In this paper, an enhanced version of IGMM, which removes unchanged variables after specified iterations, is developed. The proposed method is implemented on two numerical examples in the field of structural damage detection. The results show that the proposed method can perform well and competitive in PBDD of structures.

Keywords: enhanced ideal gas molecular movement (EIGMM), ideal gas molecular movement (IGMM), model updating method, probability-based damage detection (PBDD), uncertainty quantification

Procedia PDF Downloads 246
5296 An Adaptive CFAR Algorithm Based on Automatic Censoring in Heterogeneous Environments

Authors: Naime Boudemagh

Abstract:

In this work, we aim to improve the detection performances of radar systems. To this end, we propose and analyze a novel censoring technique of undesirable samples, of priori unknown positions, that may be present in the environment under investigation. Therefore, we consider heterogeneous backgrounds characterized by the presence of some irregularities such that clutter edge transitions and/or interfering targets. The proposed detector, termed automatic censoring constant false alarm (AC-CFAR), operates exclusively in a Gaussian background. It is built to allow the segmentation of the environment to regions and switch automatically to the appropriate detector; namely, the cell averaging CFAR (CA-CFAR), the censored mean level CFAR (CMLD-CFAR) or the order statistic CFAR (OS-CFAR). Monte Carlo simulations show that the AC-CFAR detector performs like the CA-CFAR in a homogeneous background. Moreover, the proposed processor exhibits considerable robustness in a heterogeneous background.

Keywords: CFAR, automatic censoring, heterogeneous environments, radar systems

Procedia PDF Downloads 572
5295 Distance Education: Using a Digital Platform to Improve Struggling University Students' Mathematical Skills

Authors: Robert Vanderburg, Nicholas Gibson

Abstract:

Objectives: There has been an increased focus in education students’ mathematics skills in the last two years. Universities have, specifically, had problems teaching students struggling with mathematics. This paper focuses on the ability of a digital platform to significantly improve mathematics skills for struggling students. Methods: 32 students who demonstrated low scores on a mathematics test were selected to take part in a one-month tutorial program using a digital mathematics portal. Students were provided feedback for questions posted on the portal and a fortnightly tutorial session. Results: A pre-test post-test design was analyzed using a one-way analysis of variance (ANOVA). The analysis suggested that students improved skills in algebra, geometry, statistics, probability, ratios, fractions, and probability. Conclusion: Distance university students can improve their mathematics skills using a digital platform.

Keywords: digital education, distance education, higher education, mathematics education

Procedia PDF Downloads 161
5294 Cooperative Spectrum Sensing Using Hybrid IWO/PSO Algorithm in Cognitive Radio Networks

Authors: Deepa Das, Susmita Das

Abstract:

Cognitive Radio (CR) is an emerging technology to combat the spectrum scarcity issues. This is achieved by consistently sensing the spectrum, and detecting the under-utilized frequency bands without causing undue interference to the primary user (PU). In soft decision fusion (SDF) based cooperative spectrum sensing, various evolutionary algorithms have been discussed, which optimize the weight coefficient vector for maximizing the detection performance. In this paper, we propose the hybrid invasive weed optimization and particle swarm optimization (IWO/PSO) algorithm as a fast and global optimization method, which improves the detection probability with a lesser sensing time. Then, the efficiency of this algorithm is compared with the standard invasive weed optimization (IWO), particle swarm optimization (PSO), genetic algorithm (GA) and other conventional SDF based methods on the basis of convergence and detection probability.

Keywords: cognitive radio, spectrum sensing, soft decision fusion, GA, PSO, IWO, hybrid IWO/PSO

Procedia PDF Downloads 426
5293 A Bayesian Approach for Analyzing Academic Article Structure

Authors: Jia-Lien Hsu, Chiung-Wen Chang

Abstract:

Research articles may follow a simple and succinct structure of organizational patterns, called move. For example, considering extended abstracts, we observe that an extended abstract usually consists of five moves, including Background, Aim, Method, Results, and Conclusion. As another example, when publishing articles in PubMed, authors are encouraged to provide a structured abstract, which is an abstract with distinct and labeled sections (e.g., Introduction, Methods, Results, Discussions) for rapid comprehension. This paper introduces a method for computational analysis of move structures (i.e., Background-Purpose-Method-Result-Conclusion) in abstracts and introductions of research documents, instead of manually time-consuming and labor-intensive analysis process. In our approach, sentences in a given abstract and introduction are automatically analyzed and labeled with a specific move (i.e., B-P-M-R-C in this paper) to reveal various rhetorical status. As a result, it is expected that the automatic analytical tool for move structures will facilitate non-native speakers or novice writers to be aware of appropriate move structures and internalize relevant knowledge to improve their writing. In this paper, we propose a Bayesian approach to determine move tags for research articles. The approach consists of two phases, training phase and testing phase. In the training phase, we build a Bayesian model based on a couple of given initial patterns and the corpus, a subset of CiteSeerX. In the beginning, the priori probability of Bayesian model solely relies on initial patterns. Subsequently, with respect to the corpus, we process each document one by one: extract features, determine tags, and update the Bayesian model iteratively. In the testing phase, we compare our results with tags which are manually assigned by the experts. In our experiments, the promising accuracy of the proposed approach reaches 56%.

Keywords: academic English writing, assisted writing, move tag analysis, Bayesian approach

Procedia PDF Downloads 304
5292 Performance Analysis of the Time-Based and Periodogram-Based Energy Detector for Spectrum Sensing

Authors: Sadaf Nawaz, Adnan Ahmed Khan, Asad Mahmood, Chaudhary Farrukh Javed

Abstract:

Classically, an energy detector is implemented in time domain (TD). However, frequency domain (FD) based energy detector has demonstrated an improved performance. This paper presents a comparison between the two approaches as to analyze their pros and cons. A detailed performance analysis of the classical TD energy-detector and the periodogram based detector is performed. Exact and approximate mathematical expressions for probability of false alarm (Pf) and probability of detection (Pd) are derived for both approaches. The derived expressions naturally lead to an analytical as well as intuitive reasoning for the improved performance of (Pf) and (Pd) in different scenarios. Our analysis suggests the dependence improvement on buffer sizes. Pf is improved in FD, whereas Pd is enhanced in TD based energy detectors. Finally, Monte Carlo simulations results demonstrate the analysis reached by the derived expressions.

Keywords: cognitive radio, energy detector, periodogram, spectrum sensing

Procedia PDF Downloads 346
5291 Preparedness is Overrated: Community Responses to Floods in a Context of (Perceived) Low Probability

Authors: Kim Anema, Matthias Max, Chris Zevenbergen

Abstract:

For any flood risk manager the 'safety paradox' has to be a familiar concept: low probability leads to a sense of safety, which leads to more investments in the area, which leads to higher potential consequences: keeping the aggregated risk (probability*consequences) at the same level. Therefore, it is important to mitigate potential consequences apart from probability. However, when the (perceived) probability is so low that there is no recognizable trend for society to adapt to, addressing the potential consequences will always be the lagging point on the agenda. Preparedness programs fail because of lack of interest and urgency, policy makers are distracted by their day to day business and there's always a more urgent issue to spend the taxpayer's money on. The leading question in this study was how to address the social consequences of flooding in a context of (perceived) low probability. Disruptions of everyday urban life, large or small, can be caused by a variety of (un)expected things - of which flooding is only one possibility. Variability like this is typically addressed with resilience - and we used the concept of Community Resilience as the framework for this study. Drawing on face to face interviews, an extensive questionnaire and publicly available statistical data we explored the 'whole society response' to two recent urban flood events; the Brisbane Floods (AUS) in 2011 and the Dresden Floods (GE) in 2013. In Brisbane, we studied how the societal impacts of the floods were counteracted by both authorities and the public, and in Dresden we were able to validate our findings. A large part of the reactions, both public as institutional, to these two urban flood events were not fuelled by preparedness or proper planning. Instead, more important success factors in counteracting social impacts like demographic changes in neighborhoods and (non-)economic losses were dynamics like community action, flexibility and creativity from authorities, leadership, informal connections and a shared narrative. These proved to be the determining factors for the quality and speed of recovery in both cities. The resilience of the community in Brisbane was good, due to (i) the approachability of (local) authorities, (ii) a big group of ‘secondary victims’ and (iii) clear leadership. All three of these elements were amplified by the use of social media and/ or web 2.0 by both the communities and the authorities involved. The numerous contacts and social connections made through the web were fast, need driven and, in their own way, orderly. Similarly in Dresden large groups of 'unprepared', ad hoc organized citizens managed to work together with authorities in a way that was effective and speeded up recovery. The concept of community resilience is better fitted than 'social adaptation' to deal with the potential consequences of an (im)probable flood. Community resilience is built on capacities and dynamics that are part of everyday life and which can be invested in pre-event to minimize the social impact of urban flooding. Investing in these might even have beneficial trade-offs in other policy fields.

Keywords: community resilience, disaster response, social consequences, preparedness

Procedia PDF Downloads 326
5290 Environmental Safety and Occupational Health Risk Assessment for Rocket Static Test

Authors: Phontip Kanlahasuth

Abstract:

This paper presents the environmental safety and occupational health risk assessment of rocket static test by assessing risk level from probability and severity and then appropriately applying the risk control measures. Before the environmental safety and occupational health measures are applied, the serious hazards level is 31%, medium level is 24% and low level is 45%. Once risk control measures are practically implemented, the serious hazard level can be diminished, medium level is 38%, low level is 45% and eliminated level is 17%. It is clearly shown that the environmental safety and occupational health measures can significantly reduce the risk level.

Keywords: rocket static test, hazard, risk, risk assessment, risk analysis, environment, safety, occupational health, acceptable risk, probability, severity, risk level

Procedia PDF Downloads 553
5289 A Stochastic Diffusion Process Based on the Two-Parameters Weibull Density Function

Authors: Meriem Bahij, Ahmed Nafidi, Boujemâa Achchab, Sílvio M. A. Gama, José A. O. Matos

Abstract:

Stochastic modeling concerns the use of probability to model real-world situations in which uncertainty is present. Therefore, the purpose of stochastic modeling is to estimate the probability of outcomes within a forecast, i.e. to be able to predict what conditions or decisions might happen under different situations. In the present study, we present a model of a stochastic diffusion process based on the bi-Weibull distribution function (its trend is proportional to the bi-Weibull probability density function). In general, the Weibull distribution has the ability to assume the characteristics of many different types of distributions. This has made it very popular among engineers and quality practitioners, who have considered it the most commonly used distribution for studying problems such as modeling reliability data, accelerated life testing, and maintainability modeling and analysis. In this work, we start by obtaining the probabilistic characteristics of this model, as the explicit expression of the process, its trends, and its distribution by transforming the diffusion process in a Wiener process as shown in the Ricciaardi theorem. Then, we develop the statistical inference of this model using the maximum likelihood methodology. Finally, we analyse with simulated data the computational problems associated with the parameters, an issue of great importance in its application to real data with the use of the convergence analysis methods. Overall, the use of a stochastic model reflects only a pragmatic decision on the part of the modeler. According to the data that is available and the universe of models known to the modeler, this model represents the best currently available description of the phenomenon under consideration.

Keywords: diffusion process, discrete sampling, likelihood estimation method, simulation, stochastic diffusion process, trends functions, bi-parameters weibull density function

Procedia PDF Downloads 265
5288 Cellular Automata Model for Car Accidents at a Signalized Intersection

Authors: Rachid Marzoug, Noureddine Lakouari, Beatriz Castillo Téllez, Margarita Castillo Téllez, Gerardo Alberto Mejía Pérez

Abstract:

This paper developed a two-lane cellular automata model to explain the relationship between car accidents at a signalized intersection and traffic-related parameters. It is found that the increase of the lane-changing probability P?ₕ? increases the risk of accidents, besides, the inflow α and the probability of accidents Pₐ? exhibit a nonlinear relationship. Furthermore, depending on the inflow, Pₐ? exhibits three different phases. The transition from phase I to phase II is of first (second) order when P?ₕ?=0 (P?ₕ?>0). However, the system exhibits a second (first) order transition from phase II to phase III when P?ₕ?=0 (P?ₕ?>0). In addition, when the inflow is not very high, the green light length of one road should be increased to improve road safety. Finally, simulation results show that the traffic at the intersection is safer adopting symmetric lane-changing rules than asymmetric ones.

Keywords: two-lane intersection, accidents, fatality risk, lane-changing, phase transition

Procedia PDF Downloads 180
5287 Evaluating the Probability of Foreign Tourists' Return to the City of Mashhad, Iran

Authors: Mohammad Rahim Rahnama, Amir Ali Kharazmi, Safiye Rokni

Abstract:

The tourism industry will be the most important unlimited, sustainable source of income after the oil and automotive industries by 2020 and not only countries, but cities are striving to apprehend its various facets. In line with this objective, the present descriptive-analytical study, through survey and using a questionnaire, seeks to evaluate the probability of tourists’ return and their recommendation to their countrymen to travel to Mashhad, Iran. The population under study is a sample of 384 foreign tourists who, in 2016, arrived at Mashhad, the second metropolis in Iran and its biggest religious city. The Kaplan-Meier estimator was used to analyze the data. Twenty-six percent of the tourists are female and 74% are male. On average, each tourist has had 3.02 trips abroad and 2.1 trips to Mashhad. Tourists from 14 different countries have arrived at Mashhad. Kuwait (15.9%), Armenia (15.6%), and Iraq (10.9%) were the countries where most tourists originated. Seventy-six percent of the tourists traveled with family and 90% of the tourists arrived at Mashhad via airplane. Major purposes of tourists’ trip include pilgrimage (27.9%), treatment (22.1%) followed by pilgrimage and treatment combined (35.4%). Major issues for tourists, in the order of priority, include quality of goods and services (30.2%), shopping (18%), and inhabitants’ treatment of foreigners (15.9%). Main tourist attractions, in addition to the Holy Shrine of Imam Reza, include Torqabeh and Shandiz (Torqabeh 40.9% and Shandiz 29.9%), Neyshabour (18.2%) followed by Kalat, 4.4%. The average willingness to return among tourists is 3.13, which is higher than the mean 3, indicating satisfaction with the stay in Mashhad. Similarly, the average for tourists’ recommending to their countrymen to visit Mashhad is 3.42, which is also an indicator of tourists’ satisfaction with their presence in Mashhad. According to the findings of the Kaplan-Meier estimator, an increase in the number of tourists’ trips to Mashhad, and an increase in the number of tourists’ foreign trips, reduces the probability of recommending a trip to Mashhad by tourists. Similarly, willingness to return is higher among those who stayed at a relatives’ home compared with other patterns of residence (hotels, self-catering accommodation, and pilgrim houses). Therefore, addressing the issues raised by tourists is essential for their return and their recommendation to others to travel to Mashhad.

Keywords: international tourist, probability of return, satisfaction, Mashhad

Procedia PDF Downloads 134
5286 Max-Entropy Feed-Forward Clustering Neural Network

Authors: Xiaohan Bookman, Xiaoyan Zhu

Abstract:

The outputs of non-linear feed-forward neural network are positive, which could be treated as probability when they are normalized to one. If we take Entropy-Based Principle into consideration, the outputs for each sample could be represented as the distribution of this sample for different clusters. Entropy-Based Principle is the principle with which we could estimate the unknown distribution under some limited conditions. As this paper defines two processes in Feed-Forward Neural Network, our limited condition is the abstracted features of samples which are worked out in the abstraction process. And the final outputs are the probability distribution for different clusters in the clustering process. As Entropy-Based Principle is considered into the feed-forward neural network, a clustering method is born. We have conducted some experiments on six open UCI data sets, comparing with a few baselines and applied purity as the measurement. The results illustrate that our method outperforms all the other baselines that are most popular clustering methods.

Keywords: feed-forward neural network, clustering, max-entropy principle, probabilistic models

Procedia PDF Downloads 406
5285 Replacing MOSFETs with Single Electron Transistors (SET) to Reduce Power Consumption of an Inverter Circuit

Authors: Ahmed Shariful Alam, Abu Hena M. Mustafa Kamal, M. Abdul Rahman, M. Nasmus Sakib Khan Shabbir, Atiqul Islam

Abstract:

According to the rules of quantum mechanics there is a non-vanishing probability of for an electron to tunnel through a thin insulating barrier or a thin capacitor which is not possible according to the laws of classical physics. Tunneling of electron through a thin insulating barrier or tunnel junction is a random event and the magnitude of current flowing due to the tunneling of electron is very low. As the current flowing through a Single Electron Transistor (SET) is the result of electron tunneling through tunnel junctions of its source and drain the supply voltage requirement is also very low. As a result, the power consumption across a Single Electron Transistor is ultra-low in comparison to that of a MOSFET. In this paper simulations have been done with PSPICE for an inverter built with both SETs and MOSFETs. 35mV supply voltage was used for a SET built inverter circuit and the supply voltage used for a CMOS inverter was 3.5V.

Keywords: ITRS, enhancement type MOSFET, island, DC analysis, transient analysis, power consumption, background charge co-tunneling

Procedia PDF Downloads 496
5284 S-N-Pf Relationship for Steel Fibre Reinforced Concrete Made with Cement Additives

Authors: Gurbir Kaur, Surinder Pal Singh

Abstract:

The present study is a part of the research work on the effect of limestone powder (LP), silica fume (SF) and metakaolin (MK), on the flexural fatigue performance of steel fibre reinforced concrete (SFRC). Corrugated rectangular steel fibres of size 0.6x2.0x35 mm at a constant volume fraction of 1.0% have been incorporated in all mix combinations as the reinforcing material. Three mix combinations were prepared by replacing 30% of ordinary Portland cement (OPC) by weight with these cement additives in binary and ternary fashion to demonstrate their contribution. An experimental programme was conducted to obtain the fatigue lives of all mix combinations at various stress levels. The fatigue life data have been analysed as an attempt to determine the relationship between stress level ‘S’, number of cycles to failure ‘N’ and probability of failure ‘Pf’ for all mix combinations. The experimental coefficients of the fatigue equation have also been obtained from the fatigue data to represent the S-N-Pf curves analytically.

Keywords: cement additives, fatigue life, probability of failure, steel fibre reinforced concrete

Procedia PDF Downloads 392
5283 Diagnostic Yield of CT PA and Value of Pre Test Assessments in Predicting the Probability of Pulmonary Embolism

Authors: Shanza Akram, Sameen Toor, Heba Harb Abu Alkass, Zainab Abdulsalam Altaha, Sara Taha Abdulla, Saleem Imran

Abstract:

Acute pulmonary embolism (PE) is a common disease and can be fatal. The clinical presentation is variable and nonspecific, making accurate diagnosis difficult. Testing patients with suspected acute PE has increased dramatically. However, the overuse of some tests, particularly CT and D-dimer measurement, may not improve care while potentially leading to patient harm and unnecessary expense. CTPA is the investigation of choice for PE. Its easy availability, accuracy and ability to provide alternative diagnosis has lowered the threshold for performing it, resulting in its overuse. Guidelines have recommended the use of clinical pretest probability tools such as ‘Wells score’ to assess risk of suspected PE. Unfortunately, implementation of guidelines in clinical practice is inconsistent. This has led to low risk patients being subjected to unnecessary imaging, exposure to radiation and possible contrast related complications. Aim: To study the diagnostic yield of CT PA, clinical pretest probability of patients according to wells score and to determine whether or not there was an overuse of CTPA in our service. Methods: CT scans done on patients with suspected P.E in our hospital from 1st January 2014 to 31st December 2014 were retrospectively reviewed. Medical records were reviewed to study demographics, clinical presentation, final diagnosis, and to establish if Wells score and D-Dimer were used correctly in predicting the probability of PE and the need for subsequent CTPA. Results: 100 patients (51male) underwent CT PA in the time period. Mean age was 57 years (24-91 years). Majority of patients presented with shortness of breath (52%). Other presenting symptoms included chest pain 34%, palpitations 6%, collapse 5% and haemoptysis 5%. D Dimer test was done in 69%. Overall Wells score was low (<2) in 28 %, moderate (>2 - < 6) in 47% and high (> 6) in 15% of patients. Wells score was documented in medical notes of only 20% patients. PE was confirmed in 12% (8 male) patients. 4 had bilateral PE’s. In high-risk group (Wells > 6) (n=15), there were 5 diagnosed PEs. In moderate risk group (Wells >2 - < 6) (n=47), there were 6 and in low risk group (Wells <2) (n=28), one case of PE was confirmed. CT scans negative for PE showed pleural effusion in 30, Consolidation in 20, atelactasis in 15 and pulmonary nodule in 4 patients. 31 scans were completely normal. Conclusion: Yield of CT for pulmonary embolism was low in our cohort at 12%. A significant number of our patients who underwent CT PA had low Wells score. This suggests that CT PA is over utilized in our institution. Wells score was poorly documented in medical notes. CT-PA was able to detect alternative pulmonary abnormalities explaining the patient's clinical presentation. CT-PA requires concomitant pretest clinical probability assessment to be an effective diagnostic tool for confirming or excluding PE. . Clinicians should use validated clinical prediction rules to estimate pretest probability in patients in whom acute PE is being considered. Combining Wells scores with clinical and laboratory assessment may reduce the need for CTPA.

Keywords: CT PA, D dimer, pulmonary embolism, wells score

Procedia PDF Downloads 190
5282 Probability Fuzzy Aggregation Operators in Vehicle Routing Problem

Authors: Anna Sikharulidze, Gia Sirbiladze

Abstract:

For the evaluation of unreliability levels of movement on the closed routes in the vehicle routing problem, the fuzzy operators family is constructed. The interactions between routing factors in extreme conditions on the roads are considered. A multi-criteria decision-making model (MCDM) is constructed. Constructed aggregations are based on the Choquet integral and the associated probability class of a fuzzy measure. Propositions on the correctness of the extension are proved. Connections between the operators and the compositions of dual triangular norms are described. The conjugate connections between the constructed operators are shown. Operators reflect interactions among all the combinations of the factors in the fuzzy MCDM process. Several variants of constructed operators are used in the decision-making problem regarding the assessment of unreliability and possibility levels of movement on closed routes.

Keywords: vehicle routing problem, associated probabilities of a fuzzy measure, choquet integral, fuzzy aggregation operator

Procedia PDF Downloads 284
5281 Downtime Modelling for the Post-Earthquake Building Assessment Phase

Authors: S. Khakurel, R. P. Dhakal, T. Z. Yeow

Abstract:

Downtime is one of the major sources (alongside damage and injury/death) of financial loss incurred by a structure in an earthquake. The length of downtime associated with a building after an earthquake varies depending on the time taken for the reaction (to the earthquake), decision (on the future course of action) and execution (of the decided course of action) phases. Post-earthquake assessment of buildings is a key step in the decision making process to decide the appropriate safety placarding as well as to decide whether a damaged building is to be repaired or demolished. The aim of the present study is to develop a model to quantify downtime associated with the post-earthquake building-assessment phase in terms of two parameters; i) duration of the different assessment phase; and ii) probability of different colour tagging. Post-earthquake assessment of buildings includes three stages; Level 1 Rapid Assessment including a fast external inspection shortly after the earthquake, Level 2 Rapid Assessment including a visit inside the building and Detailed Engineering Evaluation (if needed). In this study, the durations of all three assessment phases are first estimated from the total number of damaged buildings, total number of available engineers and the average time needed for assessing each building. Then, probability of different tag colours is computed from the 2010-11 Canterbury earthquake Sequence database. Finally, a downtime model for the post-earthquake building inspection phase is proposed based on the estimated phase length and probability of tag colours. This model is expected to be used for rapid estimation of seismic downtime within the Loss Optimisation Seismic Design (LOSD) framework.

Keywords: assessment, downtime, LOSD, Loss Optimisation Seismic Design, phase length, tag color

Procedia PDF Downloads 153
5280 Effects of the Affordable Care Act On Preventive Care Disparities

Authors: Cagdas Agirdas

Abstract:

Background: The Affordable Care Act (ACA) requires non-grandfathered private insurance plans, starting with plan years on or after September 23rd, 2010, to provide certain preventive care services without any cost sharing in the form of deductibles, copayments or co-insurance. This requirement may affect racial and ethnic disparities in preventive care as it provides the largest copay reduction in preventive care. Objectives: We ask whether the ACA’s free preventive care benefits are associated with a reduction in racial and ethnic disparities in the utilization of four preventive services: cholesterol screenings, colonoscopies, mammograms, and pap smears. Methods: We use a data set of over 6,000 individuals from the 2009, 2010, and 2013 Medical Expenditure Panel Surveys (MEPS). We restrict our data set only to individuals who are old enough to be eligible for each preventive service. Our difference-in-differences logistic regression model classifies privately-insured Hispanics, African Americans, and Asians as the treatment groups and 2013 as the after-policy year. Our control group consists of non-Hispanic whites on Medicaid as this program already covered preventive care services for free or at a low cost before the ACA. Results: After controlling for income, education, marital status, preferred interview language, self-reported health status, employment, having a usual source of care, age and gender, we find that the ACA is associated with increases in the probability of the median, privately-insured Hispanic person to get a colonoscopy by 3.6% and a mammogram by 3.1%, compared to a non-Hispanic white person on Medicaid. Similarly, we find that the median, privately-insured African American person’s probability of receiving these two preventive services improved by 2.3% and 2.4% compared to a non-Hispanic white person on Medicaid. We do not find any significant improvements for any racial or ethnic group for cholesterol screenings or pap smears. Furthermore, our results do not indicate any significant changes for Asians compared to non-Hispanic whites in utilizing the four preventive services. These reductions in racial/ethnic disparities are robust to reconfigurations of time periods, previous diagnosis, and residential status. Conclusions: Early effects of the ACA’s provision of free preventive care are significant for Hispanics and African Americans. Further research is needed for the later years as more individuals became aware of these benefits.

Keywords: preventive care, Affordable Care Act, cost sharing, racial disparities

Procedia PDF Downloads 117
5279 Trajectories of Conduct Problems and Cumulative Risk from Early Childhood to Adolescence

Authors: Leslie M. Gutman

Abstract:

Conduct problems (CP) represent a major dilemma, with wide-ranging and long-lasting individual and societal impacts. Children experience heterogeneous patterns of conduct problems; based on the age of onset, developmental course and related risk factors from around age 3. Early childhood represents a potential window for intervention efforts aimed at changing the trajectory of early starting conduct problems. Using the UK Millennium Cohort Study (n = 17,206 children), this study (a) identifies trajectories of conduct problems from ages 3 to 14 years and (b) assesses the cumulative and interactive effects of individual, family and socioeconomic risk factors from ages 9 months to 14 years. The same factors according to three domains were assessed, including child (i.e., low verbal ability, hyperactivity/inattention, peer problems, emotional problems), family (i.e., single families, parental poor physical and mental health, large family size) and socioeconomic (i.e., low family income, low parental education, unemployment, social housing). A cumulative risk score for the child, family, and socioeconomic domains at each age was calculated. It was then examined how the cumulative risk scores explain variation in the trajectories of conduct problems. Lastly, interactive effects among the different domains of cumulative risk were tested. Using group-based trajectory modeling, four distinct trajectories were found including a ‘low’ problem group and three groups showing childhood-onset conduct problems: ‘school-age onset’; ‘early-onset, desisting’; and ‘early-onset, persisting’. The ‘low’ group (57% of the sample) showed a low probability of conducts problems, close to zero, from 3 to 14 years. The ‘early-onset, desisting’ group (23% of the sample) demonstrated a moderate probability of CP in early childhood, with a decline from 3 to 5 years and a low probability thereafter. The ‘early-onset, persistent’ group (8%) followed a high probability of conduct problems, which declined from 11 years but was close to 70% at 14 years. In the ‘school-age onset’ group, 12% of the sample showed a moderate probability of conduct problems from 3 and 5 years, with a sharp increase by 7 years, increasing to 50% at 14 years. In terms of individual risk, all factors increased the likelihood of being in the childhood-onset groups compared to the ‘low’ group. For cumulative risk, the socioeconomic domain at 9 months and 3 years, the family domain at all ages except 14 years and child domain at all ages were found to differentiate childhood-onset groups from the ‘low’ group. Cumulative risk at 9 months and 3 years did not differentiate between the ‘school-onset’ group and ‘low’ group. Significant interactions were found between the domains for the ‘early-onset, desisting group’ suggesting that low levels of risk in one domain may buffer the effects of high risk in another domain. The implications of these findings for preventive interventions will be highlighted.

Keywords: conduct problems, cumulative risk, developmental trajectories, early childhood, adolescence

Procedia PDF Downloads 225
5278 A Semi-Markov Chain-Based Model for the Prediction of Deterioration of Concrete Bridges in Quebec

Authors: Eslam Mohammed Abdelkader, Mohamed Marzouk, Tarek Zayed

Abstract:

Infrastructure systems are crucial to every aspect of life on Earth. Existing Infrastructure is subjected to degradation while the demands are growing for a better infrastructure system in response to the high standards of safety, health, population growth, and environmental protection. Bridges play a crucial role in urban transportation networks. Moreover, they are subjected to high level of deterioration because of the variable traffic loading, extreme weather conditions, cycles of freeze and thaw, etc. The development of Bridge Management Systems (BMSs) has become a fundamental imperative nowadays especially in the large transportation networks due to the huge variance between the need for maintenance actions, and the available funds to perform such actions. Deterioration models represent a very important aspect for the effective use of BMSs. This paper presents a probabilistic time-based model that is capable of predicting the condition ratings of the concrete bridge decks along its service life. The deterioration process of the concrete bridge decks is modeled using semi-Markov process. One of the main challenges of the Markov Chain Decision Process (MCDP) is the construction of the transition probability matrix. Yet, the proposed model overcomes this issue by modeling the sojourn times based on some probability density functions. The sojourn times of each condition state are fitted to probability density functions based on some goodness of fit tests such as Kolmogorov-Smirnov test, Anderson Darling, and chi-squared test. The parameters of the probability density functions are obtained using maximum likelihood estimation (MLE). The condition ratings obtained from the Ministry of Transportation in Quebec (MTQ) are utilized as a database to construct the deterioration model. Finally, a comparison is conducted between the Markov Chain and semi-Markov chain to select the most feasible prediction model.

Keywords: bridge management system, bridge decks, deterioration model, Semi-Markov chain, sojourn times, maximum likelihood estimation

Procedia PDF Downloads 180
5277 Tolerance of Ambiguity in Relation to Listening Performance across Learners of Various Linguistic Backgrounds

Authors: Amin Kaveh Boukani

Abstract:

Foreign language learning is not straightforward and can be affected by numerous factors, among which personality features like tolerance of ambiguity (TA) are so well-known and important. Such characteristics yet can be affected by other factors like learning additional languages. The current investigation, thus, opted to explore the possible effect of linguistic background (being bilingual or trilingual) on the tolerance of ambiguity (TA) of Iranian EFL learners. Furthermore, the possible mediating effect of TA on multilingual learners' language performance (listening comprehension in this study) was expounded. This research involved 68 EFL learners (32 bilinguals, 29 trilinguals) with the age range of 19-29 doing their degrees in the Department of English Language and Literature of Urmia University. A set of questionnaires, including tolerance of ambiguity (Herman et. al., 2010) and linguistic background information (Modirkhameneh, 2005), as well as the IELTS listening comprehension test, were used for data collection purposes. The results of a set of independent samples t-test and mediation analysis (Hayes, 2022) showed that (1) linguistic background (being bilingual or trilingual) had a significant direct effect on EFL learners' TA, (2) Linguistic background had a significant direct influence on listening comprehension, (3) TA had a substantial direct influence on listening comprehension, and (4) TA moderated the influence of linguistic background on listening comprehension considerably. These results suggest that multilingualism may be considered as an advantageous asset for EFL learners and should be a prioritized characteristic in EFL instruction in multilingual contexts. Further pedagogical implications and suggestions for research are proposed in light of effective EFL instruction in multilingual contexts.

Keywords: tolerance of ambiguity, listening comprehension, multilingualism, bilingual, trilingual

Procedia PDF Downloads 30
5276 Continuous Wave Interference Effects on Global Position System Signal Quality

Authors: Fang Ye, Han Yu, Yibing Li

Abstract:

Radio interference is one of the major concerns in using the global positioning system (GPS) for civilian and military applications. Interference signals are produced not only through all electronic systems but also illegal jammers. Among different types of interferences, continuous wave (CW) interference has strong adverse impacts on the quality of the received signal. In this paper, we make more detailed analysis for CW interference effects on GPS signal quality. Based on the C/A code spectrum lines, the influence of CW interference on the acquisition performance of GPS receivers is further analysed. This influence is supported by simulation results using GPS software receiver. As the most important user parameter of GPS receivers, the mathematical expression of bit error probability is also derived in the presence of CW interference, and the expression is consistent with the Monte Carlo simulation results. The research on CW interference provides some theoretical gist and new thoughts on monitoring the radio noise environment and improving the anti-jamming ability of GPS receivers.

Keywords: GPS, CW interference, acquisition performance, bit error probability, Monte Carlo

Procedia PDF Downloads 233
5275 Probabilistic and Stochastic Analysis of a Retaining Wall for C-Φ Soil Backfill

Authors: André Luís Brasil Cavalcante, Juan Felix Rodriguez Rebolledo, Lucas Parreira de Faria Borges

Abstract:

A methodology for the probabilistic analysis of active earth pressure on retaining wall for c-Φ soil backfill is described in this paper. The Rosenblueth point estimate method is used to measure the failure probability of a gravity retaining wall. The basic principle of this methodology is to use two point estimates, i.e., the standard deviation and the mean value, to examine a variable in the safety analysis. The simplicity of this framework assures to its wide application. For the calculation is required 2ⁿ repetitions during the analysis, since the system is governed by n variables. In this study, a probabilistic model based on the Rosenblueth approach for the computation of the overturning probability of failure of a retaining wall is presented. The obtained results have shown the advantages of this kind of models in comparison with the deterministic solution. In a relatively easy way, the uncertainty on the wall and fill parameters are taken into account, and some practical results can be obtained for the retaining structure design.

Keywords: retaining wall, active earth pressure, backfill, probabilistic analysis

Procedia PDF Downloads 379