Search results for: probability approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14156

Search results for: probability approach

14006 Markov-Chain-Based Optimal Filtering and Smoothing

Authors: Garry A. Einicke, Langford B. White

Abstract:

This paper describes an optimum filter and smoother for recovering a Markov process message from noisy measurements. The developments follow from an equivalence between a state space model and a hidden Markov chain. The ensuing filter and smoother employ transition probability matrices and approximate probability distribution vectors. The properties of the optimum solutions are retained, namely, the estimates are unbiased and minimize the variance of the output estimation error, provided that the assumed parameter set are correct. Methods for estimating unknown parameters from noisy measurements are discussed. Signal recovery examples are described in which performance benefits are demonstrated at an increased calculation cost.

Keywords: optimal filtering, smoothing, Markov chains

Procedia PDF Downloads 296
14005 Low Probability of Intercept (LPI) Signal Detection and Analysis Using Choi-Williams Distribution

Authors: V. S. S. Kumar, V. Ramya

Abstract:

In the modern electronic warfare, the signal scenario is changing at a rapid pace with the introduction of Low Probability of Intercept (LPI) radars. In the modern battlefield, radar system faces serious threats from passive intercept receivers such as Electronic Attack (EA) and Anti-Radiation Missiles (ARMs). To perform necessary target detection and tracking and simultaneously hide themselves from enemy attack, radar systems should be LPI. These LPI radars use a variety of complex signal modulation schemes together with pulse compression with the aid of advancement in signal processing capabilities of the radar such that the radar performs target detection and tracking while simultaneously hiding enemy from attack such as EA etc., thus posing a major challenge to the ES/ELINT receivers. Today an increasing number of LPI radars are being introduced into the modern platforms and weapon systems so these LPI radars created a requirement for the armed forces to develop new techniques, strategies and equipment to counter them. This paper presents various modulation techniques used in generation of LPI signals and development of Time Frequency Algorithms to analyse those signals.

Keywords: anti-radiation missiles, cross terms, electronic attack, electronic intelligence, electronic warfare, intercept receiver, low probability of intercept

Procedia PDF Downloads 409
14004 Computing Transition Intensity Using Time-Homogeneous Markov Jump Process: Case of South African HIV/AIDS Disposition

Authors: A. Bayaga

Abstract:

This research provides a technical account of estimating Transition Probability using Time-homogeneous Markov Jump Process applying by South African HIV/AIDS data from the Statistics South Africa. It employs Maximum Likelihood Estimator (MLE) model to explore the possible influence of Transition Probability of mortality cases in which case the data was based on actual Statistics South Africa. This was conducted via an integrated demographic and epidemiological model of South African HIV/AIDS epidemic. The model was fitted to age-specific HIV prevalence data and recorded death data using MLE model. Though the previous model results suggest HIV in South Africa has declined and AIDS mortality rates have declined since 2002 – 2013, in contrast, our results differ evidently with the generally accepted HIV models (Spectrum/EPP and ASSA2008) in South Africa. However, there is the need for supplementary research to be conducted to enhance the demographic parameters in the model and as well apply it to each of the nine (9) provinces of South Africa.

Keywords: AIDS mortality rates, epidemiological model, time-homogeneous markov jump process, transition probability, statistics South Africa

Procedia PDF Downloads 455
14003 Sizing Residential Solar Power Systems Based on Site-Specific Energy Statistics

Authors: Maria Arechavaleta, Mark Halpin

Abstract:

In the United States, costs of solar energy systems have declined to the point that they are viable options for most consumers. However, there are no consistent procedures for specifying sufficient systems. The factors that must be considered are energy consumption, potential solar energy production, and cost. The traditional method of specifying solar energy systems is based on assumed daily levels of available solar energy and average amounts of daily energy consumption. The mismatches between energy production and consumption are usually mitigated using battery energy storage systems, and energy use is curtailed when necessary. The main consumer decision question that drives the total system cost is how much unserved (or curtailed) energy is acceptable? Of course additional solar conversion equipment can be installed to provide greater peak energy production and extra energy storage capability can be added to mitigate longer lasting low solar energy production periods. Each option increases total cost and provides a benefit which is difficult to quantify accurately. An approach to quantify the cost-benefit of adding additional resources, either production or storage or both, based on the statistical concepts of loss-of-energy probability and expected unserved energy, is presented in this paper. Relatively simple calculations, based on site-specific energy availability and consumption data, can be used to show the value of each additional increment of production or storage. With this incremental benefit-cost information, consumers can select the best overall performance combination for their application at a cost they are comfortable paying. The approach is based on a statistical analysis of energy consumption and production characteristics over time. The characteristics are in the forms of curves with each point on the curve representing an energy consumption or production value over a period of time; a one-minute period is used for the work in this paper. These curves are measured at the consumer location under the conditions that exist at the site and the duration of the measurements is a minimum of one week. While greater accuracy could be obtained with longer recording periods, the examples in this paper are based on a single week for demonstration purposes. The weekly consumption and production curves are overlaid on each other and the mismatches are used to size the battery energy storage system. Loss-of-energy probability and expected unserved energy indices are calculated in addition to the total system cost. These indices allow the consumer to recognize and quantify the benefit (probably a reduction in energy consumption curtailment) available for a given increase in cost. Consumers can then make informed decisions that are accurate for their location and conditions and which are consistent with their available funds.

Keywords: battery energy storage systems, loss of load probability, residential renewable energy, solar energy systems

Procedia PDF Downloads 211
14002 Subjective Probability and the Intertemporal Dimension of Probability to Correct the Misrelation Between Risk and Return of a Financial Asset as Perceived by Investors. Extension of Prospect Theory to Better Describe Risk Aversion

Authors: Roberta Martino, Viviana Ventre

Abstract:

From a theoretical point of view, the relationship between the risk associated with an investment and the expected value are directly proportional, in the sense that the market allows a greater result to those who are willing to take a greater risk. However, empirical evidence proves that this relationship is distorted in the minds of investors and is perceived exactly the opposite. To deepen and understand the discrepancy between the actual actions of the investor and the theoretical predictions, this paper analyzes the essential parameters used for the valuation of financial assets with greater attention to two elements: probability and the passage of time. Although these may seem at first glance to be two distinct elements, they are closely related. In particular, the error in the theoretical description of the relationship between risk and return lies in the failure to consider the impatience that is generated in the decision-maker when events that have not yet happened occur in the decision-making context. In this context, probability loses its objective meaning and in relation to the psychological aspects of the investor, it can only be understood as the degree of confidence that the investor has in the occurrence or non-occurrence of an event. Moreover, the concept of objective probability does not consider the inter-temporality that characterizes financial activities and does not consider the condition of limited cognitive capacity of the decision maker. Cognitive psychology has made it possible to understand that the mind acts with a compromise between quality and effort when faced with very complex choices. To evaluate an event that has not yet happened, it is necessary to imagine that it happens in your head. This projection into the future requires a cognitive effort and is what differentiates choices under conditions of risk and choices under conditions of uncertainty. In fact, since the receipt of the outcome in choices under risk conditions is imminent, the mechanism of self-projection into the future is not necessary to imagine the consequence of the choice and the decision makers dwell on the objective analysis of possibilities. Financial activities, on the other hand, develop over time and the objective probability is too static to consider the anticipatory emotions that the self-projection mechanism generates in the investor. Assuming that uncertainty is inherent in valuations of events that have not yet occurred, the focus must shift from risk management to uncertainty management. Only in this way the intertemporal dimension of the decision-making environment and the haste generated by the financial market can be cautioned and considered. The work considers an extension of the prospectus theory with the temporal component with the aim of providing a description of the attitude towards risk with respect to the passage of time.

Keywords: impatience, risk aversion, subjective probability, uncertainty

Procedia PDF Downloads 78
14001 Survival Pattern of Under-five Mortality in High Focus States in India

Authors: Rahul Kumar

Abstract:

Background: Under-FiveMortality Rate(U5MR)ofanationiswidelyacceptedandlong-standing indicators of well-beingofherchildren.They measuredtheprobability of dying before theageoffive(expressedper1000livebirths).TheU5MRisanappropriate indicator of the cumulative exposure totheriskofdeathduringthefirstfiveyearsoflife, and accepted globalindicator ofthehealthandsocioeconomicstatusofagiven population.Itisalsousefulforassessing theimpactofvariousintervention programmes aimed at improving child survival.Under-fivemortalitytrendsconstitutealeadingindicatorofthelevel ofchildhealthandoveralldevelopmentincountries. Objectives: The first aim of our research is to study the level, trends, and Pattern of Under-five mortality using different sources of data. The second objective is to examine the survival pattern of Under-five mortality by different background characteristics. Data Source and Methodology: SRS and NFHS data have been used forobservingthelevelandtrendofUnder-Five mortality rate. Kaplan Meier Estimate has been used to understand the survival Pattern of Under-five mortality. Result: WefindthatallmostallthestatesmadesomeprogressbyreducingU5MRin recent decades.During1992-93highestU5MR(per thousand live birth) was observed in Assam(142)followed by up(141),Odisha(131),MP(130),andBihar(127.5).While the least U5MR(perthousandlive birth)wasobservedinRajasthan(102). The highestU5MR(per thousandlive birth)isobservedinUP(78.1), followed by MP(64.9)and Chhattisgarh(63.7)which are far away from the national level(50). Among them, Uttarakhand(46.7)hadleastU5MR(perthousandlivebirth), followed by Odisha(48.6). TheU5MR(perthousandlivebirth)ofcombinedhighfocusstateis63.7whichisfar away fromthenationallevel(50). Weidentified thatthesurvivalprobability ofunder-fivechildrenfromadolescentmotherislessin comparisontootherchildrenbornby differentagegroupofmothers. thatduringneonatalperiodusually male mortality exceedsthefemale mortality butthisdifferentialreversedinthepostneonatalperiod. Astheirageincreasesand approachingtofiveyears,weidentifiedthatthesurvivalprobability ofbothsexdecreasesbut female’s survival probabilitydecrement is more than male as their ageincreases. The poorer children’s survival probability is minimum. Children using improved toilet facility has more survival probability throughout thefiveyearsthan who uses unimproved. The survival probability of children under five who got Full ANCis more than the survival probability of children under five who doesn’t get any ANC. Conclusions: Improvement of maternal education is an urgent need to improve their health seeking behavior and thus the health of their children. Awareness on reproductive health and environmental sanitation should be strengthened.

Keywords: under-five mortality, survival pattern, ANC, trend

Procedia PDF Downloads 100
14000 Reliability Based Performance Evaluation of Stone Column Improved Soft Ground

Authors: A. GuhaRay, C. V. S. P. Kiranmayi, S. Rudraraju

Abstract:

The present study considers the effect of variation of different geotechnical random variables in the design of stone column-foundation systems for assessing the bearing capacity and consolidation settlement of highly compressible soil. The soil and stone column properties, spacing, diameter and arrangement of stone columns are considered as the random variables. Probability of failure (Pf) is computed for a target degree of consolidation and a target safe load by Monte Carlo Simulation (MCS). The study shows that the variation in coefficient of radial consolidation (cr) and cohesion of soil (cs) are two most important factors influencing Pf. If the coefficient of variation (COV) of cr exceeds 20%, Pf exceeds 0.001, which is unsafe following the guidelines of US Army Corps of Engineers. The bearing capacity also exceeds its safe value for COV of cs > 30%. It is also observed that as the spacing between the stone column increases, the probability of reaching a target degree of consolidation decreases. Accordingly, design guidelines, considering both consolidation and bearing capacity of improved ground, are proposed for different spacing and diameter of stone columns and geotechnical random variables.

Keywords: bearing capacity, consolidation, geotechnical random variables, probability of failure, stone columns

Procedia PDF Downloads 332
13999 Design and Application of a Model Eliciting Activity with Civil Engineering Students on Binomial Distribution to Solve a Decision Problem Based on Samples Data Involving Aspects of Randomness and Proportionality

Authors: Martha E. Aguiar-Barrera, Humberto Gutierrez-Pulido, Veronica Vargas-Alejo

Abstract:

Identifying and modeling random phenomena is a fundamental cognitive process to understand and transform reality. Recognizing situations governed by chance and giving them a scientific interpretation, without being carried away by beliefs or intuitions, is a basic training for citizens. Hence the importance of generating teaching-learning processes, supported using technology, paying attention to model creation rather than only executing mathematical calculations. In order to develop the student's knowledge about basic probability distributions and decision making; in this work a model eliciting activity (MEA) is reported. The intention was applying the Model and Modeling Perspective to design an activity related to civil engineering that would be understandable for students, while involving them in its solution. Furthermore, the activity should imply a decision-making challenge based on sample data, and the use of the computer should be considered. The activity was designed considering the six design principles for MEA proposed by Lesh and collaborators. These are model construction, reality, self-evaluation, model documentation, shareable and reusable, and prototype. The application and refinement of the activity was carried out during three school cycles in the Probability and Statistics class for Civil Engineering students at the University of Guadalajara. The analysis of the way in which the students sought to solve the activity was made using audio and video recordings, as well as with the individual and team reports of the students. The information obtained was categorized according to the activity phase (individual or team) and the category of analysis (sample, linearity, probability, distributions, mechanization, and decision-making). With the results obtained through the MEA, four obstacles have been identified to understand and apply the binomial distribution: the first one was the resistance of the student to move from the linear to the probabilistic model; the second one, the difficulty of visualizing (infering) the behavior of the population through the sample data; the third one, viewing the sample as an isolated event and not as part of a random process that must be viewed in the context of a probability distribution; and the fourth one, the difficulty of decision-making with the support of probabilistic calculations. These obstacles have also been identified in literature on the teaching of probability and statistics. Recognizing these concepts as obstacles to understanding probability distributions, and that these do not change after an intervention, allows for the modification of these interventions and the MEA. In such a way, the students may identify themselves the erroneous solutions when they carrying out the MEA. The MEA also showed to be democratic since several students who had little participation and low grades in the first units, improved their participation. Regarding the use of the computer, the RStudio software was useful in several tasks, for example in such as plotting the probability distributions and to exploring different sample sizes. In conclusion, with the models created to solve the MEA, the Civil Engineering students improved their probabilistic knowledge and understanding of fundamental concepts such as sample, population, and probability distribution.

Keywords: linear model, models and modeling, probability, randomness, sample

Procedia PDF Downloads 87
13998 Secrecy Analysis in Downlink Cellular Networks in the Presence of D2D Pairs and Hardware Impairment

Authors: Mahdi Rahimi, Mohammad Mahdi Mojahedian, Mohammad Reza Aref

Abstract:

In this paper, a cellular communication scenario with a transmitter and an authorized user is considered to analyze its secrecy in the face of eavesdroppers and the interferences propagated unintentionally through the communication network. It is also assumed that some D2D pairs and eavesdroppers are randomly located in the cell. Assuming hardware impairment, perfect connection probability is analytically calculated, and upper bound is provided for the secrecy outage probability. In addition, a method based on random activation of D2Ds is proposed to improve network security. Finally, the analytical results are verified by simulations.

Keywords: physical layer security, stochastic geometry, device-to-device, hardware impairment

Procedia PDF Downloads 141
13997 A Novel Approach of Secret Communication Using Douglas-Peucker Algorithm

Authors: R. Kiruthika, A. Kannan

Abstract:

Steganography is the problem of hiding secret messages in 'innocent – looking' public communication so that the presence of the secret message cannot be detected. This paper introduces a steganographic security in terms of computational in-distinguishability from a channel of probability distributions on cover messages. This method first splits the cover image into two separate blocks using Douglas – Peucker algorithm. The text message and the image will be hided in the Least Significant Bit (LSB) of the cover image.

Keywords: steganography, lsb, embedding, Douglas-Peucker algorithm

Procedia PDF Downloads 327
13996 Future Trends of Mechatronics Engineering in Pakistan

Authors: Aqeela Mir, Akhtar Nawaz Malik, Javaid Iqbal

Abstract:

The paper presents a survey based approach in order to observe the level of awareness regarding Mechatronics in society of Pakistan and the factors affecting the future development trend of Mechatronics in Pakistan. With the help of these surveys a new direction for making a Mathematical model for the future development trend of Mechatronics in Pakistan is also suggested.

Keywords: mechatronics society survey, future development trend of mechatronics in pakistan, probability estimation, mathematical model

Procedia PDF Downloads 473
13995 Traffic Safety and Risk Assessment Model by Analysis of Questionnaire Survey: A Case Study of S. G. Highway, Ahmedabad, India

Authors: Abhijitsinh Gohil, Kaushal Wadhvaniya, Kuldipsinh Jadeja

Abstract:

Road Safety is a multi-sectoral and multi-dimensional issue. An effective model can assess the risk associated with highway safety. A questionnaire survey is very essential to identify the events or activities which are causing unsafe condition for traffic on an urban highway. A questionnaire of standard questions including vehicular, human and infrastructure characteristics can be made. Responses from the age wise group of road users can be taken on field. Each question or an event holds a specific risk weightage, which contributes in creating an inappropriate and unsafe flow of traffic. The probability of occurrence of an event can be calculated from the data collected from the road users. Finally, the risk score can be calculated by considering the risk factor and the probability of occurrence of individual event and addition of all risk score for the individual event will give the total risk score of a particular road. Standards for risk score can be made and total risk score can be compared with the standards. Thus road can be categorized based on risk associated and traffic safety on it. With this model, one can assess the need for traffic safety improvement on a given road, and qualitative data can be analysed.

Keywords: probability of occurrence, questionnaire, risk factor, risk score

Procedia PDF Downloads 315
13994 Probabilistic Analysis of Bearing Capacity of Isolated Footing using Monte Carlo Simulation

Authors: Sameer Jung Karki, Gokhan Saygili

Abstract:

The allowable bearing capacity of foundation systems is determined by applying a factor of safety to the ultimate bearing capacity. Conventional ultimate bearing capacity calculations routines are based on deterministic input parameters where the nonuniformity and inhomogeneity of soil and site properties are not accounted for. Hence, the laws of mathematics like probability calculus and statistical analysis cannot be directly applied to foundation engineering. It’s assumed that the Factor of Safety, typically as high as 3.0, incorporates the uncertainty of the input parameters. This factor of safety is estimated based on subjective judgement rather than objective facts. It is an ambiguous term. Hence, a probabilistic analysis of the bearing capacity of an isolated footing on a clayey soil is carried out by using the Monte Carlo Simulation method. This simulated model was compared with the traditional discrete model. It was found out that the bearing capacity of soil was found higher for the simulated model compared with the discrete model. This was verified by doing the sensitivity analysis. As the number of simulations was increased, there was a significant % increase of the bearing capacity compared with discrete bearing capacity. The bearing capacity values obtained by simulation was found to follow a normal distribution. While using the traditional value of Factor of safety 3, the allowable bearing capacity had lower probability (0.03717) of occurring in the field compared to a higher probability (0.15866), while using the simulation derived factor of safety of 1.5. This means the traditional factor of safety is giving us bearing capacity that is less likely occurring/available in the field. This shows the subjective nature of factor of safety, and hence probability method is suggested to address the variability of the input parameters in bearing capacity equations.

Keywords: bearing capacity, factor of safety, isolated footing, montecarlo simulation

Procedia PDF Downloads 155
13993 Fault Tree Analysis and Bayesian Network for Fire and Explosion of Crude Oil Tanks: Case Study

Authors: B. Zerouali, M. Kara, B. Hamaidi, H. Mahdjoub, S. Rouabhia

Abstract:

In this paper, a safety analysis for crude oil tanks to prevent undesirable events that may cause catastrophic accidents. The estimation of the probability of damage to industrial systems is carried out through a series of steps, and in accordance with a specific methodology. In this context, this work involves developing an assessment tool and risk analysis at the level of crude oil tanks system, based primarily on identification of various potential causes of crude oil tanks fire and explosion by the use of Fault Tree Analysis (FTA), then improved risk modelling by Bayesian Networks (BNs). Bayesian approach in the evaluation of failure and quantification of risks is a dynamic analysis approach. For this reason, have been selected as an analytical tool in this study. Research concludes that the Bayesian networks have a distinct and effective method in the safety analysis because of the flexibility of its structure; it is suitable for a wide variety of accident scenarios.

Keywords: bayesian networks, crude oil tank, fault tree, prediction, safety

Procedia PDF Downloads 629
13992 Frequency Analysis Using Multiple Parameter Probability Distributions for Rainfall to Determine Suitable Probability Distribution in Pakistan

Authors: Tasir Khan, Yejuan Wang

Abstract:

The study of extreme rainfall events is very important for flood management in river basins and the design of water conservancy infrastructure. Evaluation of quantiles of annual maximum rainfall (AMRF) is required in different environmental fields, agriculture operations, renewable energy sources, climatology, and the design of different structures. Therefore, the annual maximum rainfall (AMRF) was performed at different stations in Pakistan. Multiple probability distributions, log normal (LN), generalized extreme value (GEV), Gumbel (max), and Pearson type3 (P3) were used to find out the most appropriate distributions in different stations. The L moments method was used to evaluate the distribution parameters. Anderson darling test, Kolmogorov- Smirnov test, and chi-square test showed that two distributions, namely GUM (max) and LN, were the best appropriate distributions. The quantile estimate of a multi-parameter PD offers extreme rainfall through a specific location and is therefore important for decision-makers and planners who design and construct different structures. This result provides an indication of these multi-parameter distribution consequences for the study of sites and peak flow prediction and the design of hydrological maps. Therefore, this discovery can support hydraulic structure and flood management.

Keywords: RAMSE, multiple frequency analysis, annual maximum rainfall, L-moments

Procedia PDF Downloads 52
13991 Approximate Confidence Interval for Effect Size Base on Bootstrap Resampling Method

Authors: S. Phanyaem

Abstract:

This paper presents the confidence intervals for the effect size base on bootstrap resampling method. The meta-analytic confidence interval for effect size is proposed that are easy to compute. A Monte Carlo simulation study was conducted to compare the performance of the proposed confidence intervals with the existing confidence intervals. The best confidence interval method will have a coverage probability close to 0.95. Simulation results have shown that our proposed confidence intervals perform well in terms of coverage probability and expected length.

Keywords: effect size, confidence interval, bootstrap method, resampling

Procedia PDF Downloads 566
13990 Invalidation of the Start of Lunar Calendars Based on Sighting of Crescent: A Survey of 101 Years of Data between 1938 and 2038

Authors: Rafik Ouared

Abstract:

The purpose of this paper is to invalidate decisions made by the Islamic conference led at Istanbul in 2016, which had defined two basic criteria to determine the start of the lunar month: (1)they are all based on the sighting of the crescent, be it observed or computed with modern methods, and (2) they've strongly recommended the adoption of the principle of 'unification of sighting', by which any occurrence of sighting anywhere would be applicable everywhere. To demonstrate the invalidation of those statements, a survey of 101 years of data, from 1938 to 2038, have been analyzed to compare the probability density function (PDF) of time difference between different types of fajr and new moon. Two groups of fajr have been considered: the 'natural fajr', which is the very first fajr following new moon, and the 'biased fajr', which is defined by human being inclusively of all chosen definitions. The parametric and non-parametric statistical comparisons between the different groups have shown the all the biased PDFs are significantly different from the unbiased (natural) PDF with probability value (p-value) less than 0.001. The significance level was fixed to 0.05. Conclusion: the on-going reference to sighting of crescent is inducing an significant bias in defining lunar calendar. Therefore, 'natural' calendar would be more applicable requiring a more contextualized revision of issue in fiqh.

Keywords: biased fajr, lunar calendar, natural fajr, probability density function, sighting of crescent, time difference between fajr and new moon

Procedia PDF Downloads 183
13989 Constructions of Linear and Robust Codes Based on Wavelet Decompositions

Authors: Alla Levina, Sergey Taranov

Abstract:

The classical approach to the providing noise immunity and integrity of information that process in computing devices and communication channels is to use linear codes. Linear codes have fast and efficient algorithms of encoding and decoding information, but this codes concentrate their detect and correct abilities in certain error configurations. To protect against any configuration of errors at predetermined probability can robust codes. This is accomplished by the use of perfect nonlinear and almost perfect nonlinear functions to calculate the code redundancy. The paper presents the error-correcting coding scheme using biorthogonal wavelet transform. Wavelet transform applied in various fields of science. Some of the wavelet applications are cleaning of signal from noise, data compression, spectral analysis of the signal components. The article suggests methods for constructing linear codes based on wavelet decomposition. For developed constructions we build generator and check matrix that contain the scaling function coefficients of wavelet. Based on linear wavelet codes we develop robust codes that provide uniform protection against all errors. In article we propose two constructions of robust code. The first class of robust code is based on multiplicative inverse in finite field. In the second robust code construction the redundancy part is a cube of information part. Also, this paper investigates the characteristics of proposed robust and linear codes.

Keywords: robust code, linear code, wavelet decomposition, scaling function, error masking probability

Procedia PDF Downloads 463
13988 An Analysis of the Effect of Sharia Financing and Work Relation Founding towards Non-Performing Financing in Islamic Banks in Indonesia

Authors: Muhammad Bahrul Ilmi

Abstract:

The purpose of this research is to analyze the influence of Islamic financing and work relation founding simultaneously and partially towards non-performing financing in Islamic banks. This research was regression quantitative field research, and had been done in Muammalat Indonesia Bank and Islamic Danamon Bank in 3 months. The populations of this research were 15 account officers of Muammalat Indonesia Bank and Islamic Danamon Bank in Surakarta, Indonesia. The techniques of collecting data used in this research were documentation, questionnaire, literary study and interview. Regression analysis result shows that Islamic financing and work relation founding simultaneously has positive and significant effect towards non performing financing of two Islamic Banks. It is obtained with probability value 0.003 which is less than 0.05 and F value 9.584. The analysis result of Islamic financing regression towards non performing financing shows the significant effect. It is supported by double linear regression analysis with probability value 0.001 which is less than 0.05. The regression analysis of work relation founding effect towards non-performing financing shows insignificant effect. This is shown in the double linear regression analysis with probability value 0.161 which is bigger than 0.05.

Keywords: Syariah financing, work relation founding, non-performing financing (NPF), Islamic Bank

Procedia PDF Downloads 404
13987 A Knowledge-Based Development of Risk Management Approaches for Construction Projects

Authors: Masoud Ghahvechi Pour

Abstract:

Risk management is a systematic and regular process of identifying, analyzing and responding to risks throughout the project's life cycle in order to achieve the optimal level of elimination, reduction or control of risk. The purpose of project risk management is to increase the probability and effect of positive events and reduce the probability and effect of unpleasant events on the project. Risk management is one of the most fundamental parts of project management, so that unmanaged or untransmitted risks can be one of the primary factors of failure in a project. Effective risk management does not apply to risk regression, which is apparently the cheapest option of the activity. However, the main problem with this option is the economic sensitivity, because what is potentially profitable is by definition risky, and what does not pose a risk is economically interesting and does not bring tangible benefits. Therefore, in relation to the implemented project, effective risk management is finding a "middle ground" in its management, which includes, on the one hand, protection against risk from a negative direction by means of accurate identification and classification of risk, which leads to analysis And it becomes a comprehensive analysis. On the other hand, management using all mathematical and analytical tools should be based on checking the maximum benefits of these decisions. Detailed analysis, taking into account all aspects of the company, including stakeholder analysis, will allow us to add what will become tangible benefits for our project in the future to effective risk management. Identifying the risk of the project is based on the theory that which type of risk may affect the project, and also refers to specific parameters and estimating the probability of their occurrence in the project. These conditions can be divided into three groups: certainty, uncertainty, and risk, which in turn support three types of investment: risk preference, risk neutrality, specific risk deviation, and its measurement. The result of risk identification and project analysis is a list of events that indicate the cause and probability of an event, and a final assessment of its impact on the environment.

Keywords: risk, management, knowledge, risk management

Procedia PDF Downloads 29
13986 Congestion Control in Mobile Network by Prioritizing Handoff Calls

Authors: O. A. Lawal, O. A Ojesanmi

Abstract:

The demand for wireless cellular services continues to increase while the radio resources remain limited. Thus, network operators have to continuously manage the scarce radio resources in order to have an improved quality of service for mobile users. This paper proposes how to handle the problem of congestion in the mobile network by prioritizing handoff call, using the guard channel allocation scheme. The research uses specific threshold value for the time of allocation of the channel in the algorithm. The scheme would be simulated by generating various data for different traffics in the network as it would be in the real life. The result would be used to determine the probability of handoff call dropping and the probability of the new call blocking as a way of measuring the network performance.

Keywords: call block, channel, handoff, mobile cellular network

Procedia PDF Downloads 366
13985 Comparison of Wind Fragility for Window System in the Simplified 10 and 15-Story Building Considering Exposure Category

Authors: Viriyavudh Sim, WooYoung Jung

Abstract:

Window system in high rise building is occasionally subjected to an excessive wind intensity, particularly during typhoon. The failure of window system did not affect overall safety of structural performance; however, it could endanger the safety of the residents. In this paper, comparison of fragility curves for window system of two residential buildings was studied. The probability of failure for individual window was determined with Monte Carlo Simulation method. Then, lognormal cumulative distribution function was used to represent the fragility. The results showed that windows located on the edge of leeward wall were more susceptible to wind load and the probability of failure for each window panel increased at higher floors.

Keywords: wind fragility, window system, high rise building, wind disaster

Procedia PDF Downloads 292
13984 The Impact of Corporate Social Responsibility and Relationship Marketing on Relationship Maintainer and Customer Loyalty by Mediating Role of Customer Satisfaction

Authors: Anam Bhatti, Sumbal Arif, Mariam Mehar, Sohail Younas

Abstract:

CSR has become one of the imperative implements in satisfying customers. The impartial of this research is to calculate CSR, relationship marketing, and customer satisfaction. In Pakistan, there is not enough research work on the effect of CSR and relationship marketing on relationship maintainer and customer loyalty. To find out deductive approach and survey method is used as research approach and research strategy respectively. This research design is descriptive and quantitative study. For data, collection questionnaire method with semantic differential scale and seven point scales are adopted. Data has been collected by adopting the non-probability convenience technique as sampling technique and the sample size is 400. For factor confirmatory factor analysis, structure equation modeling and medication analysis, regression analysis Amos software were used. Strong empirical evidence supports that the customer’s perception of CSR performance is highly influenced by the values.

Keywords: CSR, Relationship marketing, Relationship maintainer, Customer loyalty, Customer satisfaction

Procedia PDF Downloads 441
13983 Seismic Design Approach for Areas with Low Seismicity

Authors: Mogens Saberi

Abstract:

The following article focuses on a new seismic design approach for Denmark. Denmark is located in a low seismic zone and up till now a general and very simplified approach has been used to accommodate the effect of seismic loading. The current used method is presented and it is found that the approach is on the unsafe side for many building types in Denmark. The damages during time due to earth quake is presented and a seismic map for Denmark is developed and presented. Furthermore, a new design approach is suggested and compared to the existing one. The new approach is relatively simple but captures the effect of seismic loading more realistic than the existing one. The new approach is believed to the incorporated in the Danish Deign Code for building structures.

Keywords: low seismicity, new design approach, earthquakes, Denmark

Procedia PDF Downloads 331
13982 Flexural Fatigue Performance of Self-Compacting Fibre Reinforced Concrete

Authors: Surinder Pal Singh, Sanjay Goel

Abstract:

The paper presents results of an investigation conducted to study the flexural fatigue characteristics of Self Compacting Concrete (SCC) and Self Compacting Fibre Reinforced Concrete (SCFRC). In total 360 flexural fatigue tests and 270 static flexural strength tests were conducted on SCC and SCFRC specimens to obtain the fatigue test data. The variability in the distribution of fatigue life of SCC and SCFRC have been analyzed and compared with that of NVC and NVFRC containing steel fibres of comparable size and shape. The experimental coefficients of fatigue equations have been estimated to represent relationship between stress level (S) and fatigue life (N) for SCC and SCFRC containing different fibre volume fractions. The probability of failure (Pf) has been incorporated in S-N relationships to obtain families of S-N-Pf relationships. A good agreement between the predicted curves and those obtained from the test data has been observed. The fatigue performance of SCC and SCFRC has been evaluated in terms of two-million cycles fatigue strength/endurance limit. The theoretic fatigue lives were also estimated using single-log fatigue equation for 10% probability of failure to estimate the enhanced extent of theoretic fatigue lives of SCFRC with reference to SCC and NVC. The reduction in variability in the fatigue life, increased endurance limit and increased theoretiac fatigue lives demonstrates an overall better fatigue performance for SCC and SCFRC.

Keywords: fatigue life, fibre, probability of failure, self-compacting concrete

Procedia PDF Downloads 329
13981 Wind Fragility of Window Glass in 10-Story Apartment with Two Different Window Models

Authors: Viriyavudh Sim, WooYoung Jung

Abstract:

Damage due to high wind is not limited to load resistance components such as beam and column. The majority of damage is due to breach in the building envelope such as broken roof, window, and door. In this paper, wind fragility of window glass in residential apartment was determined to compare the difference between two window configuration models. Monte Carlo Simulation method had been used to derive damage data and analytical fragilities were constructed. Fragility of window system showed that window located in leeward wall had higher probability of failure, especially those close to the edge of structure. Between the two window models, Model 2 had higher probability of failure, this was due to the number of panel in this configuration.

Keywords: wind fragility, glass window, high rise building, wind disaster

Procedia PDF Downloads 234
13980 Optimal Scheduling for Energy Storage System Considering Reliability Constraints

Authors: Wook-Won Kim, Je-Seok Shin, Jin-O Kim

Abstract:

This paper propose the method for optimal scheduling for battery energy storage system with reliability constraint of energy storage system in reliability aspect. The optimal scheduling problem is solved by dynamic programming with proposed transition matrix. Proposed optimal scheduling method guarantees the minimum fuel cost within specific reliability constraint. For evaluating proposed method, the timely capacity outage probability table (COPT) is used that is calculated by convolution of probability mass function of each generator. This study shows the result of optimal schedule of energy storage system.

Keywords: energy storage system (ESS), optimal scheduling, dynamic programming, reliability constraints

Procedia PDF Downloads 376
13979 Examining Motivational Strategies of Foreign Manufacturing Firms in Ghana

Authors: Samuel Ato Dadzie

Abstract:

The objective of this study is to examine the influence of eclectic paradigm on motivational strategy of foreign subsidiaries in Ghana. This study uses binary regression model, and the analysis was based on 75 manufacturing investments made by MNEs from different countries in 1994–2008. The results indicated that perceived market size increases the probability of foreign firms undertaking a market seeking (MS) in Ghana, while perceived cultural distance between Ghana and foreign firm’s home countries decreased the probability of foreign firms undertaking an market seeking (MS) foreign direct investment (FDI) in Ghana. Furthermore, extensive international experience decreases the probability of foreign firms undertaking a market seeking (MS) foreign direct investment (FDI) in Ghana. Most of the studies done by earlier researchers were based on the advanced and emerging countries and offered support for the theory, which was used in generalizing the result that multinational corporations (MNCs) normally used the theory regarding investment strategy outside their home country. In using the same theory in the context of Ghana, the result does not offer strong support for the theory. This means that MNCs that come to Sub-Sahara Africa cannot rely much on eclectic paradigm for their motivational strategies because prevailing economic conditions in Ghana are different from that of the advanced and emerging economies where the institutional structures work.

Keywords: foreign subsidiary, motives, Ghana, foreign direct investment

Procedia PDF Downloads 402
13978 Analytical Modeling of Globular Protein-Ferritin in α-Helical Conformation: A White Noise Functional Approach

Authors: Vernie C. Convicto, Henry P. Aringa, Wilson I. Barredo

Abstract:

This study presents a conformational model of the helical structures of globular protein particularly ferritin in the framework of white noise path integral formulation by using Associated Legendre functions, Bessel and convolution of Bessel and trigonometric functions as modulating functions. The model incorporates chirality features of proteins and their helix-turn-helix sequence structural motif.

Keywords: globular protein, modulating function, white noise, winding probability

Procedia PDF Downloads 443
13977 Kauffman Model on a Network of Containers

Authors: Johannes J. Schneider, Mathias S. Weyland, Peter Eggenberger Hotz, William D. Jamieson, Oliver Castell, Alessia Faggian, Rudolf M. Füchslin

Abstract:

In the description of the origin of life, there are still some open gaps, e.g., the formation of macromolecules cannot be fully explained so far. The Kauffman model proposes the existence of autocatalytic sets of macromolecules which mutually catalyze reactions leading to each other’s formation. Usually, this model is simulated in one well-stirred pot only, with a continuous inflow of small building blocks, from which larger molecules are created by a set of catalyzed ligation and cleavage reactions. This approach represents the picture of the primordial soup. However, the conditions on the early Earth must have differed geographically, leading to spatially different outcomes whether a specific reaction could be performed or not. Guided by this picture, the Kauffman model is simulated in a large number of containers in parallel, with neighboring containers being connected by diffusion. In each container, only a subset of the overall reaction set can be performed. Under specific conditions, this approach leads to a larger probability for the existence of an autocatalytic metabolism than in the original Kauffman model.

Keywords: agglomeration, autocatalytic set, differential equation, Kauffman model

Procedia PDF Downloads 26