Search results for: cumulative probabilities
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 536

Search results for: cumulative probabilities

356 Pension Reform in Georgia: Challenges, International Practice and Opportunities for Development

Authors: Manana Lobzhanidze

Abstract:

Reforming the pension system is urgent in Georgia due to socio-economic problems. Replacing the current pension system with a new one requires, on the one hand, an assessment of the challenges in this field and, on the other hand, a study of the best practices of foreign experience. Objectives: The aim of the research is to identify challenges in the pension reform process in Georgia, to study international experience, and to develop recommendations for the implementation of an effective pension system. Methodologies: A desk study was conducted, and methods of analysis, comparison, grouping, matrix charts, and scenario analysis were used. Findings: The advantages of accumulative pension compared to the current pension system are identified. The main challenge is the non-targeting of the pension contributions and the ineffective investment policy; the public's attitude towards the cumulative pension system is determined.

Keywords: pension reform, challenges, international practice, opportunity for development

Procedia PDF Downloads 36
355 The Theory behind Logistic Regression

Authors: Jan Henrik Wosnitza

Abstract:

The logistic regression has developed into a standard approach for estimating conditional probabilities in a wide range of applications including credit risk prediction. The article at hand contributes to the current literature on logistic regression fourfold: First, it is demonstrated that the binary logistic regression automatically meets its model assumptions under very general conditions. This result explains, at least in part, the logistic regression's popularity. Second, the requirement of homoscedasticity in the context of binary logistic regression is theoretically substantiated. The variances among the groups of defaulted and non-defaulted obligors have to be the same across the level of the aggregated default indicators in order to achieve linear logits. Third, this article sheds some light on the question why nonlinear logits might be superior to linear logits in case of a small amount of data. Fourth, an innovative methodology for estimating correlations between obligor-specific log-odds is proposed. In order to crystallize the key ideas, this paper focuses on the example of credit risk prediction. However, the results presented in this paper can easily be transferred to any other field of application.

Keywords: correlation, credit risk estimation, default correlation, homoscedasticity, logistic regression, nonlinear logistic regression

Procedia PDF Downloads 394
354 CE Method for Development of Japan's Stochastic Earthquake Catalogue

Authors: Babak Kamrani, Nozar Kishi

Abstract:

Stochastic catalog represents the events module of the earthquake loss estimation models. It includes series of events with different magnitudes and corresponding frequencies/probabilities. For the development of the stochastic catalog, random or uniform sampling methods are used to sample the events from the seismicity model. For covering all the Magnitude Frequency Distribution (MFD), a huge number of events should be generated for the above-mentioned methods. Characteristic Event (CE) method chooses the events based on the interest of the insurance industry. We divide the MFD of each source into bins. We have chosen the bins based on the probability of the interest by the insurance industry. First, we have collected the information for the available seismic sources. Sources are divided into Fault sources, subduction, and events without specific fault source. We have developed the MFD for each of the individual and areal source based on the seismicity of the sources. Afterward, we have calculated the CE magnitudes based on the desired probability. To develop the stochastic catalog, we have introduced uncertainty to the location of the events too.

Keywords: stochastic catalogue, earthquake loss, uncertainty, characteristic event

Procedia PDF Downloads 269
353 Vulnerable Paths Assessment for Distributed Denial of Service Attacks in a Cloud Computing Environment

Authors: Manas Tripathi, Arunabha Mukhopadhyay

Abstract:

In Cloud computing environment, cloud servers, sometimes may crash after receiving huge amount of request and cloud services may stop which can create huge loss to users of that cloud services. This situation is called Denial of Service (DoS) attack. In Distributed Denial of Service (DDoS) attack, an attacker targets multiple network paths by compromising various vulnerable systems (zombies) and floods the victim with huge amount of request through these zombies. There are many solutions to mitigate this challenge but most of the methods allows the attack traffic to arrive at Cloud Service Provider (CSP) and then only takes actions against mitigation. Here in this paper we are rather focusing on preventive mechanism to deal with these attacks. We analyze network topology and find most vulnerable paths beforehand without waiting for the traffic to arrive at CSP. We have used Dijkstra's and Yen’s algorithm. Finally, risk assessment of these paths can be done by multiplying the probabilities of attack for these paths with the potential loss.

Keywords: cloud computing, DDoS, Dijkstra, Yen’s k-shortest path, network security

Procedia PDF Downloads 257
352 Comprehensive, Up-to-Date Climate System Change Indicators, Trends and Interactions

Authors: Peter Carter

Abstract:

Comprehensive climate change indicators and trends inform the state of the climate (system) with respect to present and future climate change scenarios and the urgency of mitigation and adaptation. With data records now going back for many decades, indicator trends can complement model projections. They are provided as datasets by several climate monitoring centers, reviewed by state of the climate reports, and documented by the IPCC assessments. Up-to-date indicators are provided here. Rates of change are instructive, as are extremes. The indicators include greenhouse gas (GHG) emissions (natural and synthetic), cumulative CO2 emissions, atmospheric GHG concentrations (including CO2 equivalent), stratospheric ozone, surface ozone, radiative forcing, global average temperature increase, land temperature increase, zonal temperature increases, carbon sinks, soil moisture, sea surface temperature, ocean heat content, ocean acidification, ocean oxygen, glacier mass, Arctic temperature, Arctic sea ice (extent and volume), northern hemisphere snow cover, permafrost indices, Arctic GHG emissions, ice sheet mass, sea level rise, and stratospheric and surface ozone. Global warming is not the most reliable single metric for the climate state. Radiative forcing, atmospheric CO2 equivalent, and ocean heat content are more reliable. Global warming does not provide future commitment, whereas atmospheric CO2 equivalent does. Cumulative carbon is used for estimating carbon budgets. The forcing of aerosols is briefly addressed. Indicator interactions are included. In particular, indicators can provide insight into several crucial global warming amplifying feedback loops, which are explained. All indicators are increasing (adversely), most as fast as ever and some faster. One particularly pressing indicator is rapidly increasing global atmospheric methane. In this respect, methane emissions and sources are covered in more detail. In their application, indicators used in assessing safe planetary boundaries are included. Indicators are considered with respect to recent published papers on possible catastrophic climate change and climate system tipping thresholds. They are climate-change-policy relevant. In particular, relevant policies include the 2015 Paris Agreement on “holding the increase in the global average temperature to well below 2°C above pre-industrial levels and pursuing efforts to limit the temperature increase to 1.5°C above pre-industrial levels” and the 1992 UN Framework Convention on Climate change, which has “stabilization of greenhouse gas concentrations in the atmosphere at a level that would prevent dangerous anthropogenic interference with the climate system.”

Keywords: climate change, climate change indicators, climate change trends, climate system change interactions

Procedia PDF Downloads 77
351 Trace Metals in Natural Bottled Water on Montenegrin Market and Comaparison with Tap Water in Podgorica

Authors: Katarina Živković, Ivana Joksimović

Abstract:

Many different chemicals may occur in drinking water and cause significant human health risks after prolonged periods of exposure. In particular concern are contaminants that have cumulative toxic properties, such as heavy metals. This investigation was done to clarify concerns about chemical quality and safety of drinking tap water in Podgorica. For comparison, all available natural bottled water on Montenegrin market were bought. All samples (bottled water and tap water from Podgorica) were analyzed using ICP –OES on contents of Al, Cd, Pb, Cu, Zn,Cr, Fe, As and Mn. All results compared with the maximum concentration levels allowed by international standards and World Health Organization (WHO) guidelines. The results of analysis showed that all trace of heavy metals were very low and in same time below MCL according to WHO and International standard.

Keywords: inductively coupled plasma - optical emission spectrometry (ICP-OES), Montenegro (Podgorica), natural bottled water, tap water , trace of heavy metal

Procedia PDF Downloads 425
350 An Analysis of a Queueing System with Heterogeneous Servers Subject to Catastrophes

Authors: M. Reni Sagayaraj, S. Anand Gnana Selvam, R. Reynald Susainathan

Abstract:

This study analyzed a queueing system with blocking and no waiting line. The customers arrive according to a Poisson process and the service times follow exponential distribution. There are two non-identical servers in the system. The queue discipline is FCFS, and the customers select the servers on fastest server first (FSF) basis. The service times are exponentially distributed with parameters μ1 and μ2 at servers I and II, respectively. Besides, the catastrophes occur in a Poisson manner with rate γ in the system. When server I is busy or blocked, the customer who arrives in the system leaves the system without being served. Such customers are called lost customers. The probability of losing a customer was computed for the system. The explicit time dependent probabilities of system size are obtained and a numerical example is presented in order to show the managerial insights of the model. Finally, the probability that arriving customer finds system busy and average number of server busy in steady state are obtained numerically.

Keywords: queueing system, blocking, poisson process, heterogeneous servers, queue discipline FCFS, busy period

Procedia PDF Downloads 477
349 Mathematics Model Approaching: Parameter Estimation of Transmission Dynamics of HIV and AIDS in Indonesia

Authors: Endrik Mifta Shaiful, Firman Riyudha

Abstract:

Acquired Immunodeficiency Syndrome (AIDS) is one of the world's deadliest diseases caused by the Human Immunodeficiency Virus (HIV) that infects white blood cells and cause a decline in the immune system. AIDS quickly became a world epidemic disease that affects almost all countries. Therefore, mathematical modeling approach to the spread of HIV and AIDS is needed to anticipate the spread of HIV and AIDS which are widespread. The purpose of this study is to determine the parameter estimation on mathematical models of HIV transmission and AIDS using cumulative data of people with HIV and AIDS each year in Indonesia. In this model, there are parameters of r ∈ [0,1) which is the effectiveness of the treatment in patients with HIV. If the value of r is close to 1, the number of people with HIV and AIDS will decline toward zero. The estimation results indicate when the value of r is close to unity, there will be a significant decline in HIV patients, whereas in AIDS patients constantly decreases towards zero.

Keywords: HIV, AIDS, parameter estimation, mathematical models

Procedia PDF Downloads 220
348 Analysis of Patient No-Shows According to Health Conditions

Authors: Sangbok Lee

Abstract:

There has been much effort on process improvement for outpatient clinics to provide quality and acute care to patients. One of the efforts is no-show analysis or prediction. This work analyzes patient no-shows along with patient health conditions. The health conditions refer to clinical symptoms that each patient has, out of the followings; hyperlipidemia, diabetes, metastatic solid tumor, dementia, chronic obstructive pulmonary disease, hypertension, coronary artery disease, myocardial infraction, congestive heart failure, atrial fibrillation, stroke, drug dependence abuse, schizophrenia, major depression, and pain. A dataset from a regional hospital is used to find the relationship between the number of the symptoms and no-show probabilities. Additional analysis reveals how each symptom or combination of symptoms affects no-shows. In the above analyses, cross-classification of patients by age and gender is carried out. The findings from the analysis will be used to take extra care to patients with particular health conditions. They will be forced to visit clinics by being informed about their health conditions and possible consequences more clearly. Moreover, this work will be used in the preparation of making institutional guidelines for patient reminder systems.

Keywords: healthcare system, no show analysis, process improvment, statistical data analysis

Procedia PDF Downloads 209
347 Impacts of Electronic Dance Music towards Social Harmony: The Malaysian Perspective

Authors: Kok Meng Ng, Sulung Veronica

Abstract:

Electronic Dance Music (EDM), a musical event that so sought-after amongst the youth, is getting prevailed around the world. The emergence of this à la mode event has magnetized lots of attentions from the media as well as the public due to its high probabilities in creating social problems and menacing social harmony of one destination, for instance, two death cases occurred during the EDM events in Malaysia caused a feeling of consternation of the society. The arguments over the impacts of such events towards the society are endless. This paper focuses on the study of the impacts of EDM towards social harmony in Klang Valley area, Malaysia by scrutinizing the contradiction of statements from several experts and the local communities. This study sampled 15-20 people that represent different social background with face-to-face and online interview through snowball sampling method. This study helps to understand the social context as a whole based on the impacts of EDM events that take place in Malaysia. It also provides valuable information to EDMs’ organizer as well as local authorities for a proper event management to minimize EDM impacts towards society as part of the sustainable growth of the event industry.

Keywords: electronic dance music, social harmony, impacts, Klang Valley

Procedia PDF Downloads 223
346 Markov Characteristics of the Power Line Communication Channels in China

Authors: Ming-Yue Zhai

Abstract:

Due to the multipath and pulse noise nature, power line communications(PLC) channel can be modelled as a memory one with the finite states Markov model(FSMC). As the most important parameter modelling a Markov channel,the memory order in an FSMC is not solved in PLC systems yet. In the paper, the mutual information is used as a measure of the dependence between the different symbols, treated as the received SNA or amplitude of the current channel symbol or that of previous symbols. The joint distribution probabilities of the envelopes in PLC systems are computed based on the multi-path channel model, which is commonly used in PLC. we confirm that given the information of the symbol immediately preceding the current one, any other previous symbol is independent of the current one in PLC systems, which means the PLC channels is a Markov chain with the first-order. The field test is also performed to model the received OFDM signals with the help of AR model. The results show that the first-order AR model is enough to model the fading channel in PLC systems, which means the amount of uncertainty remaining in the current symbol should be negligible, given the information corresponding to the immediately preceding one.

Keywords: power line communication, channel model, markovian, information theory, first-order

Procedia PDF Downloads 380
345 Heart-Rate Resistance Electrocardiogram Identification Based on Slope-Oriented Neural Networks

Authors: Tsu-Wang Shen, Shan-Chun Chang, Chih-Hsien Wang, Te-Chao Fang

Abstract:

For electrocardiogram (ECG) biometrics system, it is a tedious process to pre-install user’s high-intensity heart rate (HR) templates in ECG biometric systems. Based on only resting enrollment templates, it is a challenge to identify human by using ECG with the high-intensity HR caused from exercises and stress. This research provides a heartbeat segment method with slope-oriented neural networks against the ECG morphology changes due to high intensity HRs. The method has overall system accuracy at 97.73% which includes six levels of HR intensities. A cumulative match characteristic curve is also used to compare with other traditional ECG biometric methods.

Keywords: high-intensity heart rate, heart rate resistant, ECG human identification, decision based artificial neural network

Procedia PDF Downloads 394
344 Volatility Switching between Two Regimes

Authors: Josip Visković, Josip Arnerić, Ante Rozga

Abstract:

Based on the fact that volatility is time varying in high frequency data and that periods of high volatility tend to cluster, the most successful and popular models in modelling time varying volatility are GARCH type models. When financial returns exhibit sudden jumps that are due to structural breaks, standard GARCH models show high volatility persistence, i.e. integrated behaviour of the conditional variance. In such situations models in which the parameters are allowed to change over time are more appropriate. This paper compares different GARCH models in terms of their ability to describe structural changes in returns caused by financial crisis at stock markets of six selected central and east European countries. The empirical analysis demonstrates that Markov regime switching GARCH model resolves the problem of excessive persistence and outperforms uni-regime GARCH models in forecasting volatility when sudden switching occurs in response to financial crisis.

Keywords: central and east European countries, financial crisis, Markov switching GARCH model, transition probabilities

Procedia PDF Downloads 201
343 Debt Reconstruction, Career Development and Famers Household Well-Being in Thailand

Authors: Yothin Sawangdee, Piyawat Katewongsa, Chutima Yousomboon, Kornkanok Pongpradit, Sakapas Saengchai, Phusit Khantikul

Abstract:

Debts reconstruction under some of moratorium projects is one of important method that highly benefits to both the Banks and farmers. The method can reduce probabilities for nonprofits loan. This paper discuss about debts reconstruction and career development training for farmers in Thailand between 2011 and 2013. The research designed is mix-method between quantitative survey and qualitative survey. Sample size for quantitative method is 1003 cases. Data gathering procedure is between October and December 2013. Main results affirmed that debts reconstruction is needed. And there are numerous benefits from farmers’ career development training. Many of farmers who attend field school activities able to bring knowledge learned to apply for the farms’ work. They can reduce production costs. Framers’ quality of life and their household well-being also improve. This program should apply in any countries where farmers have highly debts and highly risks for not return the debts.

Keywords: career development, debts reconstruction, farmers household well-being, Thailand

Procedia PDF Downloads 384
342 When Change Is the Only Constant: The Impact of Change Frequency and Diversity on Change Appraisal

Authors: Danika Pieters

Abstract:

Due to changing societal and economic demands, organizational change has become increasingly prevalent in work life. While a long time change research has focused on the effects of single discrete change events on different employee outcomes such as job satisfaction and organizational commitment, a nascent research stream has begun to look into the potential cumulative effects of change in the context of continuous intense reforms. This case study of a large Belgian public organization aims to add to this growing literature by examining how the frequency and diversity of past changes impact employees’ appraisals of a newly introduced change. Twelve hundred survey results were analyzed using standard ordinary least squares regression. Results showed a correlation between high past change frequency and diversity and a negative appraisal of the new change. Implications for practitioners and future research are discussed.

Keywords: change frequency, change diversity, organizational changes, change appraisal, change evaluation

Procedia PDF Downloads 101
341 Planning of Green Infrastructure on a City Level

Authors: James Li, Darko Joksimovic

Abstract:

Urban development changes the natural hydrologic cycle, resulting in storm water impacts such as flooding, water quality degradation, receiving water erosion, and ecosystem deterioration. An integrated storm water managementapproach utilizing source and conveyance (termed green infrastructure) and end-of-pipe control measures is an effective way to manage urban storm water impacts. This paper focuses onplanning green infrastructure (GI) at the source and along the drainage system on a city level. It consists of (1)geospatial analysis of feasible GI using physical suitability; (2) modelling of cumulative GI's stormwater performance; and (3) cost-effectiveness analysis to prioritize the implementation of GI. A case study of the City of Barrie in Ontario, Canada, was used to demonstrate the GI's planning.

Keywords: cost-effectiveness of storm water controls, green infrastructure, urban storm water, city-level master planning

Procedia PDF Downloads 73
340 Academic Performance and Therapeutic Breathing

Authors: Abha Gupta, Seema Maira, Smita Sinha

Abstract:

This paper explores using breathing techniques to boost the academic performance of students and describes how teachers can foster the technique in their classrooms. The innovative study examines the differential impact of therapeutic breathing exercises, called pranayama, on students’ academic performance. The paper introduces approaches to therapeutic breathing exercises as an alternative to improve school performance, as well as the self-regulatory behavior, which is known to correlate with academic performance. The study was conducted in a school-wide pranayama program with positive outcomes. The intervention consisted of two breathing exercises, (1) deep breathing, and (2) alternate nostril breathing. It is a quantitative study spanning over a year with about 100 third graders was conducted using daily breathing exercises to investigate the impact of pranayama on academic performance. Significant cumulative gain-scores were found for students who practiced the approach.

Keywords: academic performance, pranayama, therapeutic breathing, yoga

Procedia PDF Downloads 461
339 Some Discrepancies between Experimentally-Based Theory of Toxic Metals Combined Action and Actual Approaches to Occupational and Environmental Health Risk Assessment and Management

Authors: Ilzira A. Minigalieva

Abstract:

Assessment of cumulative health risks associated with the widely observed combined exposures to two or more metals and their compounds on the organism in industrial or general environment, as well as respective regulatory and technical risk management decision-making have presumably the theoretical and experimental toxicology of mixtures as their reliable scientific basis. Analysis of relevant literature and our own experience proves, however, that there is no full match between these different practices. Moreover, some of the contradictions between them are of a fundamental nature. This unsatisfactory state of things may be explained not only by unavoidable simplifications characteristic of the methodologies of risk assessment and permissible exposure standards setting but also by the extreme intrinsic complexity of the combined toxicity theory, the most essential issues of which are considered and briefly discussed in this paper.

Keywords: toxic metals, nanoparticles, typology of combined toxicity, mathematical modeling, health risk assessment and management

Procedia PDF Downloads 300
338 Cash Flow Optimization on Synthetic CDOs

Authors: Timothée Bligny, Clément Codron, Antoine Estruch, Nicolas Girodet, Clément Ginet

Abstract:

Collateralized Debt Obligations are not as widely used nowadays as they were before 2007 Subprime crisis. Nonetheless there remains an enthralling challenge to optimize cash flows associated with synthetic CDOs. A Gaussian-based model is used here in which default correlation and unconditional probabilities of default are highlighted. Then numerous simulations are performed based on this model for different scenarios in order to evaluate the associated cash flows given a specific number of defaults at different periods of time. Cash flows are not solely calculated on a single bought or sold tranche but rather on a combination of bought and sold tranches. With some assumptions, the simplex algorithm gives a way to find the maximum cash flow according to correlation of defaults and maturities. The used Gaussian model is not realistic in crisis situations. Besides present system does not handle buying or selling a portion of a tranche but only the whole tranche. However the work provides the investor with relevant elements on how to know what and when to buy and sell.

Keywords: synthetic collateralized debt obligation (CDO), credit default swap (CDS), cash flow optimization, probability of default, default correlation, strategies, simulation, simplex

Procedia PDF Downloads 244
337 A Kolmogorov-Smirnov Type Goodness-Of-Fit Test of Multinomial Logistic Regression Model in Case-Control Studies

Authors: Chen Li-Ching

Abstract:

The multinomial logistic regression model is used popularly for inferring the relationship of risk factors and disease with multiple categories. This study based on the discrepancy between the nonparametric maximum likelihood estimator and semiparametric maximum likelihood estimator of the cumulative distribution function to propose a Kolmogorov-Smirnov type test statistic to assess adequacy of the multinomial logistic regression model for case-control data. A bootstrap procedure is presented to calculate the critical value of the proposed test statistic. Empirical type I error rates and powers of the test are performed by simulation studies. Some examples will be illustrated the implementation of the test.

Keywords: case-control studies, goodness-of-fit test, Kolmogorov-Smirnov test, multinomial logistic regression

Procedia PDF Downloads 427
336 Sloshing-Induced Overflow Assessment of the Seismically-Isolated Nuclear Tanks

Authors: Kihyon Kwon, Hyun T. Park, Gil Y. Chung, Sang-Hoon Lee

Abstract:

This paper focuses on assessing sloshing-induced overflow of the seismically-isolated nuclear tanks based on Fluid-Structure Interaction (FSI) analysis. Typically, fluid motion in the seismically-isolated nuclear tank systems may be rather amplified and even overflowed under earthquake. Sloshing-induced overflow in those structures has to be reliably assessed and predicted since it can often cause critical damages to humans and environments. FSI analysis is herein performed to compute the total cumulative overflowed water volume more accurately, by coupling ANSYS with CFX for structural and fluid analyses, respectively. The approach is illustrated on a nuclear liquid storage tank, Spent Fuel Pool (SFP), forgiven conditions under consideration: different liquid levels, Peak Ground Accelerations (PGAs), and post earthquakes.

Keywords: FSI analysis, seismically-isolated nuclear tank system, sloshing-induced overflow

Procedia PDF Downloads 448
335 The Influence of Design Complexity of a Building Structure on the Expected Performance

Authors: Ormal Lishi

Abstract:

This research presents a computationally efficient probabilistic method to assess the performance of compartmentation walls with similar Fire Resistance Levels (FRL) but varying complexity. Specifically, a masonry brick wall and a light-steel framed (LSF) wall with comparable insulation performance are analyzed. A Monte Carlo technique, employing Latin Hypercube Sampling (LHS), is utilized to quantify uncertainties and determine the probability of failure for both walls exposed to standard and parametric fires, following ISO 834 and Eurocodes guidelines. Results show that the probability of failure for the brick masonry wall under standard fire exposure is estimated at 4.8%, while the LSF wall is 7.6%. These probabilities decrease to 0.4% and 4.8%, respectively, when subjected to parametric fires. Notably, the complex LSF wall exhibits higher variability in predicting time to failure for specific criteria compared to the less complex brick wall, especially at higher temperatures. The proposed approach highlights the need for Probabilistic Risk Assessment (PRA) to accurately evaluate the reliability and safety levels of complex designs.

Keywords: design complexity, probability of failure, monte carlo analysis, compartmentation walls, insulation

Procedia PDF Downloads 34
334 Effect of Doping Ag and N on the Photo-Catalytic Activity of ZnO/CuO Nanocomposite for Degradation of Methyl Orange under UV and Visible Radiation

Authors: O. P. Yadav

Abstract:

Nano-size Ag-N co-doped ZnO/CuO composite photo-catalyst has been synthesized by chemical method and characterized using XRD, TEM, FTIR, AAS and UV-Vis spectroscopic techniques. Photo-catalytic activity of as-synthesized nanomaterial has been studied using degradation of methyl orange as a probe under UV as well as visible radiations. Ag-N co-doped ZnO/CuO composite showed higher photo-catalytic activity than Ag- or N-doped ZnO and undoped ZnO-CuO composite photo-catalysts. The observed highest activity of Ag-N co-doped ZnO-CuO among the studied photo-catalysts is attributed to the cumulative effects of lowering of band-gap energy and decrease of recombination rate of photo-generated electrons and holes owing to doped N and Ag, respectively. Effects of photo-catalyst load, pH and substrate initial concentration on degradation of methyl orange have also been studied. Photo-catalytic degradation of methyl orange follows pseudo first order kinetics.

Keywords: degradation, nanocomposite, photocatalyst, spectroscopy, XRD

Procedia PDF Downloads 469
333 Deepnic, A Method to Transform Each Variable into Image for Deep Learning

Authors: Nguyen J. M., Lucas G., Brunner M., Ruan S., Antonioli D.

Abstract:

Deep learning based on convolutional neural networks (CNN) is a very powerful technique for classifying information from an image. We propose a new method, DeepNic, to transform each variable of a tabular dataset into an image where each pixel represents a set of conditions that allow the variable to make an error-free prediction. The contrast of each pixel is proportional to its prediction performance and the color of each pixel corresponds to a sub-family of NICs. NICs are probabilities that depend on the number of inputs to each neuron and the range of coefficients of the inputs. Each variable can therefore be expressed as a function of a matrix of 2 vectors corresponding to an image whose pixels express predictive capabilities. Our objective is to transform each variable of tabular data into images into an image that can be analysed by CNNs, unlike other methods which use all the variables to construct an image. We analyse the NIC information of each variable and express it as a function of the number of neurons and the range of coefficients used. The predictive value and the category of the NIC are expressed by the contrast and the color of the pixel. We have developed a pipeline to implement this technology and have successfully applied it to genomic expressions on an Affymetrix chip.

Keywords: tabular data, deep learning, perfect trees, NICS

Procedia PDF Downloads 56
332 Integrated Nested Laplace Approximations For Quantile Regression

Authors: Kajingulu Malandala, Ranganai Edmore

Abstract:

The asymmetric Laplace distribution (ADL) is commonly used as the likelihood function of the Bayesian quantile regression, and it offers different families of likelihood method for quantile regression. Notwithstanding their popularity and practicality, ADL is not smooth and thus making it difficult to maximize its likelihood. Furthermore, Bayesian inference is time consuming and the selection of likelihood may mislead the inference, as the Bayes theorem does not automatically establish the posterior inference. Furthermore, ADL does not account for greater skewness and Kurtosis. This paper develops a new aspect of quantile regression approach for count data based on inverse of the cumulative density function of the Poisson, binomial and Delaporte distributions using the integrated nested Laplace Approximations. Our result validates the benefit of using the integrated nested Laplace Approximations and support the approach for count data.

Keywords: quantile regression, Delaporte distribution, count data, integrated nested Laplace approximation

Procedia PDF Downloads 135
331 Co-Smoldered Digestate Ash as Additive for Anaerobic Digestion of Berry Fruit Waste: Stability and Enhanced Production Rate

Authors: Arinze Ezieke, Antonio Serrano, William Clarke, Denys Villa-Gomez

Abstract:

Berry cultivation results in discharge of high organic strength putrescible solid waste which potentially contributes to environmental degradation, making it imperative to assess options for its complete management. Anaerobic digestion (AD) could be an ideal option when the target is energy generation; however, due to berry fruit characteristics high carbohydrate composition, the technology could be limited by its high alkalinity requirement which suggests dosing of additives such as buffers and trace elements supplement. Overcoming this limitation in an economically viable way could entail replacement of synthetic additives with recycled by-product waste. Consequently, ash from co-smouldering of high COD characteristic AD digestate and coco-coir could be a promising material to be used to enhance the AD of berry fruit waste, given its characteristic high pH, alkalinity and metal concentrations which is typical of synthetic additives. Therefore, the aim of the research was to evaluate the stability and process performance from the AD of BFW when ash from co-smoldered digestate and coir are supplemented as alkalinity and trace elements (TEs) source. Series of batch experiments were performed to ascertain the necessity for alkalinity addition and to see whether the alkalinity and metals in the co-smouldered digestate ash can provide the necessary buffer and TEs for AD of berry fruit waste. Triplicate assays were performed in batch systems following I/S of 2 (in VS), using serum bottles (160 mL) sealed and placed in a heated room (35±0.5 °C), after creating anaerobic conditions. Control experiment contained inoculum and substrates only, and inoculum, substrate and NaHCO3 for optimal total alkalinity concentration and TEs assays, respectively. Total alkalinity concentration refers to alkalinity of inoculum and the additives. The alkalinity and TE potential of the ash were evaluated by supplementing ash (22.574 g/kg) of equivalent total alkalinity concentration to that of the pre-determined optimal from NaHCO3, and by dosing ash (0.012 – 7.574 g/kg) of varying concentrations of specific essential TEs (Co, Fe, Ni, Se), respectively. The result showed a stable process at all examined conditions. Supplementation of 745 mg/L CaCO3 NaHCO3 resulted to an optimum TAC of 2000 mg/L CaCO3. Equivalent ash supplementation of 22.574 g/kg allowed the achievement of this pre-determined optimum total alkalinity concentration, resulting to a stable process with a 92% increase in the methane production rate (323 versus 168 mL CH4/ (gVS.d)), but a 36% reduction in the cumulative methane production (103 versus 161 mL CH4/gVS). Addition of ashes at incremental dosage as TEs source resulted to a reduction in the Cumulative methane production, with the highest dosage of 7.574 g/kg having the highest effect of -23.5%; however, the seemingly immediate bioavailability of TE at this high dosage allowed for a +15% increase in the methane production rate. With an increased methane production rate, the results demonstrated that the ash at high dosages could be an effective supplementary material for either a buffered or none buffered berry fruit waste AD system.

Keywords: anaerobic digestion, alkalinity, co-smoldered digestate ash, trace elements

Procedia PDF Downloads 97
330 Real-Time Network Anomaly Detection Systems Based on Machine-Learning Algorithms

Authors: Zahra Ramezanpanah, Joachim Carvallo, Aurelien Rodriguez

Abstract:

This paper aims to detect anomalies in streaming data using machine learning algorithms. In this regard, we designed two separate pipelines and evaluated the effectiveness of each separately. The first pipeline, based on supervised machine learning methods, consists of two phases. In the first phase, we trained several supervised models using the UNSW-NB15 data-set. We measured the efficiency of each using different performance metrics and selected the best model for the second phase. At the beginning of the second phase, we first, using Argus Server, sniffed a local area network. Several types of attacks were simulated and then sent the sniffed data to a running algorithm at short intervals. This algorithm can display the results of each packet of received data in real-time using the trained model. The second pipeline presented in this paper is based on unsupervised algorithms, in which a Temporal Graph Network (TGN) is used to monitor a local network. The TGN is trained to predict the probability of future states of the network based on its past behavior. Our contribution in this section is introducing an indicator to identify anomalies from these predicted probabilities.

Keywords: temporal graph network, anomaly detection, cyber security, IDS

Procedia PDF Downloads 74
329 Forecasting Models for Steel Demand Uncertainty Using Bayesian Methods

Authors: Watcharin Sangma, Onsiri Chanmuang, Pitsanu Tongkhow

Abstract:

A forecasting model for steel demand uncertainty in Thailand is proposed. It consists of trend, autocorrelation, and outliers in a hierarchical Bayesian frame work. The proposed model uses a cumulative Weibull distribution function, latent first-order autocorrelation, and binary selection, to account for trend, time-varying autocorrelation, and outliers, respectively. The Gibbs sampling Markov Chain Monte Carlo (MCMC) is used for parameter estimation. The proposed model is applied to steel demand index data in Thailand. The root mean square error (RMSE), mean absolute percentage error (MAPE), and mean absolute error (MAE) criteria are used for model comparison. The study reveals that the proposed model is more appropriate than the exponential smoothing method.

Keywords: forecasting model, steel demand uncertainty, hierarchical Bayesian framework, exponential smoothing method

Procedia PDF Downloads 322
328 An Exhaustive All-Subsets Examination of Trade Theory on WTO Data

Authors: Masoud Charkhabi

Abstract:

We examine trade theory with this motivation. The full set of World Trade Organization data are organized into country-year pairs, each treated as a different entity. Topological Data Analysis reveals that among the 16 region and 240 region-year pairs there exists in fact a distinguishable group of region-period pairs. The generally accepted periods of shifts from dissimilar-dissimilar to similar-similar trade in goods among regions are examined from this new perspective. The period breaks are treated as cumulative and are flexible. This type of all-subsets analysis is motivated from computer science and is made possible with Lossy Compression and Graph Theory. The results question many patterns in similar-similar to dissimilar-dissimilar trade. They also show indications of economic shifts that only later become evident in other economic metrics.

Keywords: econometrics, globalization, network science, topological data, analysis, trade theory, visualization, world trade

Procedia PDF Downloads 341
327 Current Status of 5A Lab6 Hollow Cathode Life Tests in Lanzhou Institute of Physics, China

Authors: Yanhui Jia, Ning Guo, Juan Li, Yunkui Sun, Wei Yang, Tianping Zhang, Lin Ma, Wei Meng, Hai Geng

Abstract:

The current statuses of lifetime test of LaB6 hollow cathode at the Lanzhou institute of physics (LIP), China, was described. 5A LaB6 hollow cathode was designed for LIPS-200 40mN Xenon ion thruster and it could be used for LHT-100 80 mN Hall thruster, too. Life test of the discharge and neutralizer modes of LHC-5 hollow cathode were stared in October 2011, and cumulative operation time reached 17,300 and 16,100 hours in April 2015, respectively. The life of cathode was designed more than 11,000 hours. Parameters of discharge and key structure dimensions were monitored in different stage of life test indicated that cathodes were health enough. The test will continue until the cathode cannot work or operation parameter is not in normally. The result of the endurance test of cathode demonstrated that the LaB6 hollow cathode is satisfied for the required of thruster in life and performance.

Keywords: LaB6, hollow cathode, thruster, lifetime test, electric propulsion

Procedia PDF Downloads 567