Search results for: hierarchical Bayesian path analysis
27902 Environmental Radioactivity Analysis by a Sequential Approach
Authors: G. Medkour Ishak-Boushaki, A. Taibi, M. Allab
Abstract:
Quantitative environmental radioactivity measurements are needed to determine the level of exposure of a population to ionizing radiations and for the assessment of the associated risks. Gamma spectrometry remains a very powerful tool for the analysis of radionuclides present in an environmental sample but the basic problem in such measurements is the low rate of detected events. Using large environmental samples could help to get around this difficulty but, unfortunately, new issues are raised by gamma rays attenuation and self-absorption. Recently, a new method has been suggested, to detect and identify without quantification, in a short time, a gamma ray of a low count source. This method does not require, as usually adopted in gamma spectrometry measurements, a pulse height spectrum acquisition. It is based on a chronological record of each detected photon by simultaneous measurements of its energy ε and its arrival time τ on the detector, the pair parameters [ε,τ] defining an event mode sequence (EMS). The EMS serials are analyzed sequentially by a Bayesian approach to detect the presence of a given radioactive source. The main object of the present work is to test the applicability of this sequential approach in radioactive environmental materials detection. Moreover, for an appropriate health oversight of the public and of the concerned workers, the analysis has been extended to get a reliable quantification of the radionuclides present in environmental samples. For illustration, we consider as an example, the problem of detection and quantification of 238U. Monte Carlo simulated experience is carried out consisting in the detection, by a Ge(Hp) semiconductor junction, of gamma rays of 63 keV emitted by 234Th (progeny of 238U). The generated EMS serials are analyzed by a Bayesian inference. The application of the sequential Bayesian approach, in environmental radioactivity analysis, offers the possibility of reducing the measurements time without requiring large environmental samples and consequently avoids the attached inconvenient. The work is still in progress.Keywords: Bayesian approach, event mode sequence, gamma spectrometry, Monte Carlo method
Procedia PDF Downloads 47027901 Recurrent Neural Networks with Deep Hierarchical Mixed Structures for Chinese Document Classification
Authors: Zhaoxin Luo, Michael Zhu
Abstract:
In natural languages, there are always complex semantic hierarchies. Obtaining the feature representation based on these complex semantic hierarchies becomes the key to the success of the model. Several RNN models have recently been proposed to use latent indicators to obtain the hierarchical structure of documents. However, the model that only uses a single-layer latent indicator cannot achieve the true hierarchical structure of the language, especially a complex language like Chinese. In this paper, we propose a deep layered model that stacks arbitrarily many RNN layers equipped with latent indicators. After using EM and training it hierarchically, our model solves the computational problem of stacking RNN layers and makes it possible to stack arbitrarily many RNN layers. Our deep hierarchical model not only achieves comparable results to large pre-trained models on the Chinese short text classification problem but also achieves state of art results on the Chinese long text classification problem.Keywords: nature language processing, recurrent neural network, hierarchical structure, document classification, Chinese
Procedia PDF Downloads 3427900 A Bayesian Model with Improved Prior in Extreme Value Problems
Authors: Eva L. Sanjuán, Jacinto Martín, M. Isabel Parra, Mario M. Pizarro
Abstract:
In Extreme Value Theory, inference estimation for the parameters of the distribution is made employing a small part of the observation values. When block maxima values are taken, many data are discarded. We developed a new Bayesian inference model to seize all the information provided by the data, introducing informative priors and using the relations between baseline and limit parameters. Firstly, we studied the accuracy of the new model for three baseline distributions that lead to a Gumbel extreme distribution: Exponential, Normal and Gumbel. Secondly, we considered mixtures of Normal variables, to simulate practical situations when data do not adjust to pure distributions, because of perturbations (noise).Keywords: bayesian inference, extreme value theory, Gumbel distribution, highly informative prior
Procedia PDF Downloads 16627899 Studying Projection Distance and Flow Properties by Shape Variations of Foam Monitor
Authors: Hyun-Kyu Cho, Jun-Su Kim, Choon-Geun Huh, Geon Lee Young-Chul Park
Abstract:
In this study, the relationship between flow properties and fluid projection distance look into connection for shape variations of foam monitor. A numerical analysis technique for fluid analysis of a foam monitor was developed for the prediction. Shape of foam monitor the flow path of fluid flow according to the shape, The fluid losses were calculated from flow analysis result.. The modified model used the length increase model of the flow path, and straight line of the model. Inlet pressure was 7 [bar] and external was atmosphere codition. am. The results showed that the length increase model of the flow path and straight line of the model was improved in the nozzle projection distance.Keywords: injection performance, finite element method, foam monitor, Projection distance
Procedia PDF Downloads 31127898 The Bayesian Premium Under Entropy Loss
Authors: Farouk Metiri, Halim Zeghdoudi, Mohamed Riad Remita
Abstract:
Credibility theory is an experience rating technique in actuarial science which can be seen as one of quantitative tools that allows the insurers to perform experience rating, that is, to adjust future premiums based on past experiences. It is used usually in automobile insurance, worker's compensation premium, and IBNR (incurred but not reported claims to the insurer) where credibility theory can be used to estimate the claim size amount. In this study, we focused on a popular tool in credibility theory which is the Bayesian premium estimator, considering Lindley distribution as a claim distribution. We derive this estimator under entropy loss which is asymmetric and squared error loss which is a symmetric loss function with informative and non-informative priors. In a purely Bayesian setting, the prior distribution represents the insurer’s prior belief about the insured’s risk level after collection of the insured’s data at the end of the period. However, the explicit form of the Bayesian premium in the case when the prior is not a member of the exponential family could be quite difficult to obtain as it involves a number of integrations which are not analytically solvable. The paper finds a solution to this problem by deriving this estimator using numerical approximation (Lindley approximation) which is one of the suitable approximation methods for solving such problems, it approaches the ratio of the integrals as a whole and produces a single numerical result. Simulation study using Monte Carlo method is then performed to evaluate this estimator and mean squared error technique is made to compare the Bayesian premium estimator under the above loss functions.Keywords: bayesian estimator, credibility theory, entropy loss, monte carlo simulation
Procedia PDF Downloads 29827897 Hierarchical Filtering Method of Threat Alerts Based on Correlation Analysis
Authors: Xudong He, Jian Wang, Jiqiang Liu, Lei Han, Yang Yu, Shaohua Lv
Abstract:
Nowadays, the threats of the internet are enormous and increasing; however, the classification of huge alert messages generated in this environment is relatively monotonous. It affects the accuracy of the network situation assessment, and also brings inconvenience to the security managers to deal with the emergency. In order to deal with potential network threats effectively and provide more effective data to improve the network situation awareness. It is essential to build a hierarchical filtering method to prevent the threats. In this paper, it establishes a model for data monitoring, which can filter systematically from the original data to get the grade of threats and be stored for using again. Firstly, it filters the vulnerable resources, open ports of host devices and services. Then use the entropy theory to calculate the performance changes of the host devices at the time of the threat occurring and filter again. At last, sort the changes of the performance value at the time of threat occurring. Use the alerts and performance data collected in the real network environment to evaluate and analyze. The comparative experimental analysis shows that the threat filtering method can effectively filter the threat alerts effectively.Keywords: correlation analysis, hierarchical filtering, multisource data, network security
Procedia PDF Downloads 17627896 Coverage Probability Analysis of WiMAX Network under Additive White Gaussian Noise and Predicted Empirical Path Loss Model
Authors: Chaudhuri Manoj Kumar Swain, Susmita Das
Abstract:
This paper explores a detailed procedure of predicting a path loss (PL) model and its application in estimating the coverage probability in a WiMAX network. For this a hybrid approach is followed in predicting an empirical PL model of a 2.65 GHz WiMAX network deployed in a suburban environment. Data collection, statistical analysis, and regression analysis are the phases of operations incorporated in this approach and the importance of each of these phases has been discussed properly. The procedure of collecting data such as received signal strength indicator (RSSI) through experimental set up is demonstrated. From the collected data set, empirical PL and RSSI models are predicted with regression technique. Furthermore, with the aid of the predicted PL model, essential parameters such as PL exponent as well as the coverage probability of the network are evaluated. This research work may assist in the process of deployment and optimisation of any cellular network significantly.Keywords: WiMAX, RSSI, path loss, coverage probability, regression analysis
Procedia PDF Downloads 14027895 Improved Dynamic Bayesian Networks Applied to Arabic On Line Characters Recognition
Authors: Redouane Tlemsani, Abdelkader Benyettou
Abstract:
Work is in on line Arabic character recognition and the principal motivation is to study the Arab manuscript with on line technology. This system is a Markovian system, which one can see as like a Dynamic Bayesian Network (DBN). One of the major interests of these systems resides in the complete models training (topology and parameters) starting from training data. Our approach is based on the dynamic Bayesian Networks formalism. The DBNs theory is a Bayesians networks generalization to the dynamic processes. Among our objective, amounts finding better parameters, which represent the links (dependences) between dynamic network variables. In applications in pattern recognition, one will carry out the fixing of the structure, which obliges us to admit some strong assumptions (for example independence between some variables). Our application will relate to the Arabic isolated characters on line recognition using our laboratory database: NOUN. A neural tester proposed for DBN external optimization. The DBN scores and DBN mixed are respectively 70.24% and 62.50%, which lets predict their further development; other approaches taking account time were considered and implemented until obtaining a significant recognition rate 94.79%.Keywords: Arabic on line character recognition, dynamic Bayesian network, pattern recognition, computer vision
Procedia PDF Downloads 40227894 Identify and Prioritize the Sustainable Development of Sports Venues Using New and Degradable Energies with a Hierarchical Analysis Approach
Authors: Mahsaossadat Pourrahmati Khelejan
Abstract:
The purpose of this research was to identify and prioritize the sustainable development of sports venues using new and degradable energies with using the AHP Hierarchical Analysis approach. The research method is a descriptive strategy with regard to the direction of implementation and is a hierarchical research with a practical purpose. In this study, 30 experts (physical education faculty members, geography professors, accredited sports venues managers, and renewable energy engineers) were selected using purposeful sampling method as the research population. The research tool was a researcher-made questionnaire on the factors affecting the sustainable development of sports venues by using new technologies and degradable energy. Finally, the research questionnaire was designed with four components and 21 items. All steps were performed by using Expert Choice software. The importance of indicators that influence the sustainable development of sports venues is highlighted by the use of clean and degradable energy, for example: 1. Economic factor, weighing 0.420 2. Environmental index, weighing 0. 320 3. Physical index, weighing 0.148 4. Social index, weighing 0.122.Keywords: Sports Venues, Sustainable Development, Degradable Energies, Prioritize
Procedia PDF Downloads 10727893 An Optimal Bayesian Maintenance Policy for a Partially Observable System Subject to Two Failure Modes
Authors: Akram Khaleghei Ghosheh Balagh, Viliam Makis, Leila Jafari
Abstract:
In this paper, we present a new maintenance model for a partially observable system subject to two failure modes, namely a catastrophic failure and a failure due to the system degradation. The system is subject to condition monitoring and the degradation process is described by a hidden Markov model. A cost-optimal Bayesian control policy is developed for maintaining the system. The control problem is formulated in the semi-Markov decision process framework. An effective computational algorithm is developed and illustrated by a numerical example.Keywords: partially observable system, hidden Markov model, competing risks, multivariate Bayesian control
Procedia PDF Downloads 42827892 Employing Bayesian Artificial Neural Network for Evaluation of Cold Rolling Force
Authors: P. Kooche Baghy, S. Eskandari, E.javanmard
Abstract:
Neural network has been used as a predictive means of cold rolling force in this dissertation. Thus, imposed average force on rollers as a mere input and five pertaining parameters to its as a outputs are regarded. According to our study, feed-forward multilayer perceptron network has been selected. Besides, Bayesian algorithm based on the feed-forward back propagation method has been selected due to noisy data. Further, 470 out of 585 all tests were used for network learning and others (115 tests) were considered as assessment criteria. Eventually, by 30 times running the MATLAB software, mean error was obtained 3.84 percent as a criteria of network learning. As a consequence, this the mentioned error on par with other approaches such as numerical and empirical methods is acceptable admittedly.Keywords: artificial neural network, Bayesian, cold rolling, force evaluation
Procedia PDF Downloads 40927891 Metacognitive Processing in Early Readers: The Role of Metacognition in Monitoring Linguistic and Non-Linguistic Performance and Regulating Students' Learning
Authors: Ioanna Taouki, Marie Lallier, David Soto
Abstract:
Metacognition refers to the capacity to reflect upon our own cognitive processes. Although there is an ongoing discussion in the literature on the role of metacognition in learning and academic achievement, little is known about its neurodevelopmental trajectories in early childhood, when children begin to receive formal education in reading. Here, we evaluate the metacognitive ability, estimated under a recently developed Signal Detection Theory model, of a cohort of children aged between 6 and 7 (N=60), who performed three two-alternative-forced-choice tasks (two linguistic: lexical decision task, visual attention span task, and one non-linguistic: emotion recognition task) including trial-by-trial confidence judgements. Our study has three aims. First, we investigated how metacognitive ability (i.e., how confidence ratings track accuracy in the task) relates to performance in general standardized tasks related to students' reading and general cognitive abilities using Spearman's and Bayesian correlation analysis. Second, we assessed whether or not young children recruit common mechanisms supporting metacognition across the different task domains or whether there is evidence for domain-specific metacognition at this early stage of development. This was done by examining correlations in metacognitive measures across different task domains and evaluating cross-task covariance by applying a hierarchical Bayesian model. Third, using robust linear regression and Bayesian regression models, we assessed whether metacognitive ability in this early stage is related to the longitudinal learning of children in a linguistic and a non-linguistic task. Notably, we did not observe any association between students’ reading skills and metacognitive processing in this early stage of reading acquisition. Some evidence consistent with domain-general metacognition was found, with significant positive correlations between metacognitive efficiency between lexical and emotion recognition tasks and substantial covariance indicated by the Bayesian model. However, no reliable correlations were found between metacognitive performance in the visual attention span and the remaining tasks. Remarkably, metacognitive ability significantly predicted children's learning in linguistic and non-linguistic domains a year later. These results suggest that metacognitive skill may be dissociated to some extent from general (i.e., language and attention) abilities and further stress the importance of creating educational programs that foster students’ metacognitive ability as a tool for long term learning. More research is crucial to understand whether these programs can enhance metacognitive ability as a transferable skill across distinct domains or whether unique domains should be targeted separately.Keywords: confidence ratings, development, metacognitive efficiency, reading acquisition
Procedia PDF Downloads 11827890 Path Planning for Orchard Robot Using Occupancy Grid Map in 2D Environment
Authors: Satyam Raikwar, Thomas Herlitzius, Jens Fehrmann
Abstract:
In recent years, the autonomous navigation of orchard and field robots is an emerging technology of the mobile robotics in agriculture. One of the core aspects of autonomous navigation builds upon path planning, which is still a crucial issue. Generally, for simple representation, the path planning for a mobile robot is performed in a two-dimensional space, which creates a path between the start and goal point. This paper presents the automatic path planning approach for robots used in orchards and vineyards using occupancy grid maps with field consideration. The orchards and vineyards are usually structured environment and their topology is assumed to be constant over time; therefore, in this approach, an RGB image of a field is used as a working environment. These images undergone different image processing operations and then discretized into two-dimensional grid matrices. The individual grid or cell of these grid matrices represents the occupancy of the space, whether it is free or occupied. The grid matrix represents the robot workspace for motion and path planning. After the grid matrix is described, a probabilistic roadmap (PRM) path algorithm is used to create the obstacle-free path over these occupancy grids. The path created by this method was successfully verified in the test area. Furthermore, this approach is used in the navigation of the orchard robot.Keywords: orchard robots, automatic path planning, occupancy grid, probabilistic roadmap
Procedia PDF Downloads 13027889 Simulating the Hot Hand Phenomenon in Basketball with Bayesian Hidden Markov Models
Authors: Gabriel Calvo, Carmen Armero, Luigi Spezia
Abstract:
A basketball player is said to have a hot hand if his/her performance is better than expected in different periods of time. A way to deal with this phenomenon is to make use of latent variables, which can indicate whether the player is ‘on fire’ or not. This work aims to model the hot hand phenomenon through a Bayesian hidden Markov model (HMM) with two states (cold and hot) and two different probability of success depending on the corresponding hidden state. This task is illustrated through a comprehensive simulation study. The simulated data sets emulate the field goal attempts in an NBA season from different profile players. This model can be a powerful tool to assess the ‘streakiness’ of each player, and it provides information about the general performance of the players during the match. Finally, the Bayesian HMM allows computing the posterior probability of any type of streak.Keywords: Bernoulli trials, field goals, latent variables, posterior distribution
Procedia PDF Downloads 13827888 Modeling of System Availability and Bayesian Analysis of Bivariate Distribution
Authors: Muhammad Farooq, Ahtasham Gul
Abstract:
To meet the desired standard, it is important to monitor and analyze different engineering processes to get desired output. The bivariate distributions got a lot of attention in recent years to describe the randomness of natural as well as artificial mechanisms. In this article, a bivariate model is constructed using two independent models developed by the nesting approach to study the effect of each component on reliability for better understanding. Further, the Bayes analysis of system availability is studied by considering prior parametric variations in the failure time and repair time distributions. Basic statistical characteristics of marginal distribution, like mean median and quantile function, are discussed. We use inverse Gamma prior to study its frequentist properties by conducting Monte Carlo Markov Chain (MCMC) sampling scheme.Keywords: reliability, system availability Weibull, inverse Lomax, Monte Carlo Markov Chain, Bayesian
Procedia PDF Downloads 4827887 Measurement Errors and Misclassifications in Covariates in Logistic Regression: Bayesian Adjustment of Main and Interaction Effects and the Sample Size Implications
Authors: Shahadut Hossain
Abstract:
Measurement errors in continuous covariates and/or misclassifications in categorical covariates are common in epidemiological studies. Regression analysis ignoring such mismeasurements seriously biases the estimated main and interaction effects of covariates on the outcome of interest. Thus, adjustments for such mismeasurements are necessary. In this research, we propose a Bayesian parametric framework for eliminating deleterious impacts of covariate mismeasurements in logistic regression. The proposed adjustment method is unified and thus can be applied to any generalized linear and non-linear regression models. Furthermore, adjustment for covariate mismeasurements requires validation data usually in the form of either gold standard measurements or replicates of the mismeasured covariates on a subset of the study population. Initial investigation shows that adequacy of such adjustment depends on the sizes of main and validation samples, especially when prevalences of the categorical covariates are low. Thus, we investigate the impact of main and validation sample sizes on the adjusted estimates, and provide a general guideline about these sample sizes based on simulation studies.Keywords: measurement errors, misclassification, mismeasurement, validation sample, Bayesian adjustment
Procedia PDF Downloads 38427886 Bayesian Inference for High Dimensional Dynamic Spatio-Temporal Models
Authors: Sofia M. Karadimitriou, Kostas Triantafyllopoulos, Timothy Heaton
Abstract:
Reduced dimension Dynamic Spatio-Temporal Models (DSTMs) jointly describe the spatial and temporal evolution of a function observed subject to noise. A basic state space model is adopted for the discrete temporal variation, while a continuous autoregressive structure describes the continuous spatial evolution. Application of such a DSTM relies upon the pre-selection of a suitable reduced set of basic functions and this can present a challenge in practice. In this talk, we propose an online estimation method for high dimensional spatio-temporal data based upon DSTM and we attempt to resolve this issue by allowing the basis to adapt to the observed data. Specifically, we present a wavelet decomposition in order to obtain a parsimonious approximation of the spatial continuous process. This parsimony can be achieved by placing a Laplace prior distribution on the wavelet coefficients. The aim of using the Laplace prior, is to filter wavelet coefficients with low contribution, and thus achieve the dimension reduction with significant computation savings. We then propose a Hierarchical Bayesian State Space model, for the estimation of which we offer an appropriate particle filter algorithm. The proposed methodology is illustrated using real environmental data.Keywords: multidimensional Laplace prior, particle filtering, spatio-temporal modelling, wavelets
Procedia PDF Downloads 40027885 Analyzing Energy Consumption Behavior of Migrated Population in Turkey Using Bayesian Belief Approach
Authors: Ebru Acuner, Gulgun Kayakutlu, M. Ozgur Kayalica, Sermin Onaygil
Abstract:
In Turkey, emigration, especially from Syria, has been continuously increasing together with rapid urbanization. In parallel to this, total energy consumption has been growing, rapidly. Unfortunately, domestic energy sources could not meet this energy demand. Hence, there is a need for reliable predictions. For this reason, before making a survey study for the migrated people, an informative questionnaire was prepared to take the opinions of the experts on the main drivers that shape the energy consumption behavior of the migrated people. Totally, 17 experts were answered, and they were analyzed by means of Netica program considering Bayesian belief analysis method. In the analysis, factors affecting energy consumption behaviors as well as strategies, institutions, tools and financing methods to change these behaviors towards efficient consumption were investigated. On the basis of the results, it can be concluded that changing the energy consumption behavior of the migrated people is crucial. In order to be successful, electricity and natural gas prices and tariffs in the market should be arranged considering energy efficiency. In addition, support mechanisms by not only the government but also municipalities should be taken into account while preparing related policies. Also, electric appliance producers should develop and implement strategies and action in favor of the usage of more efficient appliances. Last but not least, non-governmental organizations should support the migrated people to improve their awareness on the efficient consumption for the sustainable future.Keywords: Bayesian belief, behavior, energy consumption, energy efficiency, migrated people
Procedia PDF Downloads 8327884 Bayesian Using Markov Chain Monte Carlo and Lindley's Approximation Based on Type-I Censored Data
Authors: Al Omari Moahmmed Ahmed
Abstract:
These papers describe the Bayesian Estimator using Markov Chain Monte Carlo and Lindley’s approximation and the maximum likelihood estimation of the Weibull distribution with Type-I censored data. The maximum likelihood method can’t estimate the shape parameter in closed forms, although it can be solved by numerical methods. Moreover, the Bayesian estimates of the parameters, the survival and hazard functions cannot be solved analytically. Hence Markov Chain Monte Carlo method and Lindley’s approximation are used, where the full conditional distribution for the parameters of Weibull distribution are obtained via Gibbs sampling and Metropolis-Hastings algorithm (HM) followed by estimate the survival and hazard functions. The methods are compared to Maximum Likelihood counterparts and the comparisons are made with respect to the Mean Square Error (MSE) and absolute bias to determine the better method in scale and shape parameters, the survival and hazard functions.Keywords: weibull distribution, bayesian method, markov chain mote carlo, survival and hazard functions
Procedia PDF Downloads 44927883 Copula-Based Estimation of Direct and Indirect Effects in Path Analysis Model
Authors: Alam Ali, Ashok Kumar Pathak
Abstract:
Path analysis is a statistical technique used to evaluate the strength of the direct and indirect effects of variables. One or more structural regression equations are used to estimate a series of parameters in order to find the better fit of data. Sometimes, exogenous variables do not show a significant strength of their direct and indirect effect when the assumption of classical regression (ordinary least squares (OLS)) are violated by the nature of the data. The main motive of this article is to investigate the efficacy of the copula-based regression approach over the classical regression approach and calculate the direct and indirect effects of variables when data violates the OLS assumption and variables are linked through an elliptical copula. We perform this study using a well-organized numerical scheme. Finally, a real data application is also presented to demonstrate the performance of the superiority of the copula approach.Keywords: path analysis, copula-based regression models, direct and indirect effects, k-fold cross validation technique
Procedia PDF Downloads 4627882 Facile Hydrothermal Synthesis of Hierarchical NiO/ZnCo₂O₄ Nanocomposite for High-Energy Supercapacitor Applications
Authors: Fayssal Ynineb, Toufik Hadjersi, Fatsah Moulai, Wafa Achour
Abstract:
Currently, tremendous attention has been paid to the rational design and synthesis of core/shell heterostructures for high-performance supercapacitors. In this study, the hierarchical NiO/ZnCo₂O₄ Core-Shell Nanorods Arrays were successfully deposited onto ITO substrate via a two-step hydrothermal and electrodeposition methods. The effect of the thin carbon layer between NiO and ZnCo₂O₄ in this multi-scale hierarchical structure was investigated. The selection of this structure was based on: (i) a high specific area of pseudo-capacitive NiO to maximize specific capacitance; (ii) an effective NiO-electrolyte interface to facilitate fast charging/discharging; and (iii) conducting carbon layer between ZnCo₂O₄ and NiO enhance the electric conductivity which reduces energy loss, and the corrosion protection of ZnCo₂O₄ in alkaline electrolyte. The obtained results indicate that hierarchical NiO/ZnCo₂O₄ present a high specific capacitance of 63 mF.cm⁻² at a current density of 0.05 mA.cm⁻² higher than that of pristine NiO and ZnCo₂O₄ of 6 and 3 mF.cm⁻², respectively. The carbon layer improves the electrical conductivity among NiO and ZnCo₂O₄ in the hierarchical NiO/C/ZnCo₂O₄ electrode. As well, the specific capacitance drastically increased to reach 125 mF.cm⁻². Moreover, this multi-scale hierarchical structure exhibits superior cycling stability with ~ 95.7 % capacitance retention after 65k cycles. These results indicate that the NiO/C/ZnCo₂O₄ nanocomposite material is an outstanding electrode material for supercapacitors.Keywords: NiO/C/ZnCo₂O₄, specific capacitance, hydrothermal, supercapacitors
Procedia PDF Downloads 7127881 Independence and Path Independence on Cayley Digraphs of Left Groups and Right Groups
Authors: Nuttawoot Nupo, Sayan Panma
Abstract:
A semigroup S is said to be a left (right) zero semigroup if S satisfies the equation xy=x (xy=y) for all x,y in S. In addition, the semigroup S is called a left (right) group if S is isomorphic to the direct product of a group and a left (right) zero semigroup. The Cayley digraph Cay(S,A) of a semigroup S with a connection set A is defined to be a digraph with the vertex set S and the arc set E(Cay(S,A))={(x,xa) | x∈S, a∈A} where A is any subset of S. All sets in this research are assumed to be finite. Let D be a digraph together with a vertex set V and an arc set E. Let u and v be two different vertices in V and I a nonempty subset of V. The vertices u and v are said to be independent if (u,v)∉E and (v,u)∉E. The set I is called an independent set of D if any two different vertices in I are independent. The independence number of D is the maximum cardinality of an independent set of D. Moreover, the vertices u and v are said to be path independent if there is no dipath from u to v and there is no dipath from v to u. The set I is called a path independent set of D if any two different vertices in I are path independent. The path independence number of D is the maximum cardinality of a path independent set of D. In this research, we describe a lower bound and an upper bound of the independence number of Cayley digraphs of left groups and right groups. Some examples corresponding to those bounds are illustrated here. Furthermore, the exact value of the path independence number of Cayley digraphs of left groups and right groups are also presented.Keywords: Cayley digraphs, independence number, left groups, path independence number, right groups
Procedia PDF Downloads 21027880 Exploring the Activity Fabric of an Intelligent Environment with Hierarchical Hidden Markov Theory
Authors: Chiung-Hui Chen
Abstract:
The Internet of Things (IoT) was designed for widespread convenience. With the smart tag and the sensing network, a large quantity of dynamic information is immediately presented in the IoT. Through the internal communication and interaction, meaningful objects provide real-time services for users. Therefore, the service with appropriate decision-making has become an essential issue. Based on the science of human behavior, this study employed the environment model to record the time sequences and locations of different behaviors and adopted the probability module of the hierarchical Hidden Markov Model for the inference. The statistical analysis was conducted to achieve the following objectives: First, define user behaviors and predict the user behavior routes with the environment model to analyze user purposes. Second, construct the hierarchical Hidden Markov Model according to the logic framework, and establish the sequential intensity among behaviors to get acquainted with the use and activity fabric of the intelligent environment. Third, establish the intensity of the relation between the probability of objects’ being used and the objects. The indicator can describe the possible limitations of the mechanism. As the process is recorded in the information of the system created in this study, these data can be reused to adjust the procedure of intelligent design services.Keywords: behavior, big data, hierarchical hidden Markov model, intelligent object
Procedia PDF Downloads 20527879 Hierarchical Porous Carbon Composite Electrode for High Performance Supercapacitor Application
Authors: Chia-Chia Chang, Jhen-Ting Huang, Hu-Cheng Weng, An-Ya Lo
Abstract:
This study developed a simple hierarchical porous carbon (HPC) synthesis process and used for supercapacitor application. In which, mesopore provides huge specific surface area, meanwhile, macropore provides excellent mass transfer. Thus the hierarchical porous electrode improves the charge-discharge performance. On the other hand, cerium oxide (CeO2) have also got a lot research attention owing to its rich in content, low in price, environmentally friendly, good catalytic properties, and easy preparation. Besides, a rapid redox reaction occurs between trivalent cerium and tetravalent cerium releases oxygen atom and increase the conductivity. In order to prevent CeO2 from disintegration under long-term charge-discharge operation, the CeO2 carbon porous materials were was integrated as composite material in this study. For in the ex-situ analysis, scanning electron microscope (SEM), X-ray diffraction (XRD), transmission electron microscope (TEM) analysis were adopted to identify the surface morphology, crystal structure, and microstructure of the composite. 77K Nitrogen adsorption-desorption analysis was used to analyze the porosity of each specimen. For the in-situ test, cyclic voltammetry (CV) and chronopotentiometry (CP) were conducted by potentiostat to understand the charge and discharge properties. Ragone plot was drawn to further analyze the resistance properties. Based on above analyses, the effect of macropores/mespores and the CeO2/HPC ratios on charge-discharge performance were investigated. As a result, the capacitance can be greatly enhanced by 2.6 times higher than pristine mesoporous carbon electrode.Keywords: hierarchical porous carbon, cerium oxide, supercapacitor
Procedia PDF Downloads 9627878 Campus Living Environments that Contribute to Mental Health: A Path Analysis Based on Environmental Characteristics
Authors: Jing Ren, Guifeng Han
Abstract:
The mental health of most college students in China is negative due to the multiple pressures of academics, life, and employment. The problem of psychological stress has been widely discussed and needs to be resolved immediately. Therefore, six typical green spaces in Chongqing University, China, were selected to explore the relationship between eight environmental characteristics and students' stress relief. A path analysis model is established using Amos26.0 to explain the paths for environmental characteristics influencing psychological stress relief. The results show that (1) tree species diversity (TSD) has a positive effect on stress relief, thus green coverage ratio (GCR), the proportion of water area (WAP), visual green index (VGI), and color richness (CR) have both positive and negative effects; (2) CR could reduce stress directly and indirectly, while GCR, TSD, WAP, and VGI could only reduce stress indirectly, and the most effective path is TSD→extent→stress relief; (3) CR can reduce stress more greatly for males than females, CR and VGI have better effects for art students than science students. The study can provide a theoretical reference for planning and designing campus living environments to improve students' mental health.Keywords: public health, residential environment, space planning and management, mental health, path analysis
Procedia PDF Downloads 3727877 Optimizing Network Latency with Fast Path Assignment for Incoming Flows
Abstract:
Various flows in the network require to go through different types of middlebox. The improper placement of network middlebox and path assignment for flows could greatly increase the network latency and also decrease the performance of network. Minimizing the total end to end latency of all the ows requires to assign path for the incoming flows. In this paper, the flow path assignment problem in regard to the placement of various kinds of middlebox is studied. The flow path assignment problem is formulated to a linear programming problem, which is very time consuming. On the other hand, a naive greedy algorithm is studied. Which is very fast but causes much more latency than the linear programming algorithm. At last, the paper presents a heuristic algorithm named FPA, which takes bottleneck link information and estimated bandwidth occupancy into consideration, and achieves near optimal latency in much less time. Evaluation results validate the effectiveness of the proposed algorithm.Keywords: flow path, latency, middlebox, network
Procedia PDF Downloads 17927876 Eliminating Cutter-Path Deviation For Five-Axis Nc Machining
Authors: Alan C. Lin, Tsong Der Lin
Abstract:
This study proposes a deviation control method to add interpolation points to numerical control (NC) codes of five-axis machining in order to achieve the required machining accuracy. Specific research issues include: (1) converting machining data between the CL (cutter location) domain and the NC domain, (2) calculating the deviation between the deviated path and the linear path, (3) finding interpolation points, and (4) determining tool orientations for the interpolation points. System implementation with practical examples will also be included to highlight the applicability of the proposed methodology.Keywords: CAD/CAM, cutter path, five-axis machining, numerical control
Procedia PDF Downloads 39927875 The Use of Layered Neural Networks for Classifying Hierarchical Scientific Fields of Study
Authors: Colin Smith, Linsey S Passarella
Abstract:
Due to the proliferation and decentralized nature of academic publication, no widely accepted scheme exists for organizing papers by their scientific field of study (FoS) to the author’s best knowledge. While many academic journals require author provided keywords for papers, these keywords range wildly in scope and are not consistent across papers, journals, or field domains, necessitating alternative approaches to paper classification. Past attempts to perform field-of-study (FoS) classification on scientific texts have largely used a-hierarchical FoS schemas or ignored the schema’s inherently hierarchical structure, e.g. by compressing the structure into a single layer for multi-label classification. In this paper, we introduce an application of a Layered Neural Network (LNN) to the problem of performing supervised hierarchical classification of scientific fields of study (FoS) on research papers. In this approach, paper embeddings from a pretrained language model are fed into a top-down LNN. Beginning with a single neural network (NN) for the highest layer of the class hierarchy, each node uses a separate local NN to classify the subsequent subfield child node(s) for an input embedding of concatenated paper titles and abstracts. We compare our LNN-FOS method to other recent machine learning methods using the Microsoft Academic Graph (MAG) FoS hierarchy and find that the LNN-FOS offers increased classification accuracy at each FoS hierarchical level.Keywords: hierarchical classification, layer neural network, scientific field of study, scientific taxonomy
Procedia PDF Downloads 10027874 A Bayesian Network Approach to Customer Loyalty Analysis: A Case Study of Home Appliances Industry in Iran
Authors: Azam Abkhiz, Abolghasem Nasir
Abstract:
To achieve sustainable competitive advantage in the market, it is necessary to provide and improve customer satisfaction and Loyalty. To reach this objective, companies need to identify and analyze their customers. Thus, it is critical to measure the level of customer satisfaction and Loyalty very carefully. This study attempts to build a conceptual model to provide clear insights of customer loyalty. Using Bayesian networks (BNs), a model is proposed to evaluate customer loyalty and its consequences, such as repurchase and positive word-of-mouth. BN is a probabilistic approach that predicts the behavior of a system based on observed stochastic events. The most relevant determinants of customer loyalty are identified by the literature review. Perceived value, service quality, trust, corporate image, satisfaction, and switching costs are the most important variables that explain customer loyalty. The data are collected by use of a questionnaire-based survey from 1430 customers of a home appliances manufacturer in Iran. Four scenarios and sensitivity analyses are performed to run and analyze the impact of different determinants on customer loyalty. The proposed model allows businesses to not only set their targets but proactively manage their customer behaviors as well.Keywords: customer satisfaction, customer loyalty, Bayesian networks, home appliances industry
Procedia PDF Downloads 10427873 Case-Based Reasoning for Modelling Random Variables in the Reliability Assessment of Existing Structures
Authors: Francesca Marsili
Abstract:
The reliability assessment of existing structures with probabilistic methods is becoming an increasingly important and frequent engineering task. However probabilistic reliability methods are based on an exhaustive knowledge of the stochastic modeling of the variables involved in the assessment; at the moment standards for the modeling of variables are absent, representing an obstacle to the dissemination of probabilistic methods. The framework according to probability distribution functions (PDFs) are established is represented by the Bayesian statistics, which uses Bayes Theorem: a prior PDF for the considered parameter is established based on information derived from the design stage and qualitative judgments based on the engineer past experience; then, the prior model is updated with the results of investigation carried out on the considered structure, such as material testing, determination of action and structural properties. The application of Bayesian statistics arises two different kind of problems: 1. The results of the updating depend on the engineer previous experience; 2. The updating of the prior PDF can be performed only if the structure has been tested, and quantitative data that can be statistically manipulated have been collected; performing tests is always an expensive and time consuming operation; furthermore, if the considered structure is an ancient building, destructive tests could compromise its cultural value and therefore should be avoided. In order to solve those problems, an interesting research path is represented by investigating Artificial Intelligence (AI) techniques that can be useful for the automation of the modeling of variables and for the updating of material parameters without performing destructive tests. Among the others, one that raises particular attention in relation to the object of this study is constituted by Case-Based Reasoning (CBR). In this application, cases will be represented by existing buildings where material tests have already been carried out and an updated PDFs for the material mechanical parameters has been computed through a Bayesian analysis. Then each case will be composed by a qualitative description of the material under assessment and the posterior PDFs that describe its material properties. The problem that will be solved is the definition of PDFs for material parameters involved in the reliability assessment of the considered structure. A CBR system represent a good candi¬date in automating the modelling of variables because: 1. Engineers already draw an estimation of the material properties based on the experience collected during the assessment of similar structures, or based on similar cases collected in literature or in data-bases; 2. Material tests carried out on structure can be easily collected from laboratory database or from literature; 3. The system will provide the user of a reliable probabilistic description of the variables involved in the assessment that will also serve as a tool in support of the engineer’s qualitative judgments. Automated modeling of variables can help in spreading probabilistic reliability assessment of existing buildings in the common engineering practice, and target at the best intervention and further tests on the structure; CBR represents a technique which may help to achieve this.Keywords: reliability assessment of existing buildings, Bayesian analysis, case-based reasoning, historical structures
Procedia PDF Downloads 313