World Academy of Science, Engineering and Technology
[Mathematical and Computational Sciences]
Online ISSN : 1307-6892
956 Comparison of Methods of Estimation for Use in Goodness of Fit Tests for Binary Multilevel Models
Authors: I. V. Pinto, M. R. Sooriyarachchi
Abstract:
It can be frequently observed that the data arising in our environment have a hierarchical or a nested structure attached with the data. Multilevel modelling is a modern approach to handle this kind of data. When multilevel modelling is combined with a binary response, the estimation methods get complex in nature and the usual techniques are derived from quasi-likelihood method. The estimation methods which are compared in this study are, marginal quasi-likelihood (order 1 & order 2) (MQL1, MQL2) and penalized quasi-likelihood (order 1 & order 2) (PQL1, PQL2). A statistical model is of no use if it does not reflect the given dataset. Therefore, checking the adequacy of the fitted model through a goodness-of-fit (GOF) test is an essential stage in any modelling procedure. However, prior to usage, it is also equally important to confirm that the GOF test performs well and is suitable for the given model. This study assesses the suitability of the GOF test developed for binary response multilevel models with respect to the method used in model estimation. An extensive set of simulations was conducted using MLwiN (v 2.19) with varying number of clusters, cluster sizes and intra cluster correlations. The test maintained the desirable Type-I error for models estimated using PQL2 and it failed for almost all the combinations of MQL. Power of the test was adequate for most of the combinations in all estimation methods except MQL1. Moreover, models were fitted using the four methods to a real-life dataset and performance of the test was compared for each model.Keywords: goodness-of-fit test, marginal quasi-likelihood, multilevel modelling, penalized quasi-likelihood, power, quasi-likelihood, type-I error
Procedia PDF Downloads 142955 Extensions of Schwarz Lemma in the Half-Plane
Authors: Nicolae Pascu
Abstract:
Aside from being a fundamental tool in Complex analysis, Schwarz Lemma-which was finalized in its most complete form at the beginning of the last century-generated an important area of research in various fields of mathematics, which continues to advance even today. We present some properties of analytic functions in the half-plane which satisfy the conditions of the classical Schwarz Lemma (Carathéodory functions) and obtain a generalization of the well-known Aleksandrov-Sobolev Lemma for analytic functions in the half-plane (the correspondent of Schwarz-Pick Lemma from the unit disk). Using this Schwarz-type lemma, we obtain a characterization for the entire class of Carathéodory functions, which might be of independent interest. We prove two monotonicity properties for Carathéodory functions that do not depend upon their normalization at infinity (the hydrodynamic normalization). The method is based on conformal mapping arguments for analytic functions in the half-plane satisfying appropriate conditions, in the spirit of Schwarz lemma. According to the research findings in this paper, our main results give estimates for the modulus and the argument for the entire class of Carathéodory functions. As applications, we give several extensions of Julia-Wolf-Carathéodory Lemma in a half-strip and show that our results are sharp.Keywords: schwarz lemma, Julia-wolf-caratéodory lemma, analytic function, normalization condition, caratéodory function
Procedia PDF Downloads 218954 Regression for Doubly Inflated Multivariate Poisson Distributions
Authors: Ishapathik Das, Sumen Sen, N. Rao Chaganty, Pooja Sengupta
Abstract:
Dependent multivariate count data occur in several research studies. These data can be modeled by a multivariate Poisson or Negative binomial distribution constructed using copulas. However, when some of the counts are inflated, that is, the number of observations in some cells are much larger than other cells, then the copula based multivariate Poisson (or Negative binomial) distribution may not fit well and it is not an appropriate statistical model for the data. There is a need to modify or adjust the multivariate distribution to account for the inflated frequencies. In this article, we consider the situation where the frequencies of two cells are higher compared to the other cells, and develop a doubly inflated multivariate Poisson distribution function using multivariate Gaussian copula. We also discuss procedures for regression on covariates for the doubly inflated multivariate count data. For illustrating the proposed methodologies, we present a real data containing bivariate count observations with inflations in two cells. Several models and linear predictors with log link functions are considered, and we discuss maximum likelihood estimation to estimate unknown parameters of the models.Keywords: copula, Gaussian copula, multivariate distributions, inflated distributios
Procedia PDF Downloads 156953 A Stepwise Approach to Automate the Search for Optimal Parameters in Seasonal ARIMA Models
Authors: Manisha Mukherjee, Diptarka Saha
Abstract:
Reliable forecasts of univariate time series data are often necessary for several contexts. ARIMA models are quite popular among practitioners in this regard. Hence, choosing correct parameter values for ARIMA is a challenging yet imperative task. Thus, a stepwise algorithm is introduced to provide automatic and robust estimates for parameters (p; d; q)(P; D; Q) used in seasonal ARIMA models. This process is focused on improvising the overall quality of the estimates, and it alleviates the problems induced due to the unidimensional nature of the methods that are currently used such as auto.arima. The fast and automated search of parameter space also ensures reliable estimates of the parameters that possess several desirable qualities, consequently, resulting in higher test accuracy especially in the cases of noisy data. After vigorous testing on real as well as simulated data, the algorithm doesn’t only perform better than current state-of-the-art methods, it also completely obviates the need for human intervention due to its automated nature.Keywords: time series, ARIMA, auto.arima, ARIMA parameters, forecast, R function
Procedia PDF Downloads 165952 The Influence of Beta Shape Parameters in Project Planning
Authors: Αlexios Kotsakis, Stefanos Katsavounis, Dimitra Alexiou
Abstract:
Networks can be utilized to represent project planning problems, using nodes for activities and arcs to indicate precedence relationship between them. For fixed activity duration, a simple algorithm calculates the amount of time required to complete a project, followed by the activities that comprise the critical path. Program Evaluation and Review Technique (PERT) generalizes the above model by incorporating uncertainty, allowing activity durations to be random variables, producing nevertheless a relatively crude solution in planning problems. In this paper, based on the findings of the relevant literature, which strongly suggests that a Beta distribution can be employed to model earthmoving activities, we utilize Monte Carlo simulation, to estimate the project completion time distribution and measure the influence of skewness, an element inherent in activities of modern technical projects. We also extract the activity criticality index, with an ultimate goal to produce more accurate planning estimations.Keywords: beta distribution, PERT, Monte Carlo simulation, skewness, project completion time distribution
Procedia PDF Downloads 149951 An Application of Modified M-out-of-N Bootstrap Method to Heavy-Tailed Distributions
Authors: Hannah F. Opayinka, Adedayo A. Adepoju
Abstract:
This study is an extension of a prior study on the modification of the existing m-out-of-n (moon) bootstrap method for heavy-tailed distributions in which modified m-out-of-n (mmoon) was proposed as an alternative method to the existing moon technique. In this study, both moon and mmoon techniques were applied to two real income datasets which followed Lognormal and Pareto distributions respectively with finite variances. The performances of these two techniques were compared using Standard Error (SE) and Root Mean Square Error (RMSE). The findings showed that mmoon outperformed moon bootstrap in terms of smaller SEs and RMSEs for all the sample sizes considered in the two datasets.Keywords: Bootstrap, income data, lognormal distribution, Pareto distribution
Procedia PDF Downloads 186950 Effects of Educational Technology Integration in Classroom Instruction to the Math Performance of Generation Z Students of a Private High School in the Philippines
Authors: May Maricel De Gracia
Abstract:
Different generations respond differently to instruction because of their diverse characteristics, learning styles and study habits. Teaching strategies that were effective many years ago may not be effective now especially to the current generation which is Gen Z. Using quantitative research design, the main goal of this paper is to determine the impact of the implementation of educational technology integration in a private high school in the math performance of its Junior High School (JHS) students on SY 2014-2018 based on their periodical exam performance and on their final math grades. In support, survey on the use of technology was administered to determine the characteristics of both students and teachers of SY 2017-2018. Another survey regarding study habits was also administered to the students to determine their readiness with regards to note-taking skills, time management, test taking/preparation skills, reading, and writing and math skills. Teaching strategies were recommended based on the need of the current Gen Z JHS students. A total of 712 JHS students and 12 math teachers participated in answering the different surveys. Periodic exam means and final math grades between the school years without technology (SY 2004-2008) and with technology (SY 2014-2018) were analyzed through correlation and regression analyses. Result shows that the periodic exam mean has a 35.29% impact to the final grade of the students. In addition, z-test result where p > 0.05 shows that the periodical exam results do not differ significantly between the school years without integration of technology and with the integration of technology. However, with p < 0.01, a significant positive difference was observed in the final math grades of students between the school years without technology integration and with technology integration.Keywords: classroom instruction, technology, generation z, math performance
Procedia PDF Downloads 147949 Forecasting Issues in Energy Markets within a Reg-ARIMA Framework
Authors: Ilaria Lucrezia Amerise
Abstract:
Electricity markets throughout the world have undergone substantial changes. Accurate, reliable, clear and comprehensible modeling and forecasting of different variables (loads and prices in the first instance) have achieved increasing importance. In this paper, we describe the actual state of the art focusing on reg-SARMA methods, which have proven to be flexible enough to accommodate the electricity price/load behavior satisfactory. More specifically, we will discuss: 1) The dichotomy between point and interval forecasts; 2) The difficult choice between stochastic (e.g. climatic variation) and non-deterministic predictors (e.g. calendar variables); 3) The confrontation between modelling a single aggregate time series or creating separated and potentially different models of sub-series. The noteworthy point that we would like to make it emerge is that prices and loads require different approaches that appear irreconcilable even though must be made reconcilable for the interests and activities of energy companies.Keywords: interval forecasts, time series, electricity prices, reg-SARIMA methods
Procedia PDF Downloads 131948 Comparison of Receiver Operating Characteristic Curve Smoothing Methods
Authors: D. Sigirli
Abstract:
The Receiver Operating Characteristic (ROC) curve is a commonly used statistical tool for evaluating the diagnostic performance of screening and diagnostic test with continuous or ordinal scale results which aims to predict the presence or absence probability of a condition, usually a disease. When the test results were measured as numeric values, sensitivity and specificity can be computed across all possible threshold values which discriminate the subjects as diseased and non-diseased. There are infinite numbers of possible decision thresholds along the continuum of the test results. The ROC curve presents the trade-off between sensitivity and the 1-specificity as the threshold changes. The empirical ROC curve which is a non-parametric estimator of the ROC curve is robust and it represents data accurately. However, especially for small sample sizes, it has a problem of variability and as it is a step function there can be different false positive rates for a true positive rate value and vice versa. Besides, the estimated ROC curve being in a jagged form, since the true ROC curve is a smooth curve, it underestimates the true ROC curve. Since the true ROC curve is assumed to be smooth, several smoothing methods have been explored to smooth a ROC curve. These include using kernel estimates, using log-concave densities, to fit parameters for the specified density function to the data with the maximum-likelihood fitting of univariate distributions or to create a probability distribution by fitting the specified distribution to the data nd using smooth versions of the empirical distribution functions. In the present paper, we aimed to propose a smooth ROC curve estimation based on the boundary corrected kernel function and to compare the performances of ROC curve smoothing methods for the diagnostic test results coming from different distributions in different sample sizes. We performed simulation study to compare the performances of different methods for different scenarios with 1000 repetitions. It is seen that the performance of the proposed method was typically better than that of the empirical ROC curve and only slightly worse compared to the binormal model when in fact the underlying samples were generated from the normal distribution.Keywords: empirical estimator, kernel function, smoothing, receiver operating characteristic curve
Procedia PDF Downloads 152947 The Permutation of Symmetric Triangular Equilateral Group in the Cryptography of Private and Public Key
Authors: Fola John Adeyeye
Abstract:
In this paper, we propose a cryptosystem private and public key base on symmetric group Pn and validates its theoretical formulation. This proposed system benefits from the algebraic properties of Pn such as noncommutative high logical, computational speed and high flexibility in selecting key which makes the discrete permutation multiplier logic (DPML) resist to attack by any algorithm such as Pohlig-Hellman. One of the advantages of this scheme is that it explore all the possible triangular symmetries. Against these properties, the only disadvantage is that the law of permutation multiplicity only allow an operation from left to right. Many other cryptosystems can be transformed into their symmetric group.Keywords: cryptosystem, private and public key, DPML, symmetric group Pn
Procedia PDF Downloads 202946 Three-Dimensional Generalized Thermoelasticity with Variable Thermal Conductivity
Authors: Hamdy M. Youssef, Mowffaq Oreijah, Hunaydi S. Alsharif
Abstract:
In this paper, a three-dimensional model of the generalized thermoelasticity with one relaxation time and variable thermal conductivity has been constructed. The resulting non-dimensional governing equations together with the Laplace and double Fourier transforms techniques have been applied to a three-dimensional half-space subjected to thermal loading with rectangular pulse and traction free in the directions of the principle co-ordinates. The inverses of double Fourier transforms, and Laplace transforms have been obtained numerically. Numerical results for the temperature increment, the invariant stress, the invariant strain, and the displacement are represented graphically. The variability of the thermal conductivity has significant effects on the thermal and the mechanical waves.Keywords: thermoelasticity, thermal conductivity, Laplace transforms, Fourier transforms
Procedia PDF Downloads 228945 Semirings of Graphs: An Approach Towards the Algebra of Graphs
Authors: Gete Umbrey, Saifur Rahman
Abstract:
Graphs are found to be most capable in computing, and its abstract structures have been applied in some specific computations and algorithms like in phase encoding controller, processor microcontroller, and synthesis of a CMOS switching network, etc. Being motivated by these works, we develop an independent approach to study semiring structures and various properties by defining the binary operations which in fact, seems analogous to an existing definition in some sense but with a different approach. This work emphasizes specifically on the construction of semigroup and semiring structures on the set of undirected graphs, and their properties are investigated therein. It is expected that the investigation done here may have some interesting applications in theoretical computer science, networking and decision making, and also on joining of two network systems.Keywords: graphs, join and union of graphs, semiring, weighted graphs
Procedia PDF Downloads 148944 A Modified Nonlinear Conjugate Gradient Algorithm for Large Scale Unconstrained Optimization Problems
Authors: Tsegay Giday Woldu, Haibin Zhang, Xin Zhang, Yemane Hailu Fissuh
Abstract:
It is well known that nonlinear conjugate gradient method is one of the widely used first order methods to solve large scale unconstrained smooth optimization problems. Because of the low memory requirement, attractive theoretical features, practical computational efficiency and nice convergence properties, nonlinear conjugate gradient methods have a special role for solving large scale unconstrained optimization problems. Large scale optimization problems are with important applications in practical and scientific world. However, nonlinear conjugate gradient methods have restricted information about the curvature of the objective function and they are likely less efficient and robust compared to some second order algorithms. To overcome these drawbacks, the new modified nonlinear conjugate gradient method is presented. The noticeable features of our work are that the new search direction possesses the sufficient descent property independent of any line search and it belongs to a trust region. Under mild assumptions and standard Wolfe line search technique, the global convergence property of the proposed algorithm is established. Furthermore, to test the practical computational performance of our new algorithm, numerical experiments are provided and implemented on the set of some large dimensional unconstrained problems. The numerical results show that the proposed algorithm is an efficient and robust compared with other similar algorithms.Keywords: conjugate gradient method, global convergence, large scale optimization, sufficient descent property
Procedia PDF Downloads 205943 Location Choice of Firms in an Unequal Length Streets Model: Game Theory Approach as an Extension of the Spoke Model
Authors: Kiumars Shahbazi, Salah Salimian, Abdolrahim Hashemi Dizaj
Abstract:
Locating is one of the key elements in success and survival of industrial centers and has great impact on cost reduction of establishment and launching of various economic activities. In this study, streets with unequal length model have been used that is the classic extension of Spoke model; however with unlimited number of streets with uneven lengths. The results showed that the spoke model is a special case of streets with unequal length model. According to the results of this study, if the strategy of enterprises and firms is to select both price and location, there would be no balance in the game. Furthermore, increased length of streets leads to increased profit of enterprises and with increased number of streets, the enterprises choose locations that are far from center (the maximum differentiation), and the enterprises' output will decrease. Moreover, the enterprise production rate will incline toward zero when the number of streets goes to infinity, and complete competition outcome will be achieved.Keywords: locating, Nash equilibrium, streets with unequal length model, streets with unequal length model
Procedia PDF Downloads 203942 Free Convective Flow in a Vertical Cylinder with Heat Sink: A Numerical Study
Authors: Emmanuel Omokhuale
Abstract:
A mathematical model is presented to study free convective boundary layer flow in a semi-infinite vertical cylinder with heat sink effect in a porous medium. The governing dimensional governing partial differential equations (PDEs) with corresponding initial and boundary conditions are approximated and solved numerically employing finite difference method (FDM) the implicit type. Stability and convergence of the scheme are also established. Furthermore, the influence of significant physical parameters on the flow characteristics was analysed and shown graphically. The obtained results are benchmarked with previously published works in order to access the accuracy of the numerical method and found to be in good agreement.Keywords: free convection flow, vertical cylinder, implicit finite difference method, heat sink and porous medium
Procedia PDF Downloads 141941 Economical Dependency Evolution and Complexity
Authors: Allé Dieng, Mamadou Bousso, Latif Dramani
Abstract:
The purpose of this work is to show the complexity behind economical interrelations in a country and provide a linear dynamic model of economical dependency evolution in a country. The model is based on National Transfer Account which is one of the most robust methodology developed in order to measure a level of demographic dividend captured in a country. It is built upon three major factors: demography, economical dependency and migration. The established mathematical model has been simulated using Netlogo software. The innovation of this study is in describing economical dependency as a complex system and simulating using mathematical equation the evolution of the two populations: the economical dependent and the non-economical dependent as defined in the National Transfer Account methodology. It also allows us to see the interactions and behaviors of both populations. The model can track individual characteristics and look at the effect of birth and death rates on the evolution of these two populations. The developed model is useful to understand how demographic and economic phenomenon are relatedKeywords: ABM, demographic dividend, National Transfer Accounts (NTA), ODE
Procedia PDF Downloads 205940 Natural Emergence of a Core Structure in Networks via Clique Percolation
Authors: A. Melka, N. Slater, A. Mualem, Y. Louzoun
Abstract:
Networks are often presented as containing a “core” and a “periphery.” The existence of a core suggests that some vertices are central and form the skeleton of the network, to which all other vertices are connected. An alternative view of graphs is through communities. Multiple measures have been proposed for dense communities in graphs, the most classical being k-cliques, k-cores, and k-plexes, all presenting groups of tightly connected vertices. We here show that the edge number thresholds for such communities to emerge and for their percolation into a single dense connectivity component are very close, in all networks studied. These percolating cliques produce a natural core and periphery structure. This result is generic and is tested in configuration models and in real-world networks. This is also true for k-cores and k-plexes. Thus, the emergence of this connectedness among communities leading to a core is not dependent on some specific mechanism but a direct result of the natural percolation of dense communities.Keywords: cliques, core structure, percolation, phase transition
Procedia PDF Downloads 171939 Joint Modeling of Longitudinal and Time-To-Event Data with Latent Variable
Authors: Xinyuan Y. Song, Kai Kang
Abstract:
Joint models for analyzing longitudinal and survival data are widely used to investigate the relationship between a failure time process and time-variant predictors. A common assumption in conventional joint models in the survival analysis literature is that all predictors are observable. However, this assumption may not always be supported because unobservable traits, namely, latent variables, which are indirectly observable and should be measured through multiple observed variables, are commonly encountered in the medical, behavioral, and financial research settings. In this study, a joint modeling approach to deal with this feature is proposed. The proposed model comprises three parts. The first part is a dynamic factor analysis model for characterizing latent variables through multiple observed indicators over time. The second part is a random coefficient trajectory model for describing the individual trajectories of latent variables. The third part is a proportional hazard model for examining the effects of time-invariant predictors and the longitudinal trajectories of time-variant latent risk factors on hazards of interest. A Bayesian approach coupled with a Markov chain Monte Carlo algorithm to perform statistical inference. An application of the proposed joint model to a study on the Alzheimer's disease neuroimaging Initiative is presented.Keywords: Bayesian analysis, joint model, longitudinal data, time-to-event data
Procedia PDF Downloads 143938 Fuzzy Ideal Topological Spaces
Authors: Ali Koam, Ismail Ibedou, S. E. Abbas
Abstract:
In this paper, it is introduced the notion of r-fuzzy ideal separation axioms Tᵢi = 0; 1; 2 based on a fuzzy ideal I on a fuzzy topological space (X; τ). An r-fuzzy ideal connectedness related to the fuzzy ideal I is introduced which has relations with a previous r-fuzzy fuzzy connectedness. An r-fuzzy ideal compactness related to Ι is introduced which has also relations with many other types of fuzzy compactness.Keywords: fuzzy ideal, fuzzy separation axioms, fuzzy compactness, fuzzy connectedness
Procedia PDF Downloads 266937 Metric Dimension on Line Graph of Honeycomb Networks
Authors: M. Hussain, Aqsa Farooq
Abstract:
Let G = (V,E) be a connected graph and distance between any two vertices a and b in G is a−b geodesic and is denoted by d(a, b). A set of vertices W resolves a graph G if each vertex is uniquely determined by its vector of distances to the vertices in W. A metric dimension of G is the minimum cardinality of a resolving set of G. In this paper line graph of honeycomb network has been derived and then we calculated the metric dimension on line graph of honeycomb network.Keywords: Resolving set, Metric dimension, Honeycomb network, Line graph
Procedia PDF Downloads 200936 Factors Influencing the Enjoyment and Performance of Students in Statistics Service Courses: A Mixed-Method Study
Authors: Wilma Coetzee
Abstract:
Statistics lecturers experience that many students who are taking a service course in statistics do not like statistics. Students in these courses tend to struggle and do not perform well. This research takes a look at the student’s perspective, with the aim to determine how to change the teaching of statistics so that students will enjoy it more and perform better. Questionnaires were used to determine the perspectives of first year service statistics students at a South African university. Factors addressed included motivation to study, attitude toward statistics, statistical anxiety, mathematical abilities and tendency to procrastinate. Logistic regression was used to determine what contributes to students performing badly in statistics. The results show that the factors that contribute the most to students performing badly are: statistical anxiety, not being motivated and having had mathematical literacy instead of mathematics in secondary school. Two open ended questions were included in the questionnaire: 'I will enjoy statistics more if…' and 'I will perform better in statistics if…'. The answers to these questions were analyzed using qualitative methods. Frequent themes were identified for each of the questions. A simulation study incorporating bootstrapping was done to determine the saturation of the themes. The majority of the students indicated that they would perform better in statistics if they studied more, managed their time better, had a flare for mathematics and if the lecturer was able to explain difficult concepts better. They also want more active learning. To ensure that students enjoy statistics more, they want an active learning experience. They want fun activities, more interaction with the lecturer and with one another, more computer based problems, and more challenges. They want a better understanding of the subject, want to understand the relevance of statistics to their future career and want excellent lecturers. These findings can be used to direct the improvement of the tuition of statistics.Keywords: active learning, performance in statistics, statistical anxiety, statistics education
Procedia PDF Downloads 147935 An IM-COH Algorithm Neural Network Optimization with Cuckoo Search Algorithm for Time Series Samples
Authors: Wullapa Wongsinlatam
Abstract:
Back propagation algorithm (BP) is a widely used technique in artificial neural network and has been used as a tool for solving the time series problems, such as decreasing training time, maximizing the ability to fall into local minima, and optimizing sensitivity of the initial weights and bias. This paper proposes an improvement of a BP technique which is called IM-COH algorithm (IM-COH). By combining IM-COH algorithm with cuckoo search algorithm (CS), the result is cuckoo search improved control output hidden layer algorithm (CS-IM-COH). This new algorithm has a better ability in optimizing sensitivity of the initial weights and bias than the original BP algorithm. In this research, the algorithm of CS-IM-COH is compared with the original BP, the IM-COH, and the original BP with CS (CS-BP). Furthermore, the selected benchmarks, four time series samples, are shown in this research for illustration. The research shows that the CS-IM-COH algorithm give the best forecasting results compared with the selected samples.Keywords: artificial neural networks, back propagation algorithm, time series, local minima problem, metaheuristic optimization
Procedia PDF Downloads 152934 Thermal Instability in Rivlin-Ericksen Elastico-Viscous Nanofluid with Connective Boundary Condition: Effect of Vertical Throughflow
Authors: Shivani Saini
Abstract:
The effect of vertical throughflow on the onset of convection in Rivlin-Ericksen Elastico-Viscous nanofluid with convective boundary condition is investigated. The flow is stimulated with modified Darcy model under the assumption that the nanoparticle volume fraction is not actively managed on the boundaries. The heat conservation equation is formulated by introducing the convective term of nanoparticle flux. A linear stability analysis based upon normal mode is performed, and an approximate solution of eigenvalue problems is obtained using the Galerkin weighted residual method. Investigation of the dependence of the Rayleigh number on various viscous and nanofluid parameter is performed. It is found that through flow and nanofluid parameters hasten the convection while capacity ratio, kinematics viscoelasticity, and Vadasz number do not govern the stationary convection. Using the convective component of nanoparticle flux, critical wave number is the function of nanofluid parameters as well as the throughflow parameter. The obtained solution provides important physical insight into the behavior of this model.Keywords: Darcy model, nanofluid, porous layer, throughflow
Procedia PDF Downloads 137933 Analyzing Large Scale Recurrent Event Data with a Divide-And-Conquer Approach
Authors: Jerry Q. Cheng
Abstract:
Currently, in analyzing large-scale recurrent event data, there are many challenges such as memory limitations, unscalable computing time, etc. In this research, a divide-and-conquer method is proposed using parametric frailty models. Specifically, the data is randomly divided into many subsets, and the maximum likelihood estimator from each individual data set is obtained. Then a weighted method is proposed to combine these individual estimators as the final estimator. It is shown that this divide-and-conquer estimator is asymptotically equivalent to the estimator based on the full data. Simulation studies are conducted to demonstrate the performance of this proposed method. This approach is applied to a large real dataset of repeated heart failure hospitalizations.Keywords: big data analytics, divide-and-conquer, recurrent event data, statistical computing
Procedia PDF Downloads 165932 Construction of Finite Woven Frames through Bounded Linear Operators
Authors: A. Bhandari, S. Mukherjee
Abstract:
Two frames in a Hilbert space are called woven or weaving if all possible merge combinations between them generate frames of the Hilbert space with uniform frame bounds. Weaving frames are powerful tools in wireless sensor networks which require distributed data processing. Considering the practical applications, this article deals with finite woven frames. We provide methods of constructing finite woven frames, in particular, bounded linear operators are used to construct woven frames from a given frame. Several examples are discussed. We also introduce the notion of woven frame sequences and characterize them through the concepts of gaps and angles between spaces.Keywords: frames, woven frames, gap, angle
Procedia PDF Downloads 193931 Polynomial Chaos Expansion Combined with Exponential Spline for Singularly Perturbed Boundary Value Problems with Random Parameter
Authors: W. K. Zahra, M. A. El-Beltagy, R. R. Elkhadrawy
Abstract:
So many practical problems in science and technology developed over the past decays. For instance, the mathematical boundary layer theory or the approximation of solution for different problems described by differential equations. When such problems consider large or small parameters, they become increasingly complex and therefore require the use of asymptotic methods. In this work, we consider the singularly perturbed boundary value problems which contain very small parameters. Moreover, we will consider these perturbation parameters as random variables. We propose a numerical method to solve this kind of problems. The proposed method is based on an exponential spline, Shishkin mesh discretization, and polynomial chaos expansion. The polynomial chaos expansion is used to handle the randomness exist in the perturbation parameter. Furthermore, the Monte Carlo Simulations (MCS) are used to validate the solution and the accuracy of the proposed method. Numerical results are provided to show the applicability and efficiency of the proposed method, which maintains a very remarkable high accuracy and it is ε-uniform convergence of almost second order.Keywords: singular perturbation problem, polynomial chaos expansion, Shishkin mesh, two small parameters, exponential spline
Procedia PDF Downloads 160930 Numerical Investigation and Optimization of the Effect of Number of Blade and Blade Type on the Suction Pressure and Outlet Mass Flow Rate of a Centrifugal Fan
Authors: Ogan Karabas, Suleyman Yigit
Abstract:
Number of blade and blade type of centrifugal fans are the most decisive factor on the field of application, noise level, suction pressure and outlet mass flow rate. Nowadays, in order to determine these effects on centrifugal fans, numerical studies are carried out in addition to experimental studies. In this study, it is aimed to numerically investigate the changes of suction pressure and outlet mass flow rate values of a centrifugal fan according to the number of blade and blade type. Centrifugal fans of the same size with forward, backward and straight blade type were analyzed by using a simulation program and compared with each other. This analysis was carried out under steady state condition by selecting k-Ɛ turbulence model and air is assumed incompressible. Then, 16, 32 and 48 blade centrifugal fans were again analyzed by using same simulation program, and the optimum number of blades was determined for the suction pressure and the outlet mass flow rate. According to the results of the analysis, it was obtained that the suction pressure in the 32 blade fan was twice the value obtained in the 16 blade fan. In addition, the outlet mass flow rate increased by 45% with the increase in the number of blade from 16 to 32. There is no significant change observed on the suction pressure and outlet mass flow rate when the number of blades increased from 32 to 48. In the light of the analysis results, the optimum blade number was determined as 32.Keywords: blade type, centrifugal fan, cfd, outlet mass flow rate, suction pressure
Procedia PDF Downloads 404929 Bayesian Meta-Analysis to Account for Heterogeneity in Studies Relating Life Events to Disease
Authors: Elizabeth Stojanovski
Abstract:
Associations between life events and various forms of cancers have been identified. The purpose of a recent random-effects meta-analysis was to identify studies that examined the association between adverse events associated with changes to financial status including decreased income and breast cancer risk. The same association was studied in four separate studies which displayed traits that were not consistent between studies such as the study design, location and time frame. It was of interest to pool information from various studies to help identify characteristics that differentiated study results. Two random-effects Bayesian meta-analysis models are proposed to combine the reported estimates of the described studies. The proposed models allow major sources of variation to be taken into account, including study level characteristics, between study variance, and within study variance and illustrate the ease with which uncertainty can be incorporated using a hierarchical Bayesian modelling approach.Keywords: random-effects, meta-analysis, Bayesian, variation
Procedia PDF Downloads 160928 Predictive Modelling of Aircraft Component Replacement Using Imbalanced Learning and Ensemble Method
Authors: Dangut Maren David, Skaf Zakwan
Abstract:
Adequate monitoring of vehicle component in other to obtain high uptime is the goal of predictive maintenance, the major challenge faced by businesses in industries is the significant cost associated with a delay in service delivery due to system downtime. Most of those businesses are interested in predicting those problems and proactively prevent them in advance before it occurs, which is the core advantage of Prognostic Health Management (PHM) application. The recent emergence of industry 4.0 or industrial internet of things (IIoT) has led to the need for monitoring systems activities and enhancing system-to-system or component-to- component interactions, this has resulted to a large generation of data known as big data. Analysis of big data represents an increasingly important, however, due to complexity inherently in the dataset such as imbalance classification problems, it becomes extremely difficult to build a model with accurate high precision. Data-driven predictive modeling for condition-based maintenance (CBM) has recently drowned research interest with growing attention to both academics and industries. The large data generated from industrial process inherently comes with a different degree of complexity which posed a challenge for analytics. Thus, imbalance classification problem exists perversely in industrial datasets which can affect the performance of learning algorithms yielding to poor classifier accuracy in model development. Misclassification of faults can result in unplanned breakdown leading economic loss. In this paper, an advanced approach for handling imbalance classification problem is proposed and then a prognostic model for predicting aircraft component replacement is developed to predict component replacement in advanced by exploring aircraft historical data, the approached is based on hybrid ensemble-based method which improves the prediction of the minority class during learning, we also investigate the impact of our approach on multiclass imbalance problem. We validate the feasibility and effectiveness in terms of the performance of our approach using real-world aircraft operation and maintenance datasets, which spans over 7 years. Our approach shows better performance compared to other similar approaches. We also validate our approach strength for handling multiclass imbalanced dataset, our results also show good performance compared to other based classifiers.Keywords: prognostics, data-driven, imbalance classification, deep learning
Procedia PDF Downloads 174927 Globally Attractive Mild Solutions for Non-Local in Time Subdiffusion Equations of Neutral Type
Authors: Jorge Gonzalez Camus, Carlos Lizama
Abstract:
In this work is proved the existence of at least one globally attractive mild solution to the Cauchy problem, for fractional evolution equation of neutral type, involving the fractional derivate in Caputo sense. An almost sectorial operator on a Banach space X and a kernel belonging to a large class appears in the equation, which covers many relevant cases from physics applications, in particular, the important case of time - fractional evolution equations of neutral type. The main tool used in this work was the Hausdorff measure of noncompactness and fixed point theorems, specifically Darbo-type. Initially, the equation is a Cauchy problem, involving a fractional derivate in Caputo sense. Then, is formulated the equivalent integral version, and defining a convenient functional, using the analytic integral resolvent operator, and verifying the hypothesis of the fixed point theorem of Darbo type, give us the existence of mild solution for the initial problem. Furthermore, each mild solution is globally attractive, a property that is desired in asymptotic behavior for that solution.Keywords: attractive mild solutions, integral Volterra equations, neutral type equations, non-local in time equations
Procedia PDF Downloads 158