Search results for: random kernel
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2224

Search results for: random kernel

2134 A Convergent Interacting Particle Method for Computing Kpp Front Speeds in Random Flows

Authors: Tan Zhang, Zhongjian Wang, Jack Xin, Zhiwen Zhang

Abstract:

We aim to efficiently compute the spreading speeds of reaction-diffusion-advection (RDA) fronts in divergence-free random flows under the Kolmogorov-Petrovsky-Piskunov (KPP) nonlinearity. We study a stochastic interacting particle method (IPM) for the reduced principal eigenvalue (Lyapunov exponent) problem of an associated linear advection-diffusion operator with spatially random coefficients. The Fourier representation of the random advection field and the Feynman-Kac (FK) formula of the principal eigenvalue (Lyapunov exponent) form the foundation of our method implemented as a genetic evolution algorithm. The particles undergo advection-diffusion and mutation/selection through a fitness function originated in the FK semigroup. We analyze the convergence of the algorithm based on operator splitting and present numerical results on representative flows such as 2D cellular flow and 3D Arnold-Beltrami-Childress (ABC) flow under random perturbations. The 2D examples serve as a consistency check with semi-Lagrangian computation. The 3D results demonstrate that IPM, being mesh-free and self-adaptive, is simple to implement and efficient for computing front spreading speeds in the advection-dominated regime for high-dimensional random flows on unbounded domains where no truncation is needed.

Keywords: KPP front speeds, random flows, Feynman-Kac semigroups, interacting particle method, convergence analysis

Procedia PDF Downloads 14
2133 Empirical Study of Running Correlations in Exam Marks: Same Statistical Pattern as Chance

Authors: Weisi Guo

Abstract:

It is well established that there may be running correlations in sequential exam marks due to students sitting in the order of course registration patterns. As such, a random and non-sequential sampling of exam marks is a standard recommended practice. Here, the paper examines a large number of exam data stretching several years across different modules to see the degree to which it is true. Using the real mark distribution as a generative process, it was found that random simulated data had no more sequential randomness than the real data. That is to say, the running correlations that one often observes are statistically identical to chance. Digging deeper, it was found that some high running correlations have students that indeed share a common course history and make similar mistakes. However, at the statistical scale of a module question, the combined effect is statistically similar to the random shuffling of papers. As such, there may not be the need to take random samples for marks, but it still remains good practice to mark papers in a random sequence to reduce the repetitive marking bias and errors.

Keywords: data analysis, empirical study, exams, marking

Procedia PDF Downloads 139
2132 Nonparametric Quantile Regression for Multivariate Spatial Data

Authors: S. H. Arnaud Kanga, O. Hili, S. Dabo-Niang

Abstract:

Spatial prediction is an issue appealing and attracting several fields such as agriculture, environmental sciences, ecology, econometrics, and many others. Although multiple non-parametric prediction methods exist for spatial data, those are based on the conditional expectation. This paper took a different approach by examining a non-parametric spatial predictor of the conditional quantile. The study especially observes the stationary multidimensional spatial process over a rectangular domain. Indeed, the proposed quantile is obtained by inverting the conditional distribution function. Furthermore, the proposed estimator of the conditional distribution function depends on three kernels, where one of them controls the distance between spatial locations, while the other two control the distance between observations. In addition, the almost complete convergence and the convergence in mean order q of the kernel predictor are obtained when the sample considered is alpha-mixing. Such approach of the prediction method gives the advantage of accuracy as it overcomes sensitivity to extreme and outliers values.

Keywords: conditional quantile, kernel, nonparametric, stationary

Procedia PDF Downloads 125
2131 Bio Ethanol Production From the Co-Mixture of Jatropha Carcus L. Kernel Cake and Rice Straw

Authors: Felix U. Asoiro, Daniel I. Eleazar, Peter O. Offor

Abstract:

As a result of increasing energy demands, research in bioethanol has increased in recent years all through the world, in abide to partially or totally replace renewable energy supplies. The first and third generation feedstocks used for biofuel production have fundamental drawbacks. Waste rice straw and cake from second generation feedstock like Jatropha curcas l. kernel (JC) is seen as non-food feedstock and promising candidates for the industrial production of bioethanol. In this study, JC and rice husk (RH) wastes were characterized for proximate composition. Bioethanol was produced from the residual polysaccharides present in rice husk (RH) and Jatropha seed cake by sequential hydrolytic and fermentative processes at varying mixing proportions (50 g JC/50 g RH, 100 g JC/10 g RH, 100 g JC/20 g RH, 100 g JC/50 g RH, 100 g JC/100 g RH, 100 g JC/200 g RH and 200 g JC/100 g RH) and particle sizes (0.25, 0.5 and 1.00 mm). Mixing proportions and particle size significantly affected both bioethanol yield and some bioethanol properties. Bioethanol yield (%) increased with an increase in particle size. The highest bioethanol (8.67%) was produced at a mixing proportion of 100 g JC/50g RH at 0.25 mm particle size. The bioethanol had the lowest values of specific gravity and density of 1.25 and 0.92 g cm-3 and the highest values of 1.57 and 0.97 g cm-3 respectively. The highest values of viscosity (4.64 cSt) were obtained with 200 g JC/100 g RH, at 1.00 mm particle size. The maximum flash point and cloud point values were 139.9 oC and 23.7oC (100 g JC/200 g RH) at 1 mm and 0.5 mm particle sizes respectively. The maximum pour point value recorded was 3.85oC (100 g JC/50 g RH) at 1 mm particle size. The paper concludes that bioethanol can be recovered from JC and RH wastes. JC and RH blending proportions as well as particle sizes are important factors in bioethanol production.

Keywords: bioethanol, hydrolysis, Jatropha curcas l. kernel, rice husk, fermentation, proximate composition

Procedia PDF Downloads 67
2130 Influence of Random Fibre Packing on the Compressive Strength of Fibre Reinforced Plastic

Authors: Y. Wang, S. Zhang, X. Chen

Abstract:

The longitudinal compressive strength of fibre reinforced plastic (FRP) possess a large stochastic variability, which limits efficient application of composite structures. This study aims to address how the random fibre packing affects the uncertainty of FRP compressive strength. An novel approach is proposed to generate random fibre packing status by a combination of Latin hypercube sampling and random sequential expansion. 3D nonlinear finite element model is built which incorporates both the matrix plasticity and fibre geometrical instability. The matrix is modeled by isotropic ideal elasto-plastic solid elements, and the fibres are modeled by linear-elastic rebar elements. Composite with a series of different nominal fibre volume fractions are studied. Premature fibre waviness at different magnitude and direction is introduced in the finite element model. Compressive tests on uni-directional CFRP (carbon fibre reinforced plastic) are conducted following the ASTM D6641. By a comparison of 3D FE models and compressive tests, it is clearly shown that the stochastic variation of compressive strength is partly caused by the random fibre packing, and normal or lognormal distribution tends to be a good fit the probabilistic compressive strength. Furthermore, it is also observed that different random fibre packing could trigger two different fibre micro-buckling modes while subjected to longitudinal compression: out-of-plane buckling and twisted buckling. The out-of-plane buckling mode results much larger compressive strength, and this is the major reason why the random fibre packing results a large uncertainty in the FRP compressive strength. This study would contribute to new approaches to the quality control of FRP considering higher compressive strength or lower uncertainty.

Keywords: compressive strength, FRP, micro-buckling, random fibre packing

Procedia PDF Downloads 247
2129 Processing Methods for Increasing the Yield, Nutritional Value and Stability of Coconut Milk

Authors: Archana G. Lamdande, Shyam R. Garud, K. S. M. S. Raghavarao

Abstract:

Coconut has two edible parts, that is, a white kernel (solid endosperm) and coconut water (liquid endosperm). The white kernel is generally used in fresh or dried form for culinary purposes. Coconut testa, is the brown skin, covering the coconut kernel. It is removed by paring of wet coconut and obtained as a by-product in coconut processing industries during the production of products such as desiccated coconut, coconut milk, whole coconut milk powder and virgin coconut oil. At present, it is used as animal feed component after drying and recovering the residual oil (by expelling). Experiments were carried out on expelling of coconut milk for shredded coconut with and without testa removal, in order to explore the possibility of increasing the milk yield and value addition in terms of increased polyphenol content. The color characteristics of coconut milk obtained from the grating without removal of testa were observed to be L* 82.79, a* 0.0125, b* 6.245, while that obtained from grating with removal of testa were L* 83.24, a* -0.7925, b* 3.1. A significant increase was observed in total phenol content of coconut milk obtained from the grating with testa (833.8 µl/ml) when compared to that from without testa (521.3 µl/ml). However, significant difference was not observed in protein content of coconut milk obtained from the grating with and without testa (4.9 and 5.0% w/w, respectively). Coconut milk obtained from grating without removal of testa showed higher milk yield (62% w/w) when compared to that obtained from grating with removal of testa (60% w/w). The fat content in coconut milk was observed to be 32% (w/w), and it is unstable due to such a high fat content. Therefore, several experiments were carried out for examining its stability by adjusting the fat content at different levels (32, 28, 24, and 20% w/w). It was found that the coconut milk was more stable with a fat content of 24 % (w/w). Homogenization and ultrasonication and their combinations were used for exploring the possibility of increasing the stability of coconut milk. The microscopic study was carried out for analyzing the size of fat globules and the degree of their uniform distribution.

Keywords: coconut milk, homogenization, stability, testa, ultrasonication

Procedia PDF Downloads 277
2128 The Use of Palm Kernel Cake in Ration and Its Influence on VFA, NH3 and pH Rumen Fluid of Goat

Authors: Arief, Noovirman Jamarun, Benni Satria

Abstract:

Background: The main problem in the development of livestock in Indonesia is feed both in terms of quality and quantity. On the other hand, conventional feed ingredients are expensive and difficult to obtain. Therefore, it is necessary to find alternative feed ingredients that have good quality, potential, and low cost. Feed ingredients that meet the above requirements are by-products of the palm oil industry, namely palm kernel cake (PKC). This study aims to obtain the best PKC composition for Etawa goat concentrate ration. Material and Methode : This research consists of 2 stages. Stage I is invitro study using Tilley and Terry method. The study used a Completely Randomized Design (CRD) with 4 treatments of rations and 4 replications. The treatment is the composition of the use of palm kernel cake (PKC) in the ration, namely, A). 10%, B). 20%, C). 30%, D). 40%. Other feed ingredients are corn, rice bran, tofu waste and minerals. The measured variables are the characteristics of the rumen fluid (pH, VFA and NH3). Stage II was done using the best ration of stage I (Ration C), followed by testing the use of Tithonia (Thitonia difersifolia) and agricultural waste of sweet potato leaves as a source of forage for livestock by in-vitro. The study used a Completely Randomized Design (CRD) with 3 treatments and 5 replications. The treatments were: Treatment A) Best Concentrate Ration Stage I + Titonia (Thitonia difersifolia), Treatment B) Best Concentrate Ration Stage I + Tithonia (Thitonia difersifolia) and Sweet potato Leaves, Treatment C) Best Concentrate Ration Stage I + Sweet potato leaves. The data obtained were analyzed using variance analysis while the differences between treatments were tested using the Duncant Multiple Range Test (DMRT) according to Steel and Torrie. Results of Stage II showed that the use of PKC in rations as concentrate feed combined with forage originating from Tithonia (Thitonia difersifolia) and sweet potato leaves produced pH, VFA and NH3-N which were still in normal conditions. The best treatment was obtained from diet B (P <0.05) with 6.9 pH, 116.29 mM VFA and 15mM NH3-N. Conclussion From the results of the study it can be concluded that PKC can be used as feed ingredients for dairy goat concentrate with a combination of forage from Tithonia (Tithonia difersifolia) and sweet potato leaves.

Keywords: palm oil by-product, palm kernel cake, concentrate, rumen fluid, Etawa goat

Procedia PDF Downloads 144
2127 Use of In-line Data Analytics and Empirical Model for Early Fault Detection

Authors: Hyun-Woo Cho

Abstract:

Automatic process monitoring schemes are designed to give early warnings for unusual process events or abnormalities as soon as possible. For this end, various techniques have been developed and utilized in various industrial processes. It includes multivariate statistical methods, representation skills in reduced spaces, kernel-based nonlinear techniques, etc. This work presents a nonlinear empirical monitoring scheme for batch type production processes with incomplete process measurement data. While normal operation data are easy to get, unusual fault data occurs infrequently and thus are difficult to collect. In this work, noise filtering steps are added in order to enhance monitoring performance by eliminating irrelevant information of the data. The performance of the monitoring scheme was demonstrated using batch process data. The results showed that the monitoring performance was improved significantly in terms of detection success rate of process fault.

Keywords: batch process, monitoring, measurement, kernel method

Procedia PDF Downloads 296
2126 Simulation of Glass Breakage Using Voronoi Random Field Tessellations

Authors: Michael A. Kraus, Navid Pourmoghaddam, Martin Botz, Jens Schneider, Geralt Siebert

Abstract:

Fragmentation analysis of tempered glass gives insight into the quality of the tempering process and defines a certain degree of safety as well. Different standard such as the European EN 12150-1 or the American ASTM C 1048/CPSC 16 CFR 1201 define a minimum number of fragments required for soda-lime safety glass on the basis of fragmentation test results for classification. This work presents an approach for the glass breakage pattern prediction using a Voronoi Tesselation over Random Fields. The random Voronoi tessellation is trained with and validated against data from several breakage patterns. The fragments in observation areas of 50 mm x 50 mm were used for training and validation. All glass specimen used in this study were commercially available soda-lime glasses at three different thicknesses levels of 4 mm, 8 mm and 12 mm. The results of this work form a Bayesian framework for the training and prediction of breakage patterns of tempered soda-lime glass using a Voronoi Random Field Tesselation. Uncertainties occurring in this process can be well quantified, and several statistical measures of the pattern can be preservation with this method. Within this work it was found, that different Random Fields as basis for the Voronoi Tesselation lead to differently well fitted statistical properties of the glass breakage patterns. As the methodology is derived and kept general, the framework could be also applied to other random tesselations and crack pattern modelling purposes.

Keywords: glass breakage predicition, Voronoi Random Field Tessellation, fragmentation analysis, Bayesian parameter identification

Procedia PDF Downloads 131
2125 Reducing the Cooking Time of Bambara Groundnut (BGN)

Authors: Auswell Amfo-Antiri, Esther Eshun, Theresa A. Amu

Abstract:

Cooking Bambara groundnut (Bambara beans) is time and energy-consuming. Over time, some substances have been used to help reduce cooking time and save energy. This experimental study was carried out to find ways of reducing the cooking time of Bambara groundnut using selected organic substances. Twenty grams (20g) each of fresh pawpaw leaves, guava leaves, ginger, onion, and palm kernel were cooked with five samples of 200g of the creamy variety of raw Bambara groundnut. A control was cooked without any organic substance added. All six samples were cooked with equal quantities of water (4L); the gas mark used for cooking the samples was marked 5, the highest for the largest burner, using the same cooking pot. Gas matter. The control sample used 192 minutes to cook thoroughly. The ginger-treated sample (AET02) had the shortest cooking time of 145 minutes, followed by the onion-treated sample (AET05), with a cooking time of 157 minutes. The sample cooked with Palm kernel (AET06) and Pawpaw (AET04) used 172 minutes and 174 minutes, respectively, while sample AET03, cooked with Guava, used 185 minutes for cooking. The difference in cooking time for the sample treated with ginger (AET02) and onion (AET05) was 47 minutes and 35 minutes, respectively, as compared with the control. The comparison between Control and Pawpaw produced [p=0.163>0.05]; Control and Ginger yielded [p=0.006<0.05]; Control and Kernel resulted in [p=0.128>0.05]; Control and Guava resulted in [p=0.560>0.05]. The study concluded that ginger and onions comparatively reduced the cooking time for Bambara ground nut appreciably. The study recommended that ginger and onions could be used to reduce the cooking time of Bambara groundnut.

Keywords: cooking time, organic substances, ginger, onions, pawpaw leaves, guava leaves, bambara groundnut

Procedia PDF Downloads 51
2124 Occurrence and Levels of Mycotoxins in On-Farm Stored Sesame in Major-Growing Districts of Ethiopia

Authors: S. Alemayehu, F. A. Abera, K. M. Ayimut, R. Mahroof, J. Harvey, B. Subramanyam

Abstract:

The occurrence of mycotoxins in sesame seeds poses a significant threat to food safety and the economy in Ethiopia. This study aimed to determine the levels and occurrence of mycotoxins in on-farm stored sesame seeds in major-growing districts of Ethiopia. A total of 470 sesame seed samples were collected from randomly selected farmers' storage structures in five major-growing districts using purposive sampling techniques. An enzyme-linked immunosorbent assay (ELISA) was used to analyze the collected samples for the presence of four mycotoxins: total aflatoxins (AFT), ochratoxin A (OTA), total fumonisins (FUM), and deoxynivalenol (DON). The study found that all samples contained varying levels of mycotoxins, with AFT and DON being the most prevalent. AFT concentrations in detected samples ranged from 2.5 to 27.8 parts per billion (ppb), with a mean concentration of 13.8 ppb. OTA levels ranged from 5.0 ppb to 9.7 ppb, with a mean level of 7.1 ppb. Total fumonisin concentrations ranged from 300 to 1300 ppb in all samples, with a mean of 800 ppb. DON concentrations ranged from 560 to 700 ppb in the analyzed samples. The majority (96.8%) of the samples were safe from AFT, FUM, and DON mean levels when compared to the Federal Drug Administration maximum limit. AFT-OTA, DON-OTA, AFT-FUM, FUM-DON, and FUM-OTA, respectively, had co-occurrence rates of 44.0, 38.3, 33.8, 30.2, 29.8 and 26.0% for mycotoxins. On average, 37.2% of the sesame samples had fungal infection, and seed germination rates ranged from 66.8% to 91.1%. The Limmu district had higher levels of total aflatoxins, kernel infection, and lower germination rates than other districts. The Wollega variety of sesame had higher kernel infection, total aflatoxins concentration, and lower germination rates than other varieties. Grain age had a statistically significant (p<0.05) effect on both kernel infection and germination. The storage methods used for sesame in major-growing districts of Ethiopia favor mycotoxin-producing fungi. As the levels of mycotoxins in sesame are of public health significance, stakeholders should come together to identify secure and suitable storage technologies to maintain the quantity and quality of sesame at the level of smallholder farmers. This study suggests the need for suitable storage technologies to maintain the quality of sesame and reduce the risk of mycotoxin contamination.

Keywords: districts, seed germination, kernel infection, moisture content, relative humidity, temperature

Procedia PDF Downloads 78
2123 Low-Cost Embedded Biometric System Based on Fingervein Modality

Authors: Randa Boukhris, Alima Damak, Dorra Sellami

Abstract:

Fingervein biometric authentication is one of the most popular and accurate technologies. However, low cost embedded solution is still an open problem. In this paper, a real-time implementation of fingervein recognition process embedded in Raspberry-Pi has been proposed. The use of Raspberry-Pi reduces overall system cost and size while allowing an easy user interface. Implementation of a target technology has guided to opt some specific parallel and simple processing algorithms. In the proposed system, we use four structural directional kernel elements for filtering finger vein images. Then, a Top-Hat and Bottom-Hat kernel filters are used to enhance the visibility and the appearance of venous images. For feature extraction step, a simple Local Directional Code (LDC) descriptor is applied. The proposed system presents an Error Equal Rate (EER) and Identification Rate (IR), respectively, equal to 0.02 and 98%. Furthermore, experimental results show that real-time operations have good performance.

Keywords: biometric, Bottom-Hat, Fingervein, LDC, Rasberry-Pi, ROI, Top-Hat

Procedia PDF Downloads 180
2122 Heterogeneity, Asymmetry and Extreme Risk Perception; Dynamic Evolution Detection From Implied Risk Neutral Density

Authors: Abderrahmen Aloulou, Younes Boujelbene

Abstract:

The current paper displays a new method of extracting information content from options prices by eliminating biases caused by daily variation of contract maturity. Based on Kernel regression tool, this non-parametric technique serves to obtain a spectrum of interpolated options with constant maturity horizons from negotiated optional contracts on the S&P TSX 60 index. This method makes it plausible to compare daily risk neutral densities from which extracting time continuous indicators allows the detection traders attitudes’ evolution, such as, belief homogeneity, asymmetry and extreme Risk Perception. Our findings indicate that the applied method contribute to develop effective trading strategies and to adjust monetary policies through controlling trader’s reactions to economic and monetary news.

Keywords: risk neutral densities, kernel, constant maturity horizons, homogeneity, asymmetry and extreme risk perception

Procedia PDF Downloads 459
2121 Segmentation of Liver Using Random Forest Classifier

Authors: Gajendra Kumar Mourya, Dinesh Bhatia, Akash Handique, Sunita Warjri, Syed Achaab Amir

Abstract:

Nowadays, Medical imaging has become an integral part of modern healthcare. Abdominal CT images are an invaluable mean for abdominal organ investigation and have been widely studied in the recent years. Diagnosis of liver pathologies is one of the major areas of current interests in the field of medical image processing and is still an open problem. To deeply study and diagnose the liver, segmentation of liver is done to identify which part of the liver is mostly affected. Manual segmentation of the liver in CT images is time-consuming and suffers from inter- and intra-observer differences. However, automatic or semi-automatic computer aided segmentation of the Liver is a challenging task due to inter-patient Liver shape and size variability. In this paper, we present a technique for automatic segmenting the liver from CT images using Random Forest Classifier. Random forests or random decision forests are an ensemble learning method for classification that operate by constructing a multitude of decision trees at training time and outputting the class that is the mode of the classes of the individual trees. After comparing with various other techniques, it was found that Random Forest Classifier provide a better segmentation results with respect to accuracy and speed. We have done the validation of our results using various techniques and it shows above 89% accuracy in all the cases.

Keywords: CT images, image validation, random forest, segmentation

Procedia PDF Downloads 281
2120 The Journey of a Malicious HTTP Request

Authors: M. Mansouri, P. Jaklitsch, E. Teiniker

Abstract:

SQL injection on web applications is a very popular kind of attack. There are mechanisms such as intrusion detection systems in order to detect this attack. These strategies often rely on techniques implemented at high layers of the application but do not consider the low level of system calls. The problem of only considering the high level perspective is that an attacker can circumvent the detection tools using certain techniques such as URL encoding. One technique currently used for detecting low-level attacks on privileged processes is the tracing of system calls. System calls act as a single gate to the Operating System (OS) kernel; they allow catching the critical data at an appropriate level of detail. Our basic assumption is that any type of application, be it a system service, utility program or Web application, “speaks” the language of system calls when having a conversation with the OS kernel. At this level we can see the actual attack while it is happening. We conduct an experiment in order to demonstrate the suitability of system call analysis for detecting SQL injection. We are able to detect the attack. Therefore we conclude that system calls are not only powerful in detecting low-level attacks but that they also enable us to detect high-level attacks such as SQL injection.

Keywords: Linux system calls, web attack detection, interception, SQL

Procedia PDF Downloads 324
2119 Three-Stage Multivariate Stratified Sample Surveys with Probabilistic Cost Constraint and Random Variance

Authors: Sanam Haseen, Abdul Bari

Abstract:

In this paper a three stage multivariate programming problem with random survey cost and variances as random variables has been formulated as a non-linear stochastic programming problem. The problem has been converted into an equivalent deterministic form using chance constraint programming and modified E-modeling. An empirical study of the problem has been done at the end of the paper using R-simulation.

Keywords: chance constraint programming, modified E-model, stochastic programming, stratified sample surveys, three stage sample surveys

Procedia PDF Downloads 429
2118 Blocking of Random Chat Apps at Home Routers for Juvenile Protection in South Korea

Authors: Min Jin Kwon, Seung Won Kim, Eui Yeon Kim, Haeyoung Lee

Abstract:

Numerous anonymous chat apps that help people to connect with random strangers have been released in South Korea. However, they become a serious problem for young people since young people often use them for channels of prostitution or sexual violence. Although ISPs in South Korea are responsible for making inappropriate content inaccessible on their networks, they do not block traffic of random chat apps since 1) the use of random chat apps is entirely legal. 2) it is reported that they use HTTP proxy blocking so that non-HTTP traffic cannot be blocked. In this paper, we propose a service model that can block random chat apps at home routers. A service provider manages a blacklist that contains blocked apps’ information. Home routers that subscribe the service filter the traffic of the apps out using deep packet inspection. We have implemented a prototype of the proposed model, including a centralized server providing the blacklist, a Raspberry Pi-based home router that can filter traffic of the apps out, and an Android app used by the router’s administrator to locally customize the blacklist.

Keywords: deep packet inspection, internet filtering, juvenile protection, technical blocking

Procedia PDF Downloads 321
2117 Tabu Random Algorithm for Guiding Mobile Robots

Authors: Kevin Worrall, Euan McGookin

Abstract:

The use of optimization algorithms is common across a large number of diverse fields. This work presents the use of a hybrid optimization algorithm applied to a mobile robot tasked with carrying out a search of an unknown environment. The algorithm is then applied to the multiple robots case, which results in a reduction in the time taken to carry out the search. The hybrid algorithm is a Random Search Algorithm fused with a Tabu mechanism. The work shows that the algorithm locates the desired points in a quicker time than a brute force search. The Tabu Random algorithm is shown to work within a simulated environment using a validated mathematical model. The simulation was run using three different environments with varying numbers of targets. As an algorithm, the Tabu Random is small, clear and can be implemented with minimal resources. The power of the algorithm is the speed at which it locates points of interest and the robustness to the number of robots involved. The number of robots can vary with no changes to the algorithm resulting in a flexible algorithm.

Keywords: algorithms, control, multi-agent, search and rescue

Procedia PDF Downloads 214
2116 Probability Sampling in Matched Case-Control Study in Drug Abuse

Authors: Surya R. Niraula, Devendra B Chhetry, Girish K. Singh, S. Nagesh, Frederick A. Connell

Abstract:

Background: Although random sampling is generally considered to be the gold standard for population-based research, the majority of drug abuse research is based on non-random sampling despite the well-known limitations of this kind of sampling. Method: We compared the statistical properties of two surveys of drug abuse in the same community: one using snowball sampling of drug users who then identified “friend controls” and the other using a random sample of non-drug users (controls) who then identified “friend cases.” Models to predict drug abuse based on risk factors were developed for each data set using conditional logistic regression. We compared the precision of each model using bootstrapping method and the predictive properties of each model using receiver operating characteristics (ROC) curves. Results: Analysis of 100 random bootstrap samples drawn from the snowball-sample data set showed a wide variation in the standard errors of the beta coefficients of the predictive model, none of which achieved statistical significance. One the other hand, bootstrap analysis of the random-sample data set showed less variation, and did not change the significance of the predictors at the 5% level when compared to the non-bootstrap analysis. Comparison of the area under the ROC curves using the model derived from the random-sample data set was similar when fitted to either data set (0.93, for random-sample data vs. 0.91 for snowball-sample data, p=0.35); however, when the model derived from the snowball-sample data set was fitted to each of the data sets, the areas under the curve were significantly different (0.98 vs. 0.83, p < .001). Conclusion: The proposed method of random sampling of controls appears to be superior from a statistical perspective to snowball sampling and may represent a viable alternative to snowball sampling.

Keywords: drug abuse, matched case-control study, non-probability sampling, probability sampling

Procedia PDF Downloads 467
2115 Programming with Grammars

Authors: Peter M. Maurer Maurer

Abstract:

DGL is a context free grammar-based tool for generating random data. Many types of simulator input data require some computation to be placed in the proper format. For example, it might be necessary to generate ordered triples in which the third element is the sum of the first two elements, or it might be necessary to generate random numbers in some sorted order. Although DGL is universal in computational power, generating these types of data is extremely difficult. To overcome this problem, we have enhanced DGL to include features that permit direct computation within the structure of a context free grammar. The features have been implemented as special types of productions, preserving the context free flavor of DGL specifications.

Keywords: DGL, Enhanced Context Free Grammars, Programming Constructs, Random Data Generation

Procedia PDF Downloads 118
2114 Reliability Analysis of Construction Schedule Plan Based on Building Information Modelling

Authors: Lu Ren, You-Liang Fang, Yan-Gang Zhao

Abstract:

In recent years, the application of BIM (Building Information Modelling) to construction schedule plan has been the focus of more and more researchers. In order to assess the reasonable level of the BIM-based construction schedule plan, that is whether the schedule can be completed on time, some researchers have introduced reliability theory to evaluate. In the process of evaluation, the uncertain factors affecting the construction schedule plan are regarded as random variables, and probability distributions of the random variables are assumed to be normal distribution, which is determined using two parameters evaluated from the mean and standard deviation of statistical data. However, in practical engineering, most of the uncertain influence factors are not normal random variables. So the evaluation results of the construction schedule plan will be unreasonable under the assumption that probability distributions of random variables submitted to the normal distribution. Therefore, in order to get a more reasonable evaluation result, it is necessary to describe the distribution of random variables more comprehensively. For this purpose, cubic normal distribution is introduced in this paper to describe the distribution of arbitrary random variables, which is determined by the first four moments (mean, standard deviation, skewness and kurtosis). In this paper, building the BIM model firstly according to the design messages of the structure and making the construction schedule plan based on BIM, then the cubic normal distribution is used to describe the distribution of the random variables due to the collecting statistical data of the random factors influencing construction schedule plan. Next the reliability analysis of the construction schedule plan based on BIM can be carried out more reasonably. Finally, the more accurate evaluation results can be given providing reference for the implementation of the actual construction schedule plan. In the last part of this paper, the more efficiency and accuracy of the proposed methodology for the reliability analysis of the construction schedule plan based on BIM are conducted through practical engineering case.

Keywords: BIM, construction schedule plan, cubic normal distribution, reliability analysis

Procedia PDF Downloads 108
2113 A New Concept for Deriving the Expected Value of Fuzzy Random Variables

Authors: Liang-Hsuan Chen, Chia-Jung Chang

Abstract:

Fuzzy random variables have been introduced as an imprecise concept of numeric values for characterizing the imprecise knowledge. The descriptive parameters can be used to describe the primary features of a set of fuzzy random observations. In fuzzy environments, the expected values are usually represented as fuzzy-valued, interval-valued or numeric-valued descriptive parameters using various metrics. Instead of the concept of area metric that is usually adopted in the relevant studies, the numeric expected value is proposed by the concept of distance metric in this study based on two characters (fuzziness and randomness) of FRVs. Comparing with the existing measures, although the results show that the proposed numeric expected value is same with those using the different metric, if only triangular membership functions are used. However, the proposed approach has the advantages of intuitiveness and computational efficiency, when the membership functions are not triangular types. An example with three datasets is provided for verifying the proposed approach.

Keywords: fuzzy random variables, distance measure, expected value, descriptive parameters

Procedia PDF Downloads 314
2112 Clustering Based Level Set Evaluation for Low Contrast Images

Authors: Bikshalu Kalagadda, Srikanth Rangu

Abstract:

The important object of images segmentation is to extract objects with respect to some input features. One of the important methods for image segmentation is Level set method. Generally medical images and synthetic images with low contrast of pixel profile, for such images difficult to locate interested features in images. In conventional level set function, develops irregularity during its process of evaluation of contour of objects, this destroy the stability of evolution process. For this problem a remedy is proposed, a new hybrid algorithm is Clustering Level Set Evolution. Kernel fuzzy particles swarm optimization clustering with the Distance Regularized Level Set (DRLS) and Selective Binary, and Gaussian Filtering Regularized Level Set (SBGFRLS) methods are used. The ability of identifying different regions becomes easy with improved speed. Efficiency of the modified method can be evaluated by comparing with the previous method for similar specifications. Comparison can be carried out by considering medical and synthetic images.

Keywords: segmentation, clustering, level set function, re-initialization, Kernel fuzzy, swarm optimization

Procedia PDF Downloads 327
2111 Radio Frequency Identification Encryption via Modified Two Dimensional Logistic Map

Authors: Hongmin Deng, Qionghua Wang

Abstract:

A modified two dimensional (2D) logistic map based on cross feedback control is proposed. This 2D map exhibits more random chaotic dynamical properties than the classic one dimensional (1D) logistic map in the statistical characteristics analysis. So it is utilized as the pseudo-random (PN) sequence generator, where the obtained real-valued PN sequence is quantized at first, then applied to radio frequency identification (RFID) communication system in this paper. This system is experimentally validated on a cortex-M0 development board, which shows the effectiveness in key generation, the size of key space and security. At last, further cryptanalysis is studied through the test suite in the National Institute of Standards and Technology (NIST).

Keywords: chaos encryption, logistic map, pseudo-random sequence, RFID

Procedia PDF Downloads 376
2110 [Keynote Speech]: Feature Selection and Predictive Modeling of Housing Data Using Random Forest

Authors: Bharatendra Rai

Abstract:

Predictive data analysis and modeling involving machine learning techniques become challenging in presence of too many explanatory variables or features. Presence of too many features in machine learning is known to not only cause algorithms to slow down, but they can also lead to decrease in model prediction accuracy. This study involves housing dataset with 79 quantitative and qualitative features that describe various aspects people consider while buying a new house. Boruta algorithm that supports feature selection using a wrapper approach build around random forest is used in this study. This feature selection process leads to 49 confirmed features which are then used for developing predictive random forest models. The study also explores five different data partitioning ratios and their impact on model accuracy are captured using coefficient of determination (r-square) and root mean square error (rsme).

Keywords: housing data, feature selection, random forest, Boruta algorithm, root mean square error

Procedia PDF Downloads 287
2109 Solving Process Planning, Weighted Apparent Tardiness Cost Dispatching, and Weighted Processing plus Weight Due-Date Assignment Simultaneously Using a Hybrid Search

Authors: Halil Ibrahim Demir, Caner Erden, Abdullah Hulusi Kokcam, Mumtaz Ipek

Abstract:

Process planning, scheduling, and due date assignment are three important manufacturing functions which are studied independently in literature. There are hundreds of works on IPPS and SWDDA problems but a few works on IPPSDDA problem. Integrating these three functions is very crucial due to the high relationship between them. Since the scheduling problem is in the NP-Hard problem class without any integration, an integrated problem is even harder to solve. This study focuses on the integration of these functions. Sum of weighted tardiness, earliness, and due date related costs are used as a penalty function. Random search and hybrid metaheuristics are used to solve the integrated problem. Marginal improvement in random search is very high in the early iterations and reduces enormously in later iterations. At that point directed search contribute to marginal improvement more than random search. In this study, random and genetic search methods are combined to find better solutions. Results show that overall performance becomes better as the integration level increases.

Keywords: process planning, genetic algorithm, hybrid search, random search, weighted due-date assignment, weighted scheduling

Procedia PDF Downloads 337
2108 Long Term Love Relationships Analyzed as a Dynamic System with Random Variations

Authors: Nini Johana Marín Rodríguez, William Fernando Oquendo Patino

Abstract:

In this work, we model a coupled system where we explore the effects of steady and random behavior on a linear system like an extension of the classic Strogatz model. This is exemplified by modeling a couple love dynamics as a linear system of two coupled differential equations and studying its stability for four types of lovers chosen as CC='Cautious- Cautious', OO='Only other feelings', OP='Opposites' and RR='Romeo the Robot'. We explore the effects of, first, introducing saturation, and second, adding a random variation to one of the CC-type lover, which will shape his character by trying to model how its variability influences the dynamics between love and hate in couple in a long run relationship. This work could also be useful to model other kind of systems where interactions can be modeled as linear systems with external or internal random influence. We found the final results are not easy to predict and a strong dependence on initial conditions appear, which a signature of chaos.

Keywords: differential equations, dynamical systems, linear system, love dynamics

Procedia PDF Downloads 324
2107 Multiclass Support Vector Machines with Simultaneous Multi-Factors Optimization for Corporate Credit Ratings

Authors: Hyunchul Ahn, William X. S. Wong

Abstract:

Corporate credit rating prediction is one of the most important topics, which has been studied by researchers in the last decade. Over the last decade, researchers are pushing the limit to enhance the exactness of the corporate credit rating prediction model by applying several data-driven tools including statistical and artificial intelligence methods. Among them, multiclass support vector machine (MSVM) has been widely applied due to its good predictability. However, heuristics, for example, parameters of a kernel function, appropriate feature and instance subset, has become the main reason for the critics on MSVM, as they have dictate the MSVM architectural variables. This study presents a hybrid MSVM model that is intended to optimize all the parameter such as feature selection, instance selection, and kernel parameter. Our model adopts genetic algorithm (GA) to simultaneously optimize multiple heterogeneous design factors of MSVM.

Keywords: corporate credit rating prediction, Feature selection, genetic algorithms, instance selection, multiclass support vector machines

Procedia PDF Downloads 265
2106 Facial Expression Recognition Using Sparse Gaussian Conditional Random Field

Authors: Mohammadamin Abbasnejad

Abstract:

The analysis of expression and facial Action Units (AUs) detection are very important tasks in fields of computer vision and Human Computer Interaction (HCI) due to the wide range of applications in human life. Many works have been done during the past few years which has their own advantages and disadvantages. In this work, we present a new model based on Gaussian Conditional Random Field. We solve our objective problem using ADMM and we show how well the proposed model works. We train and test our work on two facial expression datasets, CK+, and RU-FACS. Experimental evaluation shows that our proposed approach outperform state of the art expression recognition.

Keywords: Gaussian Conditional Random Field, ADMM, convergence, gradient descent

Procedia PDF Downloads 325
2105 Credit Risk Prediction Based on Bayesian Estimation of Logistic Regression Model with Random Effects

Authors: Sami Mestiri, Abdeljelil Farhat

Abstract:

The aim of this current paper is to predict the credit risk of banks in Tunisia, over the period (2000-2005). For this purpose, two methods for the estimation of the logistic regression model with random effects: Penalized Quasi Likelihood (PQL) method and Gibbs Sampler algorithm are applied. By using the information on a sample of 528 Tunisian firms and 26 financial ratios, we show that Bayesian approach improves the quality of model predictions in terms of good classification as well as by the ROC curve result.

Keywords: forecasting, credit risk, Penalized Quasi Likelihood, Gibbs Sampler, logistic regression with random effects, curve ROC

Procedia PDF Downloads 510