Search results for: probabilistic inference
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 599

Search results for: probabilistic inference

239 Deep Neural Networks for Restoration of Sky Images Affected by Static and Anisotropic Aberrations

Authors: Constanza A. Barriga, Rafael Bernardi, Amokrane Berdja, Christian D. Guzman

Abstract:

Most image restoration methods in astronomy rely upon probabilistic tools that infer the best solution for a deconvolution problem. They achieve good performances when the point spread function (PSF) is spatially invariable in the image plane. However, this latter condition is not always satisfied with real optical systems. PSF angular variations cannot be evaluated directly from the observations, neither be corrected at a pixel resolution. We have developed a method for the restoration of images affected by static and anisotropic aberrations using deep neural networks that can be directly applied to sky images. The network is trained using simulated sky images corresponding to the T-80 telescope optical system, an 80 cm survey imager at Cerro Tololo (Chile), which are synthesized using a Zernike polynomial representation of the optical system. Once trained, the network can be used directly on sky images, outputting a corrected version of the image, which has a constant and known PSF across its field-of-view. The method was tested with the T-80 telescope, achieving better results than with PSF deconvolution techniques. We present the method and results on this telescope.

Keywords: aberrations, deep neural networks, image restoration, variable point spread function, wide field images

Procedia PDF Downloads 112
238 A Theorem Related to Sample Moments and Two Types of Moment-Based Density Estimates

Authors: Serge B. Provost

Abstract:

Numerous statistical inference and modeling methodologies are based on sample moments rather than the actual observations. A result justifying the validity of this approach is introduced. More specifically, it will be established that given the first n moments of a sample of size n, one can recover the original n sample points. This implies that a sample of size n and its first associated n moments contain precisely the same amount of information. However, it is efficient to make use of a limited number of initial moments as most of the relevant distributional information is included in them. Two types of density estimation techniques that rely on such moments will be discussed. The first one expresses a density estimate as the product of a suitable base density and a polynomial adjustment whose coefficients are determined by equating the moments of the density estimate to the sample moments. The second one assumes that the derivative of the logarithm of a density function can be represented as a rational function. This gives rise to a system of linear equations involving sample moments, the density estimate is then obtained by solving a differential equation. Unlike kernel density estimation, these methodologies are ideally suited to model ‘big data’ as they only require a limited number of moments, irrespective of the sample size. What is more, they produce simple closed form expressions that are amenable to algebraic manipulations. They also turn out to be more accurate as will be shown in several illustrative examples.

Keywords: density estimation, log-density, polynomial adjustments, sample moments

Procedia PDF Downloads 137
237 Spatial Organization of Organelles in Living Cells: Insights from Mathematical Modelling

Authors: Congping Lin

Abstract:

Intracellular transport in fungi has a number of important roles in, e.g., filamentous fungal growth and cellular metabolism. Two basic mechanisms for intracellular transport are motor-driven trafficking along microtubules (MTs) and diffusion. Mathematical modelling has been actively developed to understand such intracellular transport and provide unique insight into cellular complexity. Based on live-cell imaging data in Ustilago hyphal cells, probabilistic models have been developed to study mechanism underlying spatial organization of molecular motors and organelles. In particular, anther mechanism - stochastic motility of dynein motors along MTs has been found to contribute to half of its accumulation at hyphal tip in order to support early endosome (EE) recycling. The EE trafficking not only facilitates the directed motion of peroxisomes but also enhances their diffusive motion. Considering the importance of spatial organization of early endosomes in supporting peroxisome movement, computational and experimental approaches have been combined to a whole-cell level. Results from this interdisciplinary study promise insights into requirements for other membrane trafficking systems (e.g., in neurons), but also may inform future 'synthetic biology' studies.

Keywords: intracellular transport, stochastic process, molecular motors, spatial organization

Procedia PDF Downloads 107
236 Intelligent System and Renewable Energy: A Farming Platform in Precision Agriculture

Authors: Ryan B. Escorial, Elmer A. Maravillas, Chris Jordan G. Aliac

Abstract:

This study presents a small-scale water pumping system utilizing a fuzzy logic inference system attached to a renewable energy source. The fuzzy logic controller was designed and simulated in MATLAB fuzzy logic toolbox to examine the properties and characteristics of the input and output variables. The result of the simulation was implemented in a microcontroller, together with sensors, modules, and photovoltaic cells. The study used a grand rapid variety of lettuce, organic substrates, and foliar for observation of the capability of the device to irrigate crops. Two plant boxes intended for manual and automated irrigation were prepared with each box having 48 heads of lettuce. The observation of the system took 22-31 days, which is one harvest period of the crop. Results showed a 22.55% increase in agricultural productivity compared to manual irrigation. Aside from reducing human effort, and time, the smart irrigation system could help lessen some of the shortcomings of manual irrigations. It could facilitate the economical utilization of water, reducing consumption by 25%. The use of renewable energy could also help farmers reduce the cost of production by minimizing the use of diesel and gasoline.

Keywords: fuzzy logic, intelligent system, precision agriculture, renewable energy

Procedia PDF Downloads 101
235 Detection and Classification of Myocardial Infarction Using New Extracted Features from Standard 12-Lead ECG Signals

Authors: Naser Safdarian, Nader Jafarnia Dabanloo

Abstract:

In this paper we used four features i.e. Q-wave integral, QRS complex integral, T-wave integral and total integral as extracted feature from normal and patient ECG signals to detection and localization of myocardial infarction (MI) in left ventricle of heart. In our research we focused on detection and localization of MI in standard ECG. We use the Q-wave integral and T-wave integral because this feature is important impression in detection of MI. We used some pattern recognition method such as Artificial Neural Network (ANN) to detect and localize the MI. Because these methods have good accuracy for classification of normal and abnormal signals. We used one type of Radial Basis Function (RBF) that called Probabilistic Neural Network (PNN) because of its nonlinearity property, and used other classifier such as k-Nearest Neighbors (KNN), Multilayer Perceptron (MLP) and Naive Bayes Classification. We used PhysioNet database as our training and test data. We reached over 80% for accuracy in test data for localization and over 95% for detection of MI. Main advantages of our method are simplicity and its good accuracy. Also we can improve accuracy of classification by adding more features in this method. A simple method based on using only four features which extracted from standard ECG is presented which has good accuracy in MI localization.

Keywords: ECG signal processing, myocardial infarction, features extraction, pattern recognition

Procedia PDF Downloads 433
234 Extraction of Road Edge Lines from High-Resolution Remote Sensing Images Based on Energy Function and Snake Model

Authors: Zuoji Huang, Haiming Qian, Chunlin Wang, Jinyan Sun, Nan Xu

Abstract:

In this paper, the strategy to extract double road edge lines from acquired road stripe image was explored. The workflow is as follows: the road stripes are acquired by probabilistic boosting tree algorithm and morphological algorithm immediately, and road centerlines are detected by thinning algorithm, so the initial road edge lines can be acquired along the road centerlines. Then we refine the results with big variation of local curvature of centerlines. Specifically, the energy function of edge line is constructed by gradient feature and spectral information, and Dijkstra algorithm is used to optimize the initial road edge lines. The Snake model is constructed to solve the fracture problem of intersection, and the discrete dynamic programming algorithm is used to solve the model. After that, we could get the final road network. Experiment results show that the strategy proposed in this paper can be used to extract the continuous and smooth road edge lines from high-resolution remote sensing images with an accuracy of 88% in our study area.

Keywords: road edge lines extraction, energy function, intersection fracture, Snake model

Procedia PDF Downloads 320
233 Progressive Type-I Interval Censoring with Binomial Removal-Estimation and Its Properties

Authors: Sonal Budhiraja, Biswabrata Pradhan

Abstract:

This work considers statistical inference based on progressive Type-I interval censored data with random removal. The scheme of progressive Type-I interval censoring with random removal can be described as follows. Suppose n identical items are placed on a test at time T0 = 0 under k pre-fixed inspection times at pre-specified times T1 < T2 < . . . < Tk, where Tk is the scheduled termination time of the experiment. At inspection time Ti, Ri of the remaining surviving units Si, are randomly removed from the experiment. The removal follows a binomial distribution with parameters Si and pi for i = 1, . . . , k, with pk = 1. In this censoring scheme, the number of failures in different inspection intervals and the number of randomly removed items at pre-specified inspection times are observed. Asymptotic properties of the maximum likelihood estimators (MLEs) are established under some regularity conditions. A β-content γ-level tolerance interval (TI) is determined for two parameters Weibull lifetime model using the asymptotic properties of MLEs. The minimum sample size required to achieve the desired β-content γ-level TI is determined. The performance of the MLEs and TI is studied via simulation.

Keywords: asymptotic normality, consistency, regularity conditions, simulation study, tolerance interval

Procedia PDF Downloads 221
232 Landslide Susceptibility Mapping: A Comparison between Logistic Regression and Multivariate Adaptive Regression Spline Models in the Municipality of Oudka, Northern of Morocco

Authors: S. Benchelha, H. C. Aoudjehane, M. Hakdaoui, R. El Hamdouni, H. Mansouri, T. Benchelha, M. Layelmam, M. Alaoui

Abstract:

The logistic regression (LR) and multivariate adaptive regression spline (MarSpline) are applied and verified for analysis of landslide susceptibility map in Oudka, Morocco, using geographical information system. From spatial database containing data such as landslide mapping, topography, soil, hydrology and lithology, the eight factors related to landslides such as elevation, slope, aspect, distance to streams, distance to road, distance to faults, lithology map and Normalized Difference Vegetation Index (NDVI) were calculated or extracted. Using these factors, landslide susceptibility indexes were calculated by the two mentioned methods. Before the calculation, this database was divided into two parts, the first for the formation of the model and the second for the validation. The results of the landslide susceptibility analysis were verified using success and prediction rates to evaluate the quality of these probabilistic models. The result of this verification was that the MarSpline model is the best model with a success rate (AUC = 0.963) and a prediction rate (AUC = 0.951) higher than the LR model (success rate AUC = 0.918, rate prediction AUC = 0.901).

Keywords: landslide susceptibility mapping, regression logistic, multivariate adaptive regression spline, Oudka, Taounate

Procedia PDF Downloads 164
231 Inference for Compound Truncated Poisson Lognormal Model with Application to Maximum Precipitation Data

Authors: M. Z. Raqab, Debasis Kundu, M. A. Meraou

Abstract:

In this paper, we have analyzed maximum precipitation data during a particular period of time obtained from different stations in the Global Historical Climatological Network of the USA. One important point to mention is that some stations are shut down on certain days for some reason or the other. Hence, the maximum values are recorded by excluding those readings. It is assumed that the number of stations that operate follows zero-truncated Poisson random variables, and the daily precipitation follows a lognormal random variable. We call this model a compound truncated Poisson lognormal model. The proposed model has three unknown parameters, and it can take a variety of shapes. The maximum likelihood estimators can be obtained quite conveniently using Expectation-Maximization (EM) algorithm. Approximate maximum likelihood estimators are also derived. The associated confidence intervals also can be obtained from the observed Fisher information matrix. Simulation results have been performed to check the performance of the EM algorithm, and it is observed that the EM algorithm works quite well in this case. When we analyze the precipitation data set using the proposed model, it is observed that the proposed model provides a better fit than some of the existing models.

Keywords: compound Poisson lognormal distribution, EM algorithm, maximum likelihood estimation, approximate maximum likelihood estimation, Fisher information, skew distribution

Procedia PDF Downloads 86
230 Subsurface Elastic Properties Determination for Site Characterization Using Seismic Refraction Tomography at the Pwalugu Dam Area

Authors: Van-Dycke Sarpong Asare, Vincent Adongo

Abstract:

Field measurement of subsurface seismic p-wave velocities was undertaken through seismic refraction tomography. The aim of this work is to obtain a model of the shallow subsurface material elastic properties relevant for geotechnical site characterization. The survey area is at Pwalugu in Northern Ghana, where a multipurpose dam, for electricity generation, irrigation, and potable water delivery, is being planned. A 24-channel seismograph and 24, 10 Hz electromagnetic geophones, deployed 5 m apart constituted the acquisition hardware. Eleven (2-D) seismic refraction profiles, nine of which ran almost perpendicular and two parallel to the White Volta at Pwalugu, were acquired. The refraction tomograms of the thirteen profiles revealed a subsurface model consisting of one minor and one major acoustic impedance boundaries – the top dry/loose sand and the variably weathered sandstone contact, and the overburden-sandstones bedrock contact respectively. The p-wave velocities and by inference, with a priori values of poison ratios, the s-wave velocities, assisted in characterizing the geotechnical conditions of the proposed site and also in evaluating the dynamic properties such as the maximum shear modulus, the bulk modulus, and the Young modulus.

Keywords: tomography, characterization, consolidated, Pwalugu and seismograph

Procedia PDF Downloads 105
229 Effects of Applying Low-Dye Taping in Performing Double-Leg Squat on Electromyographic Activity of Lower Extremity Muscles for Collegiate Basketball Players with Excessive Foot Pronation

Authors: I. M. K. Ho, S. K. Y. Chan, K. H. P. Lam, G. M. W. Tong, N. C. Y. Yeung, J. T. C. Luk

Abstract:

Low-dye taping (LDT) is commonly used for treating foot problems, such as plantar fasciitis, and supporting foot arch for runners and non-athletes patients with pes planus. The potential negative impact of pronated feet leading to tibial and femoral internal rotation via the entire kinetic chain reaction was postulated and identified. The changed lower limb biomechanics potentially leading to poor activation of hip and knee stabilizers, such as gluteus maximus and medius, may associate with higher risk of knee injuries including patellofemoral pain syndrome and ligamentous sprain in many team sports players. It is therefore speculated that foot arch correction with LDT might enhance the use of gluteal muscles. The purpose of this study was to investigate the effect of applying LDT on surface electromyographic (sEMG) activity of superior gluteus maximus (SGMax), inferior gluteus maximus (IGMax), gluteus medius (GMed) and tibialis anterior (TA) during double-leg squat. 12 male collegiate basketball players (age: 21.72.5 years; body fat: 12.43.6%; navicular drop: 13.72.7mm) with at least three years regular basketball training experience participated in this study. Participants were excluded if they had recent history of lower limb injuries, over 16.6% body fat and lesser than 10mm drop in navicular drop (ND) test. Recruited subjects visited the laboratory once for the within-subject crossover study. Maximum voluntary isometric contraction (MVIC) tests on all selected muscles were performed in randomized order followed by sEMG test on double-leg squat during LDT and non-LDT conditions in counterbalanced order. SGMax, IGMax, GMed and TA activities during the entire 2-second concentric and 2-second eccentric phases were normalized and interpreted as %MVIC. The magnitude of the difference between taped and non-taped conditions of each muscle was further assessed via standardized effect90% confidence intervals (CI) with non-clinical magnitude-based inference. Paired samples T-test showed a significant decrease (4.71.4mm) in ND (95% CI: 3.8, 5.6; p < 0.05) while no significant difference was observed between taped and non-taped conditions in sEMG tests for all muscles and contractions (p > 0.05). On top of traditional significant testing, magnitude-based inference showed possibly increase in IGMax activity (small standardized effect: 0.270.44), likely increase in GMed activity (small standardized effect: 0.340.34) and possibly increase in TA activity (small standardized effect: 0.220.29) during eccentric phase. It is speculated that the decrease of navicular drop supported by LDT application could potentially enhance the use of inferior gluteus maximus and gluteus medius especially during eccentric phase in this study. As the eccentric phase of double-leg squat is an important component of landing activities in basketball, further studies on the onset and amount of gluteal activation during jumping and landing activities with LDT are recommended. Since both hip and knee kinematics were not measured in this study, the underlying cause of the observed increase in gluteal activation during squat after LDT is inconclusive. In this regard, the investigation of relationships between LDT application, ND, hip and knee kinematics, and gluteal muscle activity during sports specific jumping and landing tasks should be focused in the future.

Keywords: flat foot, gluteus maximus, gluteus medius, injury prevention

Procedia PDF Downloads 135
228 An Approach for Estimation in Hierarchical Clustered Data Applicable to Rare Diseases

Authors: Daniel C. Bonzo

Abstract:

Practical considerations lead to the use of unit of analysis within subjects, e.g., bleeding episodes or treatment-related adverse events, in rare disease settings. This is coupled with data augmentation techniques such as extrapolation to enlarge the subject base. In general, one can think about extrapolation of data as extending information and conclusions from one estimand to another estimand. This approach induces hierarchichal clustered data with varying cluster sizes. Extrapolation of clinical trial data is being accepted increasingly by regulatory agencies as a means of generating data in diverse situations during drug development process. Under certain circumstances, data can be extrapolated to a different population, a different but related indication, and different but similar product. We consider here the problem of estimation (point and interval) using a mixed-models approach under an extrapolation. It is proposed that estimators (point and interval) be constructed using weighting schemes for the clusters, e.g., equally weighted and with weights proportional to cluster size. Simulated data generated under varying scenarios are then used to evaluate the performance of this approach. In conclusion, the evaluation result showed that the approach is a useful means for improving statistical inference in rare disease settings and thus aids not only signal detection but risk-benefit evaluation as well.

Keywords: clustered data, estimand, extrapolation, mixed model

Procedia PDF Downloads 112
227 Estimation of Transition and Emission Probabilities

Authors: Aakansha Gupta, Neha Vadnere, Tapasvi Soni, M. Anbarsi

Abstract:

Protein secondary structure prediction is one of the most important goals pursued by bioinformatics and theoretical chemistry; it is highly important in medicine and biotechnology. Some aspects of protein functions and genome analysis can be predicted by secondary structure prediction. This is used to help annotate sequences, classify proteins, identify domains, and recognize functional motifs. In this paper, we represent protein secondary structure as a mathematical model. To extract and predict the protein secondary structure from the primary structure, we require a set of parameters. Any constants appearing in the model are specified by these parameters, which also provide a mechanism for efficient and accurate use of data. To estimate these model parameters there are many algorithms out of which the most popular one is the EM algorithm or called the Expectation Maximization Algorithm. These model parameters are estimated with the use of protein datasets like RS126 by using the Bayesian Probabilistic method (data set being categorical). This paper can then be extended into comparing the efficiency of EM algorithm to the other algorithms for estimating the model parameters, which will in turn lead to an efficient component for the Protein Secondary Structure Prediction. Further this paper provides a scope to use these parameters for predicting secondary structure of proteins using machine learning techniques like neural networks and fuzzy logic. The ultimate objective will be to obtain greater accuracy better than the previously achieved.

Keywords: model parameters, expectation maximization algorithm, protein secondary structure prediction, bioinformatics

Procedia PDF Downloads 451
226 FLIME - Fast Low Light Image Enhancement for Real-Time Video

Authors: Vinay P., Srinivas K. S.

Abstract:

Low Light Image Enhancement is of utmost impor- tance in computer vision based tasks. Applications include vision systems for autonomous driving, night vision devices for defence systems, low light object detection tasks. Many of the existing deep learning methods are resource intensive during the inference step and take considerable time for processing. The algorithm should take considerably less than 41 milliseconds in order to process a real-time video feed with 24 frames per second and should be even less for a video with 30 or 60 frames per second. The paper presents a fast and efficient solution which has two main advantages, it has the potential to be used for a real-time video feed, and it can be used in low compute environments because of the lightweight nature. The proposed solution is a pipeline of three steps, the first one is the use of a simple function to map input RGB values to output RGB values, the second is to balance the colors and the final step is to adjust the contrast of the image. Hence a custom dataset is carefully prepared using images taken in low and bright lighting conditions. The preparation of the dataset, the proposed model, the processing time are discussed in detail and the quality of the enhanced images using different methods is shown.

Keywords: low light image enhancement, real-time video, computer vision, machine learning

Procedia PDF Downloads 171
225 Hybrid Structure Learning Approach for Assessing the Phosphate Laundries Impact

Authors: Emna Benmohamed, Hela Ltifi, Mounir Ben Ayed

Abstract:

Bayesian Network (BN) is one of the most efficient classification methods. It is widely used in several fields (i.e., medical diagnostics, risk analysis, bioinformatics research). The BN is defined as a probabilistic graphical model that represents a formalism for reasoning under uncertainty. This classification method has a high-performance rate in the extraction of new knowledge from data. The construction of this model consists of two phases for structure learning and parameter learning. For solving this problem, the K2 algorithm is one of the representative data-driven algorithms, which is based on score and search approach. In addition, the integration of the expert's knowledge in the structure learning process allows the obtainment of the highest accuracy. In this paper, we propose a hybrid approach combining the improvement of the K2 algorithm called K2 algorithm for Parents and Children search (K2PC) and the expert-driven method for learning the structure of BN. The evaluation of the experimental results, using the well-known benchmarks, proves that our K2PC algorithm has better performance in terms of correct structure detection. The real application of our model shows its efficiency in the analysis of the phosphate laundry effluents' impact on the watershed in the Gafsa area (southwestern Tunisia).

Keywords: Bayesian network, classification, expert knowledge, structure learning, surface water analysis

Procedia PDF Downloads 100
224 An Insite to the Probabilistic Assessment of Reserves in Conventional Reservoirs

Authors: Sai Sudarshan, Harsh Vyas, Riddhiman Sherlekar

Abstract:

The oil and gas industry has been unwilling to adopt stochastic definition of reserves. Nevertheless, Monte Carlo simulation methods have gained acceptance by engineers, geoscientists and other professionals who want to evaluate prospects or otherwise analyze problems that involve uncertainty. One of the common applications of Monte Carlo simulation is the estimation of recoverable hydrocarbon from a reservoir.Monte Carlo Simulation makes use of random samples of parameters or inputs to explore the behavior of a complex system or process. It finds application whenever one needs to make an estimate, forecast or decision where there is significant uncertainty. First, the project focuses on performing Monte-Carlo Simulation on a given data set using U. S Department of Energy’s MonteCarlo Software, which is a freeware e&p tool. Further, an algorithm for simulation has been developed for MATLAB and program performs simulation by prompting user for input distributions and parameters associated with each distribution (i.e. mean, st.dev, min., max., most likely, etc.). It also prompts user for desired probability for which reserves are to be calculated. The algorithm so developed and tested in MATLAB further finds implementation in Python where existing libraries on statistics and graph plotting have been imported to generate better outcome. With PyQt designer, codes for a simple graphical user interface have also been written. The graph so plotted is then validated with already available results from U.S DOE MonteCarlo Software.

Keywords: simulation, probability, confidence interval, sensitivity analysis

Procedia PDF Downloads 352
223 Dynamic Route Optimization in Vehicle Adhoc Networks: A Heuristics Routing Protocol

Authors: Rafi Ullah, Shah Muhammad Emaduddin, Taha Jilani

Abstract:

Vehicle Adhoc Networks (VANET) belongs to a special class of Mobile Adhoc Network (MANET) with high mobility. Network is created by road side vehicles equipped with communication devices like GPS and Wifi etc. Since the environment is highly dynamic due to difference in speed and high mobility of vehicles and weak stability of the network connection, it is a challenging task to design an efficient routing protocol for such an unstable environment. Our proposed algorithm uses heuristic for the calculation of optimal path for routing the packet efficiently in collaboration with several other parameters like geographical location, speed, priority, the distance among the vehicles, communication range, and networks congestion. We have incorporated probabilistic, heuristic and machine learning based approach inconsistency with the relay function of the memory buffer to keep the packet moving towards the destination. These parameters when used in collaboration provide us a very strong and admissible heuristics. We have mathematically proved that the proposed technique is efficient for the routing of packets, especially in a medical emergency situation. These networks can be used for medical emergency, security, entertainment and routing purposes.

Keywords: heuristics routing, intelligent routing, VANET, route optimization

Procedia PDF Downloads 154
222 Parameter Estimation for Contact Tracing in Graph-Based Models

Authors: Augustine Okolie, Johannes Müller, Mirjam Kretzchmar

Abstract:

We adopt a maximum-likelihood framework to estimate parameters of a stochastic susceptible-infected-recovered (SIR) model with contact tracing on a rooted random tree. Given the number of detectees per index case, our estimator allows to determine the degree distribution of the random tree as well as the tracing probability. Since we do not discover all infectees via contact tracing, this estimation is non-trivial. To keep things simple and stable, we develop an approximation suited for realistic situations (contract tracing probability small, or the probability for the detection of index cases small). In this approximation, the only epidemiological parameter entering the estimator is the basic reproduction number R0. The estimator is tested in a simulation study and applied to covid-19 contact tracing data from India. The simulation study underlines the efficiency of the method. For the empirical covid-19 data, we are able to compare different degree distributions and perform a sensitivity analysis. We find that particularly a power-law and a negative binomial degree distribution meet the data well and that the tracing probability is rather large. The sensitivity analysis shows no strong dependency on the reproduction number.

Keywords: stochastic SIR model on graph, contact tracing, branching process, parameter inference

Procedia PDF Downloads 57
221 Exploring the Activity Fabric of an Intelligent Environment with Hierarchical Hidden Markov Theory

Authors: Chiung-Hui Chen

Abstract:

The Internet of Things (IoT) was designed for widespread convenience. With the smart tag and the sensing network, a large quantity of dynamic information is immediately presented in the IoT. Through the internal communication and interaction, meaningful objects provide real-time services for users. Therefore, the service with appropriate decision-making has become an essential issue. Based on the science of human behavior, this study employed the environment model to record the time sequences and locations of different behaviors and adopted the probability module of the hierarchical Hidden Markov Model for the inference. The statistical analysis was conducted to achieve the following objectives: First, define user behaviors and predict the user behavior routes with the environment model to analyze user purposes. Second, construct the hierarchical Hidden Markov Model according to the logic framework, and establish the sequential intensity among behaviors to get acquainted with the use and activity fabric of the intelligent environment. Third, establish the intensity of the relation between the probability of objects’ being used and the objects. The indicator can describe the possible limitations of the mechanism. As the process is recorded in the information of the system created in this study, these data can be reused to adjust the procedure of intelligent design services.

Keywords: behavior, big data, hierarchical hidden Markov model, intelligent object

Procedia PDF Downloads 211
220 Maximum-likelihood Inference of Multi-Finger Movements Using Neural Activities

Authors: Kyung-Jin You, Kiwon Rhee, Marc H. Schieber, Nitish V. Thakor, Hyun-Chool Shin

Abstract:

It remains unknown whether M1 neurons encode multi-finger movements independently or as a certain neural network of single finger movements although multi-finger movements are physically a combination of single finger movements. We present an evidence of correlation between single and multi-finger movements and also attempt a challenging task of semi-blind decoding of neural data with minimum training of the neural decoder. Data were collected from 115 task-related neurons in M1 of a trained rhesus monkey performing flexion and extension of each finger and the wrist (12 single and 6 two-finger-movements). By exploiting correlation of temporal firing pattern between movements, we found that correlation coefficient for physically related movements pairs is greater than others; neurons tuned to single finger movements increased their firing rate when multi-finger commands were instructed. According to this knowledge, neural semi-blind decoding is done by choosing the greatest and the second greatest likelihood for canonical candidates. We achieved a decoding accuracy about 60% for multiple finger movement without corresponding training data set. this results suggest that only with the neural activities on single finger movements can be exploited to control dexterous multi-fingered neuroprosthetics.

Keywords: finger movement, neural activity, blind decoding, M1

Procedia PDF Downloads 291
219 Seismic Hazard Analysis for a Multi Layer Fault System: Antalya (SW Turkey) Example

Authors: Nihat Dipova, Bulent Cangir

Abstract:

This article presents the results of probabilistic seismic hazard analysis (PSHA) for Antalya (SW Turkey). South west of Turkey is characterized by large earthquakes resulting from the continental collision between the African, Arabian and Eurasian plates and crustal faults. Earthquakes around the study area are grouped into two; crustal earthquakes (D=0-50 km) and subduction zone earthquakes (50-140 km). Maximum observed magnitude of subduction earthquakes is Mw=6.0. Maximum magnitude of crustal earthquakes is Mw=6.6. Sources for crustal earthquakes are faults which are related with Isparta Angle and Cyprus Arc tectonic structures. A new earthquake catalogue for Antalya, with unified moment magnitude scale has been prepared and seismicity of the area around Antalya city has been evaluated by defining ‘a’ and ‘b’ parameters of the Gutenberg-Richter recurrence relationship. The Standard Cornell-McGuire method has been used for hazard computation utilizing CRISIS2007 software. Attenuation relationships proposed by Chiou and Youngs (2008) has been used for 0-50 km earthquakes and Youngs et. al (1997) for deep subduction earthquakes. Finally, Seismic hazard map for peak horizontal acceleration on a uniform site condition of firm rock (average shear wave velocity of about 1130 m/s) at a hazard level of 10% probability of exceedance in 50 years has been prepared.

Keywords: Antalya, peak ground acceleration, seismic hazard assessment, subduction

Procedia PDF Downloads 352
218 Phylogenetic Relationships of Common Reef Fish Species in Vietnam

Authors: Dang Thuy Binh, Truong Thi Oanh, Le Phan Khanh Hung, Luong thi Tuong Vy

Abstract:

One of the greatest environmental challenges facing Asia is the management and conservation of the marine biodiversity threaten by fisheries overexploitation, pollution, habitat destruction, and climate change. To date, a few molecular taxonomical studies has been conducted on marine fauna in Vietnam. The purpose of this study was to clarify the phylogeny of economic and ecological reef fish species in Vietnam Reef fish species covering Labridae, Scaridae, Nemipteridae, Serranidae, Acanthuridae, Lutjanidae, Lethrinidae, Mullidae, Balistidae, Pseudochromidae, Pinguipedidae, Fistulariidae, Holocentridae, Synodontidae, and Pomacentridae representing 28 genera were collected from South and Center, Vietnam. Combine with Genbank sequences, a phylogenetic tree was constructed based on 16S gene of mitochondrial DNA using maximum parsimony, maximum likelihood, and Bayesian inference approaches. The phylogram showed the well-resolved clades at genus and family level. Perciformes is the major order of reef fish species in Vietnam. The monophyly of Perciformes is not strongly supported as it was clustered in the same clade with Tetraodontiformes syngnathiformes and Beryciformes. Continue sampling of commercial fish species and classification based on morphology and genetics to build DNA barcoding of fish species in Vietnam is really necessary.

Keywords: reef fish, 16s rDNA, Vietnam, phylogeny

Procedia PDF Downloads 411
217 A High Efficiency Reduced Rules Neuro-Fuzzy Based Maximum Power Point Tracking Controller for Photovoltaic Array Connected to Grid

Authors: Lotfi Farah, Nadir Farah, Zaiem Kamar

Abstract:

This paper achieves a maximum power point tracking (MPPT) controller using a high-efficiency reduced rules neuro-fuzzy inference system (HE2RNF) for a 100 kW stand-alone photovoltaic (PV) system connected to the grid. The suggested HE2RNF based MPPT seeks the optimal duty cycle for the boost DC-DC converter, making the designed PV system working at the maximum power point (MPP), then transferring this power to the grid via a three levels voltage source converter (VSC). PV current variation and voltage variation are chosen as HE2RNF-based MPPT controller inputs. By using these inputs with the duty cycle as the only single output, a six rules ANFIS is generated. The high performance of the proposed HE2RNF numerically in the MATLAB/Simulink environment is shown. The 0.006% steady-state error, 0.006s of tracking time, and 0.088s of starting time prove the robustness of this six reduced rules against the widely used twenty-five ones.

Keywords: PV, MPPT, ANFIS, HE2RNF-based MPPT controller, VSC, grid connection

Procedia PDF Downloads 157
216 The Effects of Three Levels of Contextual Inference among adult Athletes

Authors: Abdulaziz Almustafa

Abstract:

Considering the critical role permanence has on predictions related to the contextual interference effect on laboratory and field research, this study sought to determine whether the paradigm of the effect depends on the complexity of the skill during the acquisition and transfer phases. The purpose of the present study was to investigate the effects of contextual interference CI by extending previous laboratory and field research with adult athletes through the acquisition and transfer phases. Male (n=60) athletes age 18-22 years-old, were chosen randomly from Eastern Province Clubs. They were assigned to complete blocked, random, or serial practices. Analysis of variance with repeated measures MANOVA indicated that, the results did not support the notion of CI. There were no significant differences in acquisition phase between blocked, serial and random practice groups. During the transfer phase, there were no major differences between the practice groups. Apparently, due to the task complexity, participants were probably confused and not able to use the advantages of contextual interference. This is another contradictory result to contextual interference effects in acquisition and transfer phases in sport settings. One major factor that can influence the effect of contextual interference is task characteristics as the nature of level of difficulty in sport-related skill.

Keywords: contextual interference, acquisition, transfer, task difficulty

Procedia PDF Downloads 437
215 An Improved Convolution Deep Learning Model for Predicting Trip Mode Scheduling

Authors: Amin Nezarat, Naeime Seifadini

Abstract:

Trip mode selection is a behavioral characteristic of passengers with immense importance for travel demand analysis, transportation planning, and traffic management. Identification of trip mode distribution will allow transportation authorities to adopt appropriate strategies to reduce travel time, traffic and air pollution. The majority of existing trip mode inference models operate based on human selected features and traditional machine learning algorithms. However, human selected features are sensitive to changes in traffic and environmental conditions and susceptible to personal biases, which can make them inefficient. One way to overcome these problems is to use neural networks capable of extracting high-level features from raw input. In this study, the convolutional neural network (CNN) architecture is used to predict the trip mode distribution based on raw GPS trajectory data. The key innovation of this paper is the design of the layout of the input layer of CNN as well as normalization operation, in a way that is not only compatible with the CNN architecture but can also represent the fundamental features of motion including speed, acceleration, jerk, and Bearing rate. The highest prediction accuracy achieved with the proposed configuration for the convolutional neural network with batch normalization is 85.26%.

Keywords: predicting, deep learning, neural network, urban trip

Procedia PDF Downloads 111
214 Comparison of Seismic Response for Two RC Curved Bridges with Different Column Shapes

Authors: Nina N. Serdar, Jelena R. Pejović

Abstract:

This paper presents seismic risk assessment of two bridge structure, based on the probabilistic performance-based seismic assessment methodology. Both investigated bridges are tree span continuous RC curved bridges with the difference in column shapes. First bridge (type A) has a wall-type pier and second (type B) has a two-column bent with circular columns. Bridges are designed according to European standards: EN 1991-2, EN1992-1-1 and EN 1998-2. Aim of the performed analysis is to compare seismic behavior of these two structures and to detect the influence of column shapes on the seismic response. Seismic risk assessment is carried out by obtaining demand fragility curves. Non-linear model was constructed and time-history analysis was performed using thirty five pairs of horizontal ground motions selected to match site specific hazard. In performance based analysis, peak column drift ratio (CDR) was selected as engineering demand parameter (EDP). For seismic intensity measure (IM) spectral displacement was selected. Demand fragility curves that give probability of exceedance of certain value for chosen EDP were constructed and based on them conclusions were made.

Keywords: RC curved bridge, demand fragility curve, wall type column, nonlinear time-history analysis, circular column

Procedia PDF Downloads 313
213 Etiological Factors for Renal Cell Carcinoma: Five-Year Study at Mayo Hospital Lahore

Authors: Muhammad Umar Hassan

Abstract:

Renal cell carcinoma is a subset of kidney cancer that arises in the lining of DCT and is present in parenchymal tissue. Diagnosis is based on lab reports, including urinalysis, renal function tests (RFTs), and electrolyte balance, along with imaging techniques. Organ failure and other complications have been commonly observed in these cases. Over the years, the presentation of patients has varied, so carcinoma was classified on the basis of site, shape, and consistency for detailed analysis. Lifestyle patterns and occupational history were inquired about and recorded. Methods: Data from 100 patients presenting to the oncology and nephrology department of Mayo Hospital in the year 2015-2020 were included in this retrospective study on a random basis. The study was specifically focused on three risk factors. Smoking, occupational exposures, and Hakim medicine are taken by the patient for any cause. After procurement of data, follow-up contacts of these patients were established, resulting in a detailed analysis of lifestyle. Conclusion: The inference drawn is a direct causal link between smoking, industrial workplace exposure, and Hakim medicine with the development of Renal Cell Carcinoma. It was shown in the majority of the patients and hence confirmed our hypothesis.

Keywords: renal cell carcinoma, kidney cancer, clear cell carcinoma

Procedia PDF Downloads 75
212 An Ontology-Based Framework to Support Asset Integrity Modeling: Case Study of Offshore Riser Integrity

Authors: Mohammad Sheikhalishahi, Vahid Ebrahimipour, Amir Hossein Radman-Kian

Abstract:

This paper proposes an Ontology framework for knowledge modeling and representation of the equipment integrity process in a typical oil and gas production plant. Our aim is to construct a knowledge modeling that facilitates translation, interpretation, and conversion of human-readable integrity interpretation into computer-readable representation. The framework provides a function structure related to fault propagation using ISO 14224 and ISO 15926 OWL-Lite/ Resource Description Framework (RDF) to obtain a generic system-level model of asset integrity that can be utilized in the integrity engineering process during the equipment life cycle. It employs standard terminology developed by ISO 15926 and ISO 14224 to map textual descriptions of equipment failure and then convert it to a causality-driven logic by semantic interpretation and computer-based representation using Lite/RDF. The framework applied for an offshore gas riser. The result shows that the approach can cross-link the failure-related integrity words and domain-specific logic to obtain a representation structure of equipment integrity with causality inference based on semantic extraction of inspection report context.

Keywords: asset integrity modeling, interoperability, OWL, RDF/XML

Procedia PDF Downloads 160
211 Detecting and Thwarting Interest Flooding Attack in Information Centric Network

Authors: Vimala Rani P, Narasimha Malikarjunan, Mercy Shalinie S

Abstract:

Data Networking was brought forth as an instantiation of information-centric networking. The attackers can send a colossal number of spoofs to take hold of the Pending Interest Table (PIT) named an Interest Flooding attack (IFA) since the in- interests are recorded in the PITs of the intermediate routers until they receive corresponding Data Packets are go beyond the time limit. These attacks can be detrimental to network performance. PIT expiration rate or the Interest satisfaction rate, which cannot differentiate the IFA from attacks, is the criterion Traditional IFA detection techniques are concerned with. Threshold values can casually affect Threshold-based traditional methods. This article proposes an accurate IFA detection mechanism based on a Multiple Feature-based Extreme Learning Machine (MF-ELM). Accuracy of the attack detection can be increased by presenting the entropy of Internet names, Interest satisfaction rate and PIT usage as features extracted in the MF-ELM classifier. Furthermore, we deploy a queue-based hostile Interest prefix mitigation mechanism. The inference of this real-time test bed is that the mechanism can help the network to resist IFA with higher accuracy and efficiency.

Keywords: information-centric network, pending interest table, interest flooding attack, MF-ELM classifier, queue-based mitigation strategy

Procedia PDF Downloads 185
210 Synergy and Complementarity in Technology-Intensive Manufacturing Networks

Authors: Daidai Shen, Jean Claude Thill, Wenjia Zhang

Abstract:

This study explores the dynamics of synergy and complementarity within city networks, specifically focusing on the headquarters-subsidiary relations of firms. We begin by defining these two types of networks and establishing their pivotal roles in shaping city network structures. Utilizing the mesoscale analytic approach of weighted stochastic block modeling, we discern relational patterns between city pairs and determine connection strengths through statistical inference. Furthermore, we introduce a community detection approach to uncover the underlying structure of these networks using advanced statistical methods. Our analysis, based on comprehensive network data up to 2017, reveals the coexistence of both complementarity and synergy networks within China’s technology-intensive manufacturing cities. Notably, firms in technology hardware and office & computing machinery predominantly contribute to the complementarity city networks. In contrast, a distinct synergy city network, underpinned by the cities of Suzhou and Dongguan, emerges amidst the expansive complementarity structures in technology hardware and equipment. These findings provide new insights into the relational dynamics and structural configurations of city networks in the context of technology-intensive manufacturing, highlighting the nuanced interplay between synergy and complementarity.

Keywords: city system, complementarity, synergy network, higher-order network

Procedia PDF Downloads 0