Search results for: perceptual linear prediction (PLP’s)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5477

Search results for: perceptual linear prediction (PLP’s)

4787 3D Linear and Cyclic Homo-Peptide Crystals Forged by Supramolecular Swelling Self-Assembly

Authors: Wenliang Song, Yu Zhang, Hua Jin, Il Kim

Abstract:

The self-assembly of the polypeptide (PP) into well-defined structures at different length scales is both biomimetic relevant and fundamentally interesting. Although there are various reports of nanostructures fabricated by the self-assembly of various PPs, directed self-assembly of PP into three-dimensional (3D) hierarchical structure has proven to be difficult, despite their importance for biological applications. Herein, an efficient method has been developed through living polymerization of phenylalanine N-Carboxy anhydride (NCA) towards the linear and cyclic polyphenylalanine, and the new invented swelling methodology can form diverse hierarchical polypeptide crystals. The solvent-dependent self-assembly behaviors of these homopolymers were characterized by high-resolution imaging tools such as atomic force microscopy, transmission electron microscopy, scanning electron microscope. The linear and cyclic polypeptide formed 3D nano hierarchical shapes, such as a sphere, cubic, stratiform and hexagonal star in different solvents. Notably, a crystalline packing model was proposed to explain the formation of 3D nanostructures based on the various diffraction patterns, looking forward to give an insight for their dissimilar shape inflection during the self-assembly process.

Keywords: self-assembly, polypeptide, bio-polymer, crystalline polymer

Procedia PDF Downloads 240
4786 Fast and Efficient Algorithms for Evaluating Uniform and Nonuniform Lagrange and Newton Curves

Authors: Taweechai Nuntawisuttiwong, Natasha Dejdumrong

Abstract:

Newton-Lagrange Interpolations are widely used in numerical analysis. However, it requires a quadratic computational time for their constructions. In computer aided geometric design (CAGD), there are some polynomial curves: Wang-Ball, DP and Dejdumrong curves, which have linear time complexity algorithms. Thus, the computational time for Newton-Lagrange Interpolations can be reduced by applying the algorithms of Wang-Ball, DP and Dejdumrong curves. In order to use Wang-Ball, DP and Dejdumrong algorithms, first, it is necessary to convert Newton-Lagrange polynomials into Wang-Ball, DP or Dejdumrong polynomials. In this work, the algorithms for converting from both uniform and non-uniform Newton-Lagrange polynomials into Wang-Ball, DP and Dejdumrong polynomials are investigated. Thus, the computational time for representing Newton-Lagrange polynomials can be reduced into linear complexity. In addition, the other utilizations of using CAGD curves to modify the Newton-Lagrange curves can be taken.

Keywords: Lagrange interpolation, linear complexity, monomial matrix, Newton interpolation

Procedia PDF Downloads 234
4785 A Neural Network Modelling Approach for Predicting Permeability from Well Logs Data

Authors: Chico Horacio Jose Sambo

Abstract:

Recently neural network has gained popularity when come to solve complex nonlinear problems. Permeability is one of fundamental reservoir characteristics system that are anisotropic distributed and non-linear manner. For this reason, permeability prediction from well log data is well suited by using neural networks and other computer-based techniques. The main goal of this paper is to predict reservoir permeability from well logs data by using neural network approach. A multi-layered perceptron trained by back propagation algorithm was used to build the predictive model. The performance of the model on net results was measured by correlation coefficient. The correlation coefficient from testing, training, validation and all data sets was evaluated. The results show that neural network was capable of reproducing permeability with accuracy in all cases, so that the calculated correlation coefficients for training, testing and validation permeability were 0.96273, 0.89991 and 0.87858, respectively. The generalization of the results to other field can be made after examining new data, and a regional study might be possible to study reservoir properties with cheap and very fast constructed models.

Keywords: neural network, permeability, multilayer perceptron, well log

Procedia PDF Downloads 403
4784 Neural Synchronization - The Brain’s Transfer of Sensory Data

Authors: David Edgar

Abstract:

To understand how the brain’s subconscious and conscious functions, we must conquer the physics of Unity, which leads to duality’s algorithm. Where the subconscious (bottom-up) and conscious (top-down) processes function together to produce and consume intelligence, we use terms like ‘time is relative,’ but we really do understand the meaning. In the brain, there are different processes and, therefore, different observers. These different processes experience time at different rates. A sensory system such as the eyes cycles measurement around 33 milliseconds, the conscious process of the frontal lobe cycles at 300 milliseconds, and the subconscious process of the thalamus cycle at 5 milliseconds. Three different observers experience time differently. To bridge observers, the thalamus, which is the fastest of the processes, maintains a synchronous state and entangles the different components of the brain’s physical process. The entanglements form a synchronous cohesion between the brain components allowing them to share the same state and execute in the same measurement cycle. The thalamus uses the shared state to control the firing sequence of the brain’s linear subconscious process. Sharing state also allows the brain to cheat on the amount of sensory data that must be exchanged between components. Only unpredictable motion is transferred through the synchronous state because predictable motion already exists in the shared framework. The brain’s synchronous subconscious process is entirely based on energy conservation, where prediction regulates energy usage. So, the eyes every 33 milliseconds dump their sensory data into the thalamus every day. The thalamus is going to perform a motion measurement to identify the unpredictable motion in the sensory data. Here is the trick. The thalamus conducts its measurement based on the original observation time of the sensory system (33 ms), not its own process time (5 ms). This creates a data payload of synchronous motion that preserves the original sensory observation. Basically, a frozen moment in time (Flat 4D). The single moment in time can then be processed through the single state maintained by the synchronous process. Other processes, such as consciousness (300 ms), can interface with the synchronous state to generate awareness of that moment. Now, synchronous data traveling through a separate faster synchronous process creates a theoretical time tunnel where observation time is tunneled through the synchronous process and is reproduced on the other side in the original time-relativity. The synchronous process eliminates time dilation by simply removing itself from the equation so that its own process time does not alter the experience. To the original observer, the measurement appears to be instantaneous, but in the thalamus, a linear subconscious process generating sensory perception and thought production is being executed. It is all just occurring in the time available because other observation times are slower than thalamic measurement time. For life to exist in the physical universe requires a linear measurement process, it just hides by operating at a faster time relativity. What’s interesting is time dilation is not the problem; it’s the solution. Einstein said there was no universal time.

Keywords: neural synchronization, natural intelligence, 99.95% IoT data transmission savings, artificial subconscious intelligence (ASI)

Procedia PDF Downloads 126
4783 Linear Semi Active Controller of Magneto-Rheological Damper for Seismic Vibration Attenuation

Authors: Zizouni Khaled, Fali Leyla, Sadek Younes, Bousserhane Ismail Khalil

Abstract:

In structural vibration caused principally by an earthquake excitation, the most vibration’s attenuation system used recently is the semi active control with a Magneto Rheological Damper device. This control was a subject of many researches and works in the last years. The big challenges of searchers in this case is to propose an adequate controller with a robust algorithm of current or tension adjustment. In this present paper, a linear controller is proposed to control the MR damper using to reduce a vibrations of three story structure exposed to El Centro’s 1940 and Boumerdès 2003 earthquakes. In this example, the MR damper is installed in the first floor of the structure. The numerical simulations results of the proposed linear control with a feedback law based on clipped optimal algorithm showed the feasibility of the semi active control to protecting civil structures. The comparison of the controlled structure and uncontrolled structures responses illustrate clearly the performance and the effectiveness of the simple proposed approach.

Keywords: MR damper, seismic vibration, semi-active control

Procedia PDF Downloads 284
4782 Reliability Assessment of Various Empirical Formulas for Prediction of Scour Hole Depth (Plunge Pool) Using a Comprehensive Physical Model

Authors: Majid Galoie, Khodadad Safavi, Abdolreza Karami Nejad, Reza Roshan

Abstract:

In this study, a comprehensive scouring model has been developed in order to evaluate the accuracy of various empirical relationships which were suggested for prediction of scour hole depth in plunge pools by Martins, Mason, Chian and Veronese. For this reason, scour hole depths caused by free falling jets from a flip bucket to a plunge pool were investigated. In this study various discharges, angles, scouring times, etc. have been considered. The final results demonstrated that the all mentioned empirical formulas, except Mason formula, were reasonably agreement with the experimental data.

Keywords: scour hole depth, plunge pool, physical model, reliability assessment

Procedia PDF Downloads 535
4781 Neural Network Based Path Loss Prediction for Global System for Mobile Communication in an Urban Environment

Authors: Danladi Ali

Abstract:

In this paper, we measured GSM signal strength in the Dnepropetrovsk city in order to predict path loss in study area using nonlinear autoregressive neural network prediction and we also, used neural network clustering to determine average GSM signal strength receive at the study area. The nonlinear auto-regressive neural network predicted that the GSM signal is attenuated with the mean square error (MSE) of 2.6748dB, this attenuation value is used to modify the COST 231 Hata and the Okumura-Hata models. The neural network clustering revealed that -75dB to -95dB is received more frequently. This means that the signal strength received at the study is mostly weak signal

Keywords: one-dimensional multilevel wavelets, path loss, GSM signal strength, propagation, urban environment and model

Procedia PDF Downloads 382
4780 Hierarchical Piecewise Linear Representation of Time Series Data

Authors: Vineetha Bettaiah, Heggere S. Ranganath

Abstract:

This paper presents a Hierarchical Piecewise Linear Approximation (HPLA) for the representation of time series data in which the time series is treated as a curve in the time-amplitude image space. The curve is partitioned into segments by choosing perceptually important points as break points. Each segment between adjacent break points is recursively partitioned into two segments at the best point or midpoint until the error between the approximating line and the original curve becomes less than a pre-specified threshold. The HPLA representation achieves dimensionality reduction while preserving prominent local features and general shape of time series. The representation permits course-fine processing at different levels of details, allows flexible definition of similarity based on mathematical measures or general time series shape, and supports time series data mining operations including query by content, clustering and classification based on whole or subsequence similarity.

Keywords: data mining, dimensionality reduction, piecewise linear representation, time series representation

Procedia PDF Downloads 275
4779 Movie Genre Preference Prediction Using Machine Learning for Customer-Based Information

Authors: Haifeng Wang, Haili Zhang

Abstract:

Most movie recommendation systems have been developed for customers to find items of interest. This work introduces a predictive model usable by small and medium-sized enterprises (SMEs) who are in need of a data-based and analytical approach to stock proper movies for local audiences and retain more customers. We used classification models to extract features from thousands of customers’ demographic, behavioral and social information to predict their movie genre preference. In the implementation, a Gaussian kernel support vector machine (SVM) classification model and a logistic regression model were established to extract features from sample data and their test error-in-sample were compared. Comparison of error-out-sample was also made under different Vapnik–Chervonenkis (VC) dimensions in the machine learning algorithm to find and prevent overfitting. Gaussian kernel SVM prediction model can correctly predict movie genre preferences in 85% of positive cases. The accuracy of the algorithm increased to 93% with a smaller VC dimension and less overfitting. These findings advance our understanding of how to use machine learning approach to predict customers’ preferences with a small data set and design prediction tools for these enterprises.

Keywords: computational social science, movie preference, machine learning, SVM

Procedia PDF Downloads 260
4778 Hybrid Renewable Energy System Development Towards Autonomous Operation: The Deployment Potential in Greece

Authors: Afroditi Zamanidou, Dionysios Giannakopoulos, Konstantinos Manolitsis

Abstract:

A notable amount of electrical energy demand in many countries worldwide is used to cover public energy demand for road, square and other public spaces’ lighting. Renewable energy can contribute in a significant way to the electrical energy demand coverage for public lighting. This paper focuses on the sizing and design of a hybrid energy system (HES) exploiting the solar-wind energy potential to meet the electrical energy needs of lighting roads, squares and other public spaces. Moreover, the proposed HES provides coverage of the electrical energy demand for a Wi-Fi hotspot and a charging hotspot for the end-users. Alongside the sizing of the energy production system of the proposed HES, in order to ensure a reliable supply without interruptions, a storage system is added and sized. Multiple scenarios of energy consumption are assumed and applied in order to optimize the sizing of the energy production system and the energy storage system. A database with meteorological prediction data for 51 areas in Greece is developed in order to assess the possible deployment of the proposed HES. Since there are detailed meteorological prediction data for all 51 areas under investigation, the use of these data is evaluated, comparing them to real meteorological data. The meteorological prediction data are exploited to form three hourly production profiles for each area for every month of the year; minimum, average and maximum energy production. The energy production profiles are combined with the energy consumption scenarios and the sizing results of the energy production system and the energy storage system are extracted and presented for every area. Finally, the economic performance of the proposed HES in terms of Levelized cost of energy is estimated by calculating and assessing construction, operation and maintenance costs.

Keywords: energy production system sizing, Greece’s deployment potential, meteorological prediction data, wind-solar hybrid energy system, levelized cost of energy

Procedia PDF Downloads 154
4777 Synthesis and Prediction of Activity Spectra of Substances-Assisted Evaluation of Heterocyclic Compounds Containing Hydroquinoline Scaffolds

Authors: Gizachew Mulugeta Manahelohe, Khidmet Safarovich Shikhaliev

Abstract:

There has been a significant surge in interest in the synthesis of heterocyclic compounds that contain hydroquinoline fragments. This surge can be attributed to the broad range of pharmaceutical and industrial applications that these compounds possess. The present study provides a comprehensive account of the synthesis of both linear and fused heterocyclic systems that incorporate hydroquinoline fragments. Furthermore, the pharmacological activity spectra of the synthesized compounds were assessed using the in silico method, employing the prediction of activity spectra of substances (PASS) program. Hydroquinoline nitriles 7 and 8 were prepared through the reaction of the corresponding hydroquinolinecarbaldehyde using a hydroxylammonium chloride/pyridine/toluene system and iodine in aqueous ammonia under ambient conditions, respectively. 2-Phenyl-1,3-oxazol-5(4H)-ones 9a,b and 10a,b were synthesized via the condensation of compounds 5a,b and 6a,b with hippuric acid in acetic acid in 30–60% yield. When activated, 7-methylazolopyrimidines 11a and b were reacted with N-alkyl-2,2,4-trimethyl-1,2,3,4-tetrahydroquinoline-6-carbaldehydes 6a and b, and triazolo/pyrazolo[1,5-a]pyrimidin-6-yl carboxylic acids 12a and b were obtained in 60–70% yield. The condensation of 7-hydroxy-1,2,3,4-tetramethyl-1,2-dihydroquinoline 3 h with dimethylacetylenedicarboxylate (DMAD) and ethyl acetoacetate afforded cyclic products 16 and 17, respectively. The condensation reaction of 6-formyl-7-hydroxy-1,2,2,4-tetramethyl-1,2-dihydroquinoline 5e with methylene-active compounds such as ethyl cyanoacetate/dimethyl-3-oxopentanedioate/ethyl acetoacetate/diethylmalonate/Meldrum’s acid afforded 3-substituted coumarins containing dihydroquinolines 19 and 21. Pentacyclic coumarin 22 was obtained via the random condensation of malononitrile with 5e in the presence of a catalytic amount of piperidine in ethanol. The biological activities of the synthesized compounds were assessed using the PASS program. Based on the prognosis, compounds 13a, b, and 14 exhibited a high likelihood of being active as inhibitors of gluconate 2-dehydrogenase, as well as possessing antiallergic, antiasthmatic, and antiarthritic properties, with a probability value (Pa) ranging from 0.849 to 0.870. Furthermore, it was discovered that hydroquinoline carbonitriles 7 and 8 tended to act as effective progesterone antagonists and displayed antiallergic, antiasthmatic, and antiarthritic effects (Pa = 0.276–0.827). Among the hydroquinolines containing coumarin moieties, compounds 17, 19a, and 19c were predicted to be potent progesterone antagonists, with Pa values of 0.710, 0.630, and 0.615, respectively.

Keywords: heterocyclic compound, hydroquinoline, Vilsmeier–Haack formulation, quinolone

Procedia PDF Downloads 42
4776 Supervised-Component-Based Generalised Linear Regression with Multiple Explanatory Blocks: THEME-SCGLR

Authors: Bry X., Trottier C., Mortier F., Cornu G., Verron T.

Abstract:

We address component-based regularization of a Multivariate Generalized Linear Model (MGLM). A set of random responses Y is assumed to depend, through a GLM, on a set X of explanatory variables, as well as on a set T of additional covariates. X is partitioned into R conceptually homogeneous blocks X1, ... , XR , viewed as explanatory themes. Variables in each Xr are assumed many and redundant. Thus, Generalised Linear Regression (GLR) demands regularization with respect to each Xr. By contrast, variables in T are assumed selected so as to demand no regularization. Regularization is performed searching each Xr for an appropriate number of orthogonal components that both contribute to model Y and capture relevant structural information in Xr. We propose a very general criterion to measure structural relevance (SR) of a component in a block, and show how to take SR into account within a Fisher-scoring-type algorithm in order to estimate the model. We show how to deal with mixed-type explanatory variables. The method, named THEME-SCGLR, is tested on simulated data.

Keywords: Component-Model, Fisher Scoring Algorithm, GLM, PLS Regression, SCGLR, SEER, THEME

Procedia PDF Downloads 396
4775 Prediction Factor of Recurrence Supraventricular Tachycardia After Adenosine Treatment in the Emergency Department

Authors: Chaiyaporn Yuksen

Abstract:

Backgroud: Supraventricular tachycardia (SVT) is an abnormally fast atrial tachycardia characterized by narrow (≤ 120 ms) and constant QRS. Adenosine was the drug of choice; the first dose was 6 mg. It can be repeated with the second and third doses of 12 mg, with greater than 90% success. The study found that patients observed at 4 hours after normal sinus rhythm was no recurrence within 24 hours. The objective of this study was to investigate the factors that influence the recurrence of SVT after adenosine in the emergency department (ED). Method: The study was conducted retrospectively exploratory model, prognostic study at the Emergency Department (ED) in Faculty of Medicine, Ramathibodi Hospital, a university-affiliated super tertiary care hospital in Bangkok, Thailand. The study was conducted for ten years period between 2010 and 2020. The inclusion criteria were age > 15 years, visiting the ED with SVT, and treating with adenosine. Those patients were recorded with the recurrence SVT in ED. The multivariable logistic regression model developed the predictive model and prediction score for recurrence PSVT. Result: 264 patients met the study criteria. Of those, 24 patients (10%) had recurrence PSVT. Five independent factors were predictive of recurrence PSVT. There was age>65 years, heart rate (after adenosine) > 100 per min, structural heart disease, and dose of adenosine. The clinical risk score to predict recurrence PSVT is developed accuracy 74.41%. The score of >6 had the likelihood ratio of recurrence PSVT by 5.71 times Conclusion: The clinical predictive score of > 6 was associated with recurrence PSVT in ED.

Keywords: clinical prediction score, SVT, recurrence, emergency department

Procedia PDF Downloads 155
4774 Implementation of Deep Neural Networks for Pavement Condition Index Prediction

Authors: M. Sirhan, S. Bekhor, A. Sidess

Abstract:

In-service pavements deteriorate with time due to traffic wheel loads, environment, and climate conditions. Pavement deterioration leads to a reduction in their serviceability and structural behavior. Consequently, proper maintenance and rehabilitation (M&R) are necessary actions to keep the in-service pavement network at the desired level of serviceability. Due to resource and financial constraints, the pavement management system (PMS) prioritizes roads most in need of maintenance and rehabilitation action. It recommends a suitable action for each pavement based on the performance and surface condition of each road in the network. The pavement performance and condition are usually quantified and evaluated by different types of roughness-based and stress-based indices. Examples of such indices are Pavement Serviceability Index (PSI), Pavement Serviceability Ratio (PSR), Mean Panel Rating (MPR), Pavement Condition Rating (PCR), Ride Number (RN), Profile Index (PI), International Roughness Index (IRI), and Pavement Condition Index (PCI). PCI is commonly used in PMS as an indicator of the extent of the distresses on the pavement surface. PCI values range between 0 and 100; where 0 and 100 represent a highly deteriorated pavement and a newly constructed pavement, respectively. The PCI value is a function of distress type, severity, and density (measured as a percentage of the total pavement area). PCI is usually calculated iteratively using the 'Paver' program developed by the US Army Corps. The use of soft computing techniques, especially Artificial Neural Network (ANN), has become increasingly popular in the modeling of engineering problems. ANN techniques have successfully modeled the performance of the in-service pavements, due to its efficiency in predicting and solving non-linear relationships and dealing with an uncertain large amount of data. Typical regression models, which require a pre-defined relationship, can be replaced by ANN, which was found to be an appropriate tool for predicting the different pavement performance indices versus different factors as well. Subsequently, the objective of the presented study is to develop and train an ANN model that predicts the PCI values. The model’s input consists of percentage areas of 11 different damage types; alligator cracking, swelling, rutting, block cracking, longitudinal/transverse cracking, edge cracking, shoving, raveling, potholes, patching, and lane drop off, at three severity levels (low, medium, high) for each. The developed model was trained using 536,000 samples and tested on 134,000 samples. The samples were collected and prepared by The National Transport Infrastructure Company. The predicted results yielded satisfactory compliance with field measurements. The proposed model predicted PCI values with relatively low standard deviations, suggesting that it could be incorporated into the PMS for PCI determination. It is worth mentioning that the most influencing variables for PCI prediction are damages related to alligator cracking, swelling, rutting, and potholes.

Keywords: artificial neural networks, computer programming, pavement condition index, pavement management, performance prediction

Procedia PDF Downloads 137
4773 Numerical Solution of Space Fractional Order Linear/Nonlinear Reaction-Advection Diffusion Equation Using Jacobi Polynomial

Authors: Shubham Jaiswal

Abstract:

During modelling of many physical problems and engineering processes, fractional calculus plays an important role. Those are greatly described by fractional differential equations (FDEs). So a reliable and efficient technique to solve such types of FDEs is needed. In this article, a numerical solution of a class of fractional differential equations namely space fractional order reaction-advection dispersion equations subject to initial and boundary conditions is derived. In the proposed approach shifted Jacobi polynomials are used to approximate the solutions together with shifted Jacobi operational matrix of fractional order and spectral collocation method. The main advantage of this approach is that it converts such problems in the systems of algebraic equations which are easier to be solved. The proposed approach is effective to solve the linear as well as non-linear FDEs. To show the reliability, validity and high accuracy of proposed approach, the numerical results of some illustrative examples are reported, which are compared with the existing analytical results already reported in the literature. The error analysis for each case exhibited through graphs and tables confirms the exponential convergence rate of the proposed method.

Keywords: space fractional order linear/nonlinear reaction-advection diffusion equation, shifted Jacobi polynomials, operational matrix, collocation method, Caputo derivative

Procedia PDF Downloads 445
4772 An Experimental Study on Service Life Prediction of Self: Compacting Concrete Using Sorptivity as a Durability Index

Authors: S. Girish, N. Ajay

Abstract:

Permeation properties have been widely used to quantify durability characteristics of concrete for assessing long term performance and sustainability. The processes of deterioration in concrete are mediated largely by water. There is a strong interest in finding a better way of assessing the material properties of concrete in terms of durability. Water sorptivity is a useful single material property which can be one of the measures of durability useful in service life planning and prediction, especially in severe environmental conditions. This paper presents the results of the comparative study of sorptivity of Self-Compacting Concrete (SCC) with conventionally vibrated concrete. SCC is a new, special type of concrete mixture, characterized by high resistance to segregation that can flow through intricate geometrical configuration in the presence of reinforcement, under its own mass, without vibration and compaction. SCC mixes were developed for the paste contents of 0.38, 0.41 and 0.43 with fly ash as the filler for different cement contents ranging from 300 to 450 kg/m3. The study shows better performance by SCC in terms of capillary absorption. The sorptivity value decreased as the volume of paste increased. The use of higher paste content in SCC can make the concrete robust with better densification of the micro-structure, improving the durability and making the concrete more sustainable with improved long term performance. The sorptivity based on secondary absorption can be effectively used as a durability index to predict the time duration required for the ingress of water to penetrate the concrete, which has practical significance.

Keywords: self-compacting concrete, service life prediction, sorptivity, volume of paste

Procedia PDF Downloads 321
4771 Non-Linear Control in Positioning of PMLSM by Estimates of the Load Force by MRAS Method

Authors: Maamar Yahiaoui, Abdelrrahmene Kechich, Ismail Elkhallile Bousserhene

Abstract:

This article presents a study in simulation by means of MATLAB/Simulink software of the nonlinear control in positioning of a linear synchronous machine with the esteemed force of load, to have effective control in the estimator in all tests the wished trajectory follows and the disturbance of load start. The results of simulation prove clearly that the control proposed can detect the reference of positioning the value estimates of load force equal to the actual value.

Keywords: mathematical model, Matlab, PMLSM, control, linearization, estimator, force, load, current

Procedia PDF Downloads 608
4770 'CardioCare': A Cutting-Edge Fusion of IoT and Machine Learning to Bridge the Gap in Cardiovascular Risk Management

Authors: Arpit Patil, Atharav Bhagwat, Rajas Bhope, Pramod Bide

Abstract:

This research integrates IoT and ML to predict heart failure risks, utilizing the Framingham dataset. IoT devices gather real-time physiological data, focusing on heart rate dynamics, while ML, specifically Random Forest, predicts heart failure. Rigorous feature selection enhances accuracy, achieving over 90% prediction rate. This amalgamation marks a transformative step in proactive healthcare, highlighting early detection's critical role in cardiovascular risk mitigation. Challenges persist, necessitating continual refinement for improved predictive capabilities.

Keywords: cardiovascular diseases, internet of things, machine learning, cardiac risk assessment, heart failure prediction, early detection, cardio data analysis

Procedia PDF Downloads 11
4769 Effect of Linear Thermal Gradient on Steady-State Creep Behavior of Isotropic Rotating Disc

Authors: Minto Rattan, Tania Bose, Neeraj Chamoli

Abstract:

The present paper investigates the effect of linear thermal gradient on the steady-state creep behavior of rotating isotropic disc using threshold stress based Sherby’s creep law. The composite discs made of aluminum matrix reinforced with silicon carbide particulate has been taken for analysis. The stress and strain rate distributions have been calculated for discs rotating at linear thermal gradation using von Mises’ yield criterion. The material parameters have been estimated by regression fit of the available experimental data. The results are displayed and compared graphically in designer friendly format for the above said temperature profile with the disc operating under uniform temperature profile. It is observed that radial and tangential stresses show minor variation and the strain rates vary significantly in the presence of thermal gradation as compared to disc having uniform temperature.

Keywords: creep, isotropic, steady-state, thermal gradient

Procedia PDF Downloads 269
4768 Why Do We Need Hierachical Linear Models?

Authors: Mustafa Aydın, Ali Murat Sunbul

Abstract:

Hierarchical or nested data structures usually are seen in many research areas. Especially, in the field of education, if we examine most of the studies, we can see the nested structures. Students in classes, classes in schools, schools in cities and cities in regions are similar nested structures. In a hierarchical structure, students being in the same class, sharing the same physical conditions and similar experiences and learning from the same teachers, they demonstrate similar behaviors between them rather than the students in other classes.

Keywords: hierarchical linear modeling, nested data, hierarchical structure, data structure

Procedia PDF Downloads 652
4767 Prevalence of Cerebral Microbleeds in Apparently Healthy, Elderly Population: A Meta-Analysis

Authors: Vidishaa Jali, Amit Sinha, Kameshwar Prasad

Abstract:

Background and Objective: Cerebral microbleeds are frequently found in healthy elderly individuals. We performed a meta- analysis to determine the prevalence of cerebral microbleeds in apparently healthy, elderly population and to determine the effect of age, smoking and hypertension on the occurrence of cerebral microbleeds. Methods: Relevant literature was searched using electronic databases such as MEDLINE, EMBASE, PubMed, Cochrane database, Google scholar to identify studies on the prevalence of cerebral microbleeds in general elderly population till March 2016. STATA version 13 software was used for analysis. Fixed effect model was used if heterogeneity was less than 50%. Otherwise, random effect model was used. Meta- regression analysis was performed to check any effect of important variables such as age, smoking, hypertension. Selection Criteria: We included cross-sectional studies performed in apparently healthy elderly population, who had age more than 50 years. Results: The pooled proportion of cerebral microbleeds in healthy population is 12% (95% CI, 0.11 to 0.13). No significant effect of age was found on the prevalence of cerebral microbleeds (p= 0.99). A linear relationship between increase in hypertension and the prevalence of cerebral microbleeds was found, however, this linear relationship was not statistically significant (p=0.16). Similarly, A linear relationship between increase in smoking and the prevalence of cerebral microbleeds was found, however, this linear relationship was also not statistically significant (p=0.21). Conclusion: Presence of cerebral microbleeds is evident in apparently healthy, elderly population, in more than 10% of individuals.

Keywords: apparently healthy, elderly, prevalence, cerebral microbleeds

Procedia PDF Downloads 296
4766 Learning to Recommend with Negative Ratings Based on Factorization Machine

Authors: Caihong Sun, Xizi Zhang

Abstract:

Rating prediction is an important problem for recommender systems. The task is to predict the rating for an item that a user would give. Most of the existing algorithms for the task ignore the effect of negative ratings rated by users on items, but the negative ratings have a significant impact on users’ purchasing decisions in practice. In this paper, we present a rating prediction algorithm based on factorization machines that consider the effect of negative ratings inspired by Loss Aversion theory. The aim of this paper is to develop a concave and a convex negative disgust function to evaluate the negative ratings respectively. Experiments are conducted on MovieLens dataset. The experimental results demonstrate the effectiveness of the proposed methods by comparing with other four the state-of-the-art approaches. The negative ratings showed much importance in the accuracy of ratings predictions.

Keywords: factorization machines, feature engineering, negative ratings, recommendation systems

Procedia PDF Downloads 242
4765 Reducing Uncertainty of Monte Carlo Estimated Fatigue Damage in Offshore Wind Turbines Using FORM

Authors: Jan-Tore H. Horn, Jørgen Juncher Jensen

Abstract:

Uncertainties related to fatigue damage estimation of non-linear systems are highly dependent on the tail behaviour and extreme values of the stress range distribution. By using a combination of the First Order Reliability Method (FORM) and Monte Carlo simulations (MCS), the accuracy of the fatigue estimations may be improved for the same computational efforts. The method is applied to a bottom-fixed, monopile-supported large offshore wind turbine, which is a non-linear and dynamically sensitive system. Different curve fitting techniques to the fatigue damage distribution have been used depending on the sea-state dependent response characteristics, and the effect of a bi-linear S-N curve is discussed. Finally, analyses are performed on several environmental conditions to investigate the long-term applicability of this multistep method. Wave loads are calculated using state-of-the-art theory, while wind loads are applied with a simplified model based on rotor thrust coefficients.

Keywords: fatigue damage, FORM, monopile, Monte Carlo, simulation, wind turbine

Procedia PDF Downloads 260
4764 Improved Computational Efficiency of Machine Learning Algorithm Based on Evaluation Metrics to Control the Spread of Coronavirus in the UK

Authors: Swathi Ganesan, Nalinda Somasiri, Rebecca Jeyavadhanam, Gayathri Karthick

Abstract:

The COVID-19 crisis presents a substantial and critical hazard to worldwide health. Since the occurrence of the disease in late January 2020 in the UK, the number of infected people confirmed to acquire the illness has increased tremendously across the country, and the number of individuals affected is undoubtedly considerably high. The purpose of this research is to figure out a predictive machine learning archetypal that could forecast COVID-19 cases within the UK. This study concentrates on the statistical data collected from 31st January 2020 to 31st March 2021 in the United Kingdom. Information on total COVID cases registered, new cases encountered on a daily basis, total death registered, and patients’ death per day due to Coronavirus is collected from World Health Organisation (WHO). Data preprocessing is carried out to identify any missing values, outliers, or anomalies in the dataset. The data is split into 8:2 ratio for training and testing purposes to forecast future new COVID cases. Support Vector Machines (SVM), Random Forests, and linear regression algorithms are chosen to study the model performance in the prediction of new COVID-19 cases. From the evaluation metrics such as r-squared value and mean squared error, the statistical performance of the model in predicting the new COVID cases is evaluated. Random Forest outperformed the other two Machine Learning algorithms with a training accuracy of 99.47% and testing accuracy of 98.26% when n=30. The mean square error obtained for Random Forest is 4.05e11, which is lesser compared to the other predictive models used for this study. From the experimental analysis Random Forest algorithm can perform more effectively and efficiently in predicting the new COVID cases, which could help the health sector to take relevant control measures for the spread of the virus.

Keywords: COVID-19, machine learning, supervised learning, unsupervised learning, linear regression, support vector machine, random forest

Procedia PDF Downloads 121
4763 Physical Habitat Simulation and Comparison within a Lerma River Reach, with Respect to the Same but Modified Reach, to Create a Linear Park

Authors: Garcia-Rodriguez Ezequiel, Luis A. Ochoa-Franco, Adrian I. Cervantes-Servin

Abstract:

In this work, the Ictalurus punctatus species estimated available physical habitat is compared with the estimated physical habitat for the same but modified river reach, with the aim of creating a linear park, along a length of 5 500 m. To determine the effect of ecological park construction, on physical habitat of the Lerma river stretch of study, first, the available habitat for the Ictalurus punctatus species was estimated through the simulation of the physical habitat, by using surveying, hydraulics, and habitat information gotten at the river reach in its actual situation. Second, it was estimated the available habitat for the above species, upon the simulation of the physical habitat through the proposed modification for the ecological park creation. Third, it is presented a comparison between both scenarios in terms of available habitat estimated for Ictalurus punctatus species, concluding that in cases of adult and spawning life stages, changes in the channel to create an ecological park would produce a considerable loss of potentially usable habitat (PUH), while in the case of the juvenile life stage PUH remains virtually unchanged, and in the case of life stage fry the PUH would increase due to the presence of velocities and depths of lesser magnitude, due to the presence of minor flow rates and lower volume of the wet channel. It is expected that habitat modification for linear park construction may produce the lack of Ictalurus punktatus species conservation at the river reach of the study.

Keywords: Habitat modification, Ictalurus punctatus, Lerma, river, linear park

Procedia PDF Downloads 475
4762 Evaluation of a Piecewise Linear Mixed-Effects Model in the Analysis of Randomized Cross-over Trial

Authors: Moses Mwangi, Geert Verbeke, Geert Molenberghs

Abstract:

Cross-over designs are commonly used in randomized clinical trials to estimate efficacy of a new treatment with respect to a reference treatment (placebo or standard). The main advantage of using cross-over design over conventional parallel design is its flexibility, where every subject become its own control, thereby reducing confounding effect. Jones & Kenward, discuss in detail more recent developments in the analysis of cross-over trials. We revisit the simple piecewise linear mixed-effects model, proposed by Mwangi et. al, (in press) for its first application in the analysis of cross-over trials. We compared performance of the proposed piecewise linear mixed-effects model with two commonly cited statistical models namely, (1) Grizzle model; and (2) Jones & Kenward model, used in estimation of the treatment effect, in the analysis of randomized cross-over trial. We estimate two performance measurements (mean square error (MSE) and coverage probability) for the three methods, using data simulated from the proposed piecewise linear mixed-effects model. Piecewise linear mixed-effects model yielded lowest MSE estimates compared to Grizzle and Jones & Kenward models for both small (Nobs=20) and large (Nobs=600) sample sizes. It’s coverage probability were highest compared to Grizzle and Jones & Kenward models for both small and large sample sizes. A piecewise linear mixed-effects model is a better estimator of treatment effect than its two competing estimators (Grizzle and Jones & Kenward models) in the analysis of cross-over trials. The data generating mechanism used in this paper captures two time periods for a simple 2-Treatments x 2-Periods cross-over design. Its application is extendible to more complex cross-over designs with multiple treatments and periods. In addition, it is important to note that, even for single response models, adding more random effects increases the complexity of the model and thus may be difficult or impossible to fit in some cases.

Keywords: Evaluation, Grizzle model, Jones & Kenward model, Performance measures, Simulation

Procedia PDF Downloads 122
4761 Prediction of Cutting Tool Life in Drilling of Reinforced Aluminum Alloy Composite Using a Fuzzy Method

Authors: Mohammed T. Hayajneh

Abstract:

Machining of Metal Matrix Composites (MMCs) is very significant process and has been a main problem that draws many researchers to investigate the characteristics of MMCs during different machining process. The poor machining properties of hard particles reinforced MMCs make drilling process a rather interesting task. Unlike drilling of conventional materials, many problems can be seriously encountered during drilling of MMCs, such as tool wear and cutting forces. Cutting tool wear is a very significant concern in industries. Cutting tool wear not only influences the quality of the drilled hole, but also affects the cutting tool life. Prediction the cutting tool life during drilling is essential for optimizing the cutting conditions. However, the relationship between tool life and cutting conditions, tool geometrical factors and workpiece material properties has not yet been established by any machining theory. In this research work, fuzzy subtractive clustering system has been used to model the cutting tool life in drilling of Al2O3 particle reinforced aluminum alloy composite to investigate of the effect of cutting conditions on cutting tool life. This investigation can help in controlling and optimizing of cutting conditions when the process parameters are adjusted. The built model for prediction the tool life is identified by using drill diameter, cutting speed, and cutting feed rate as input data. The validity of the model was confirmed by the examinations under various cutting conditions. Experimental results have shown the efficiency of the model to predict cutting tool life.

Keywords: composite, fuzzy, tool life, wear

Procedia PDF Downloads 295
4760 Developing a Machine Learning-based Cost Prediction Model for Construction Projects using Particle Swarm Optimization

Authors: Soheila Sadeghi

Abstract:

Accurate cost prediction is essential for effective project management and decision-making in the construction industry. This study aims to develop a cost prediction model for construction projects using Machine Learning techniques and Particle Swarm Optimization (PSO). The research utilizes a comprehensive dataset containing project cost estimates, actual costs, resource details, and project performance metrics from a road reconstruction project. The methodology involves data preprocessing, feature selection, and the development of an Artificial Neural Network (ANN) model optimized using PSO. The study investigates the impact of various input features, including cost estimates, resource allocation, and project progress, on the accuracy of cost predictions. The performance of the optimized ANN model is evaluated using metrics such as Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and R-squared. The results demonstrate the effectiveness of the proposed approach in predicting project costs, outperforming traditional benchmark models. The feature selection process identifies the most influential variables contributing to cost variations, providing valuable insights for project managers. However, this study has several limitations. Firstly, the model's performance may be influenced by the quality and quantity of the dataset used. A larger and more diverse dataset covering different types of construction projects would enhance the model's generalizability. Secondly, the study focuses on a specific optimization technique (PSO) and a single Machine Learning algorithm (ANN). Exploring other optimization methods and comparing the performance of various ML algorithms could provide a more comprehensive understanding of the cost prediction problem. Future research should focus on several key areas. Firstly, expanding the dataset to include a wider range of construction projects, such as residential buildings, commercial complexes, and infrastructure projects, would improve the model's applicability. Secondly, investigating the integration of additional data sources, such as economic indicators, weather data, and supplier information, could enhance the predictive power of the model. Thirdly, exploring the potential of ensemble learning techniques, which combine multiple ML algorithms, may further improve cost prediction accuracy. Additionally, developing user-friendly interfaces and tools to facilitate the adoption of the proposed cost prediction model in real-world construction projects would be a valuable contribution to the industry. The findings of this study have significant implications for construction project management, enabling proactive cost estimation, resource allocation, budget planning, and risk assessment, ultimately leading to improved project performance and cost control. This research contributes to the advancement of cost prediction techniques in the construction industry and highlights the potential of Machine Learning and PSO in addressing this critical challenge. However, further research is needed to address the limitations and explore the identified future research directions to fully realize the potential of ML-based cost prediction models in the construction domain.

Keywords: cost prediction, construction projects, machine learning, artificial neural networks, particle swarm optimization, project management, feature selection, road reconstruction

Procedia PDF Downloads 59
4759 Real Time Detection, Prediction and Reconstitution of Rain Drops

Authors: R. Burahee, B. Chassinat, T. de Laclos, A. Dépée, A. Sastim

Abstract:

The purpose of this paper is to propose a solution to detect, predict and reconstitute rain drops in real time – during the night – using an embedded material with an infrared camera. To prevent the system from needing too high hardware resources, simple models are considered in a powerful image treatment algorithm reducing considerably calculation time in OpenCV software. Using a smart model – drops will be matched thanks to a process running through two consecutive pictures for implementing a sophisticated tracking system. With this system drops computed trajectory gives information for predicting their future location. Thanks to this technique, treatment part can be reduced. The hardware system composed by a Raspberry Pi is optimized to host efficiently this code for real time execution.

Keywords: reconstitution, prediction, detection, rain drop, real time, raspberry, infrared

Procedia PDF Downloads 419
4758 Performance Analysis of Artificial Neural Network with Decision Tree in Prediction of Diabetes Mellitus

Authors: J. K. Alhassan, B. Attah, S. Misra

Abstract:

Human beings have the ability to make logical decisions. Although human decision - making is often optimal, it is insufficient when huge amount of data is to be classified. medical dataset is a vital ingredient used in predicting patients health condition. In other to have the best prediction, there calls for most suitable machine learning algorithms. This work compared the performance of Artificial Neural Network (ANN) and Decision Tree Algorithms (DTA) as regards to some performance metrics using diabetes data. The evaluations was done using weka software and found out that DTA performed better than ANN. Multilayer Perceptron (MLP) and Radial Basis Function (RBF) were the two algorithms used for ANN, while RegTree and LADTree algorithms were the DTA models used. The Root Mean Squared Error (RMSE) of MLP is 0.3913,that of RBF is 0.3625, that of RepTree is 0.3174 and that of LADTree is 0.3206 respectively.

Keywords: artificial neural network, classification, decision tree algorithms, diabetes mellitus

Procedia PDF Downloads 408