Search results for: higher dimensional pmf estimation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14277

Search results for: higher dimensional pmf estimation

13917 Quality Assurance as an Educational Development Tool: Case from the European Higher Education

Authors: Maha Mourad

Abstract:

Higher education in any competitive European economy should serve the new information society by increasing the supply of good quality education services and by creating good international brands in the international higher education market. Hence, continuous risk management techniques through higher educational reforms programs became one of the top priorities within the European Union to control the quality of higher education. Risk is higher education is studies by several researchers who agreed that the risk in higher education has a direct influence on continuity of quality education and research contribution. The focus of this research is to highlights the Internal Quality Assurance (IQA) activities in the Polish higher education system as a risk management tool used to control the quality of education. This paper presents a qualitative empirical analysis in 5 different universities in Poland. In addition, it aims to help in finding global practical and create benchmark for policy makers concerning the risk management techniques based on the Polish experience.

Keywords: education development, quality assurance, sustainability, european higher education

Procedia PDF Downloads 439
13916 The Evaluation of the Performance of Different Filtering Approaches in Tracking Problem and the Effect of Noise Variance

Authors: Mohammad Javad Mollakazemi, Farhad Asadi, Aref Ghafouri

Abstract:

Performance of different filtering approaches depends on modeling of dynamical system and algorithm structure. For modeling and smoothing the data the evaluation of posterior distribution in different filtering approach should be chosen carefully. In this paper different filtering approaches like filter KALMAN, EKF, UKF, EKS and smoother RTS is simulated in some trajectory tracking of path and accuracy and limitation of these approaches are explained. Then probability of model with different filters is compered and finally the effect of the noise variance to estimation is described with simulations results.

Keywords: Gaussian approximation, Kalman smoother, parameter estimation, noise variance

Procedia PDF Downloads 404
13915 Genetic Algorithm and Multi Criteria Decision Making Approach for Compressive Sensing Based Direction of Arrival Estimation

Authors: Ekin Nurbaş

Abstract:

One of the essential challenges in array signal processing, which has drawn enormous research interest over the past several decades, is estimating the direction of arrival (DOA) of plane waves impinging on an array of sensors. In recent years, the Compressive Sensing based DoA estimation methods have been proposed by researchers, and it has been discovered that the Compressive Sensing (CS)-based algorithms achieved significant performances for DoA estimation even in scenarios where there are multiple coherent sources. On the other hand, the Genetic Algorithm, which is a method that provides a solution strategy inspired by natural selection, has been used in sparse representation problems in recent years and provides significant improvements in performance. With all of those in consideration, in this paper, a method that combines the Genetic Algorithm (GA) and the Multi-Criteria Decision Making (MCDM) approaches for Direction of Arrival (DoA) estimation in the Compressive Sensing (CS) framework is proposed. In this method, we generate a multi-objective optimization problem by splitting the norm minimization and reconstruction loss minimization parts of the Compressive Sensing algorithm. With the help of the Genetic Algorithm, multiple non-dominated solutions are achieved for the defined multi-objective optimization problem. Among the pareto-frontier solutions, the final solution is obtained with the multiple MCDM methods. Moreover, the performance of the proposed method is compared with the CS-based methods in the literature.

Keywords: genetic algorithm, direction of arrival esitmation, multi criteria decision making, compressive sensing

Procedia PDF Downloads 112
13914 Dimensionality and Superconducting Parameters of YBa2Cu3O7 Foams

Authors: Michael Koblischka, Anjela Koblischka-Veneva, XianLin Zeng, Essia Hannachi, Yassine Slimani

Abstract:

Superconducting foams of YBa2Cu3O7 (abbreviated Y-123) were produced using the infiltration growth (IG) technique from Y2BaCuO5 (Y-211) foams. The samples were investigated by SEM (scanning electron microscopy) and electrical resistivity measurements. SEM observations indicated the specific microstructure of the foam struts with numerous tiny Y-211 particles (50-100 nm diameter) embedded in channel-like structures between the Y-123 grains. The investigation of the excess conductivity of different prepared composites was analyzed using Aslamazov-Larkin (AL) model. The investigated samples comprised of five distinct fluctuation regimes, namely short-wave (SWF), one-dimensional (1D), two-dimensional (2D), three-dimensional (3D), and critical (CR) fluctuations regimes. The coherence length along the c-axis at zero-temperature (ξc(0)), lower and upper critical magnetic fields (Bc1 and Bc2), critical current density (Jc) and numerous other superconducting parameters were estimated from the data. The analysis reveals that the presence of the tiny Y-211 particles alters the excess conductivity and the fluctuation behavior observed in standard YBCO samples.

Keywords: Excess conductivity, Foam, Microstructure, Superconductor YBa2Cu3Oy

Procedia PDF Downloads 140
13913 Ensemble Sampler For Infinite-Dimensional Inverse Problems

Authors: Jeremie Coullon, Robert J. Webber

Abstract:

We introduce a Markov chain Monte Carlo (MCMC) sam-pler for infinite-dimensional inverse problems. Our sam-pler is based on the affine invariant ensemble sampler, which uses interacting walkers to adapt to the covariance structure of the target distribution. We extend this ensem-ble sampler for the first time to infinite-dimensional func-tion spaces, yielding a highly efficient gradient-free MCMC algorithm. Because our ensemble sampler does not require gradients or posterior covariance estimates, it is simple to implement and broadly applicable. In many Bayes-ian inverse problems, Markov chain Monte Carlo (MCMC) meth-ods are needed to approximate distributions on infinite-dimensional function spaces, for example, in groundwater flow, medical imaging, and traffic flow. Yet designing efficient MCMC methods for function spaces has proved challenging. Recent gradi-ent-based MCMC methods preconditioned MCMC methods, and SMC methods have improved the computational efficiency of functional random walk. However, these samplers require gradi-ents or posterior covariance estimates that may be challenging to obtain. Calculating gradients is difficult or impossible in many high-dimensional inverse problems involving a numerical integra-tor with a black-box code base. Additionally, accurately estimating posterior covariances can require a lengthy pilot run or adaptation period. These concerns raise the question: is there a functional sampler that outperforms functional random walk without requir-ing gradients or posterior covariance estimates? To address this question, we consider a gradient-free sampler that avoids explicit covariance estimation yet adapts naturally to the covariance struc-ture of the sampled distribution. This sampler works by consider-ing an ensemble of walkers and interpolating and extrapolating between walkers to make a proposal. This is called the affine in-variant ensemble sampler (AIES), which is easy to tune, easy to parallelize, and efficient at sampling spaces of moderate dimen-sionality (less than 20). The main contribution of this work is to propose a functional ensemble sampler (FES) that combines func-tional random walk and AIES. To apply this sampler, we first cal-culate the Karhunen–Loeve (KL) expansion for the Bayesian prior distribution, assumed to be Gaussian and trace-class. Then, we use AIES to sample the posterior distribution on the low-wavenumber KL components and use the functional random walk to sample the posterior distribution on the high-wavenumber KL components. Alternating between AIES and functional random walk updates, we obtain our functional ensemble sampler that is efficient and easy to use without requiring detailed knowledge of the target dis-tribution. In past work, several authors have proposed splitting the Bayesian posterior into low-wavenumber and high-wavenumber components and then applying enhanced sampling to the low-wavenumber components. Yet compared to these other samplers, FES is unique in its simplicity and broad applicability. FES does not require any derivatives, and the need for derivative-free sam-plers has previously been emphasized. FES also eliminates the requirement for posterior covariance estimates. Lastly, FES is more efficient than other gradient-free samplers in our tests. In two nu-merical examples, we apply FES to challenging inverse problems that involve estimating a functional parameter and one or more scalar parameters. We compare the performance of functional random walk, FES, and an alternative derivative-free sampler that explicitly estimates the posterior covariance matrix. We conclude that FES is the fastest available gradient-free sampler for these challenging and multimodal test problems.

Keywords: Bayesian inverse problems, Markov chain Monte Carlo, infinite-dimensional inverse problems, dimensionality reduction

Procedia PDF Downloads 128
13912 Chemical Fingerprinting of the Ephedrine Pathway to Methamphetamine

Authors: Luke Andrighetto, Paul G. Stevenson, Luke C. Henderson, Jim Pearson, Xavier A. Conlan

Abstract:

As pseudoephedrine, a common ingredient in cold and flu medications is closely monitored and restricted in Australia, alternative methods of accessing it are of interest. The impurities and by-products of every reaction step of pseudoephedrine/ephedrine and methamphetamine synthesis have been mapped in order to develop a chemical fingerprint based on synthetic route. Likewise, seized methamphetamine contains a combination of different cutting agents and starting materials. Therefore, in-silico optimised two-dimensional HPLC with DryLab® and OpenMS® software has been used to efficiently separate complex seizure samples. An excellent match between simulated and real separations was observed. Targeted separation of model compounds was completed with significantly reduced method development time. This study produced a two-dimensional separation regime that offers unprecedented separation power (separation space) while maintaining a rapid analysis time that is faster than those previously reported for gas chromatography, single dimension high performance liquid chromatography or capillary electrophoresis.

Keywords: chemical fingerprint, ephedrine, methamphetamine, two-dimensional HPLC

Procedia PDF Downloads 437
13911 Dimensionality Reduction in Modal Analysis for Structural Health Monitoring

Authors: Elia Favarelli, Enrico Testi, Andrea Giorgetti

Abstract:

Autonomous structural health monitoring (SHM) of many structures and bridges became a topic of paramount importance for maintenance purposes and safety reasons. This paper proposes a set of machine learning (ML) tools to perform automatic feature selection and detection of anomalies in a bridge from vibrational data and compare different feature extraction schemes to increase the accuracy and reduce the amount of data collected. As a case study, the Z-24 bridge is considered because of the extensive database of accelerometric data in both standard and damaged conditions. The proposed framework starts from the first four fundamental frequencies extracted through operational modal analysis (OMA) and clustering, followed by density-based time-domain filtering (tracking). The fundamental frequencies extracted are then fed to a dimensionality reduction block implemented through two different approaches: feature selection (intelligent multiplexer) that tries to estimate the most reliable frequencies based on the evaluation of some statistical features (i.e., mean value, variance, kurtosis), and feature extraction (auto-associative neural network (ANN)) that combine the fundamental frequencies to extract new damage sensitive features in a low dimensional feature space. Finally, one class classifier (OCC) algorithms perform anomaly detection, trained with standard condition points, and tested with normal and anomaly ones. In particular, a new anomaly detector strategy is proposed, namely one class classifier neural network two (OCCNN2), which exploit the classification capability of standard classifiers in an anomaly detection problem, finding the standard class (the boundary of the features space in normal operating conditions) through a two-step approach: coarse and fine boundary estimation. The coarse estimation uses classics OCC techniques, while the fine estimation is performed through a feedforward neural network (NN) trained that exploits the boundaries estimated in the coarse step. The detection algorithms vare then compared with known methods based on principal component analysis (PCA), kernel principal component analysis (KPCA), and auto-associative neural network (ANN). In many cases, the proposed solution increases the performance with respect to the standard OCC algorithms in terms of F1 score and accuracy. In particular, by evaluating the correct features, the anomaly can be detected with accuracy and an F1 score greater than 96% with the proposed method.

Keywords: anomaly detection, frequencies selection, modal analysis, neural network, sensor network, structural health monitoring, vibration measurement

Procedia PDF Downloads 94
13910 A Unified Deep Framework for Joint 3d Pose Estimation and Action Recognition from a Single Color Camera

Authors: Huy Hieu Pham, Houssam Salmane, Louahdi Khoudour, Alain Crouzil, Pablo Zegers, Sergio Velastin

Abstract:

We present a deep learning-based multitask framework for joint 3D human pose estimation and action recognition from color video sequences. Our approach proceeds along two stages. In the first, we run a real-time 2D pose detector to determine the precise pixel location of important key points of the body. A two-stream neural network is then designed and trained to map detected 2D keypoints into 3D poses. In the second, we deploy the Efficient Neural Architecture Search (ENAS) algorithm to find an optimal network architecture that is used for modeling the Spatio-temporal evolution of the estimated 3D poses via an image-based intermediate representation and performing action recognition. Experiments on Human3.6M, Microsoft Research Redmond (MSR) Action3D, and Stony Brook University (SBU) Kinect Interaction datasets verify the effectiveness of the proposed method on the targeted tasks. Moreover, we show that our method requires a low computational budget for training and inference.

Keywords: human action recognition, pose estimation, D-CNN, deep learning

Procedia PDF Downloads 118
13909 Urban Big Data: An Experimental Approach to Building-Value Estimation Using Web-Based Data

Authors: Sun-Young Jang, Sung-Ah Kim, Dongyoun Shin

Abstract:

Current real-estate value estimation, difficult for laymen, usually is performed by specialists. This paper presents an automated estimation process based on big data and machine-learning technology that calculates influences of building conditions on real-estate price measurement. The present study analyzed actual building sales sample data for Nonhyeon-dong, Gangnam-gu, Seoul, Korea, measuring the major influencing factors among the various building conditions. Further to that analysis, a prediction model was established and applied using RapidMiner Studio, a graphical user interface (GUI)-based tool for derivation of machine-learning prototypes. The prediction model is formulated by reference to previous examples. When new examples are applied, it analyses and predicts accordingly. The analysis process discerns the crucial factors effecting price increases by calculation of weighted values. The model was verified, and its accuracy determined, by comparing its predicted values with actual price increases.

Keywords: apartment complex, big data, life-cycle building value analysis, machine learning

Procedia PDF Downloads 352
13908 Bayesian Estimation under Different Loss Functions Using Gamma Prior for the Case of Exponential Distribution

Authors: Md. Rashidul Hasan, Atikur Rahman Baizid

Abstract:

The Bayesian estimation approach is a non-classical estimation technique in statistical inference and is very useful in real world situation. The aim of this paper is to study the Bayes estimators of the parameter of exponential distribution under different loss functions and then compared among them as well as with the classical estimator named maximum likelihood estimator (MLE). In our real life, we always try to minimize the loss and we also want to gather some prior information (distribution) about the problem to solve it accurately. Here the gamma prior is used as the prior distribution of exponential distribution for finding the Bayes estimator. In our study, we also used different symmetric and asymmetric loss functions such as squared error loss function, quadratic loss function, modified linear exponential (MLINEX) loss function and non-linear exponential (NLINEX) loss function. Finally, mean square error (MSE) of the estimators are obtained and then presented graphically.

Keywords: Bayes estimator, maximum likelihood estimator (MLE), modified linear exponential (MLINEX) loss function, Squared Error (SE) loss function, non-linear exponential (NLINEX) loss function

Procedia PDF Downloads 357
13907 Comprehensive Experimental Study to Determine Energy Dissipation of Nappe Flows on Stepped Chutes

Authors: Abdollah Ghasempour, Mohammad Reza Kavianpour, Majid Galoie

Abstract:

This study has investigated the fundamental parameters which have effective role on energy dissipation of nappe flows on stepped chutes in order to estimate an empirical relationship using dimensional analysis. To gain this goal, comprehensive experimental study on some large-scale physical models with various step geometries, slopes, discharges, etc. were carried out. For all models, hydraulic parameters such as velocity, pressure, water depth, flow regime and etc. were measured precisely. The effective parameters, then, could be determined by analysis of experimental data. Finally, a dimensional analysis was done in order to estimate an empirical relationship for evaluation of energy dissipation of nappe flows on stepped chutes. Because of using the large-scale physical models in this study, the empirical relationship is in very good agreement with the experimental results.

Keywords: nappe flow, energy dissipation, stepped chute, dimensional analysis

Procedia PDF Downloads 338
13906 A One Dimensional Particle in Cell Model for Excimer Lamps

Authors: W. Benstaali, A. Belasri

Abstract:

In this work we study a planar lamp filled with neon-xenon gas. We use a one-dimensional particle in a cell with Monte Carlo simulation (PIC-MCC) to investigate the effect xenon concentration on the energy deposited on excitation, ionization and ions. A Xe-Ne discharge is studied for a gas pressure of 400 torr. The results show an efficient Xe20-Ne mixture with an applied voltage of 1.2KV; the xenon excitation energy represents 65% form total energy dissipated in the discharge. We have also studied electrical properties and the energy balance a discharge for Xe50-Ne which needs a voltage of 2kv; the xenon energy is than more important.

Keywords: dielectric barrier discharge, efficiency, excitation, lamps

Procedia PDF Downloads 138
13905 Improvement of Direct Torque and Flux Control of Dual Stator Induction Motor Drive Using Intelligent Techniques

Authors: Kouzi Katia

Abstract:

This paper proposes a Direct Torque Control (DTC) algorithm of dual Stator Induction Motor (DSIM) drive using two approach intelligent techniques: Artificial Neural Network (ANN) approach replaces the switching table selector block of conventional DTC and Mamdani Fuzzy Logic controller (FLC) is used for stator resistance estimation. The fuzzy estimation method is based on an online stator resistance correction through the variations of stator current estimation error and its variation. The fuzzy logic controller gives the future stator resistance increment at the output. The main advantage of suggested algorithm control is to reduce the hardware complexity of conventional selectors, to avoid the drive instability that may occur in certain situation and ensure the tracking of the actual of the stator resistance. The effectiveness of the technique and the improvement of the whole system performance are proved by results.

Keywords: artificial neural network, direct torque control, dual stator induction motor, fuzzy logic estimator, switching table

Procedia PDF Downloads 318
13904 On the Blocked-off Finite-Volume Radiation Solutions in a Two-Dimensional Enclosure

Authors: Gyo Woo Lee, Man Young Kim

Abstract:

The blocked-off formulations for the analysis of radiative heat transfer are formulated and examined in order to find the solutions in a two-dimensional complex enclosure. The final discretization equations using the step scheme for spatial differencing practice are proposed with the additional source term to incorporate the blocked-off procedure. After introducing the implementation for inactive region into the general discretization equation, three different problems are examined to find the performance of the solution methods.

Keywords: radiative heat transfer, Finite Volume Method (FVM), blocked-off solution procedure, body-fitted coordinate

Procedia PDF Downloads 263
13903 Self –Engineering Strategy of Six Dimensional Inter-Subcultural Mental Images

Authors: Mostafa Jafari

Abstract:

How the people continually create and recreate the six dimensional inter- sub-cultural relationships from the strategic point of view? Can they engineer and direct it toward creating a set of peaceful subcultures? This paper answers to these questions. Our mental images shape the quantity and quality of our relationships. The six dimensions of mental images are: my mental image about myself, your mental image about yourself, my mental image about you, your mental image about me, my imagination about your image about me and your imagination about my mental image about you. Strategic engineering is dynamically shaping these images and imaginations.Methodology: This survey, which is based on object and the relation between the variables, is explanatory, correlative and quantitative. The target community members are 90 educated people from universities. The data has been collected through questionnaire and interview and has been analyzed by descriptive statistical techniques and qualitative method. Results: Our findings show that engineering and deliberatly managing the process of inter- sub-cultural transactions in the national and global level can enable us to continually reform a peaceful set of learner sub-culturals toward recreate a peaceful unit global Home.

Keywords: strategic engineering, mental image, six dimensional mental images strategy , cultural literacy, radar technique

Procedia PDF Downloads 374
13902 Robustified Asymmetric Logistic Regression Model for Global Fish Stock Assessment

Authors: Osamu Komori, Shinto Eguchi, Hiroshi Okamura, Momoko Ichinokawa

Abstract:

The long time-series data on population assessments are essential for global ecosystem assessment because the temporal change of biomass in such a database reflects the status of global ecosystem properly. However, the available assessment data usually have limited sample sizes and the ratio of populations with low abundance of biomass (collapsed) to those with high abundance (non-collapsed) is highly imbalanced. To allow for the imbalance and uncertainty involved in the ecological data, we propose a binary regression model with mixed effects for inferring ecosystem status through an asymmetric logistic model. In the estimation equation, we observe that the weights for the non-collapsed populations are relatively reduced, which in turn puts more importance on the small number of observations of collapsed populations. Moreover, we extend the asymmetric logistic regression model using propensity score to allow for the sample biases observed in the labeled and unlabeled datasets. It robustified the estimation procedure and improved the model fitting.

Keywords: double robust estimation, ecological binary data, mixed effect logistic regression model, propensity score

Procedia PDF Downloads 237
13901 Temperature Field Measurement of Premixed Landfill Gas Laminar Flame in a Cylindrical Slot Burner Using Mach-Zehnder Interferometry

Authors: Bahareh Najafian Ashrafi, Hossein Zeidabadinejad, Mehdi Ashjaee

Abstract:

The temperature field is a key factor of flame heat transfer rate and therefore should be measured accurately. In this study, the Mach-Zehnder Interferometry method is applied to measure the temperature field of premixed air/landfill gas (LFG60:60% CH4+40% CO2) laminar flame. The three-dimensional flame of cylindrical slot burner can assume to be two-dimensional due to the high aspect ratio (L/W=10) of the rectangular slot. So, the method converts two-dimensional flame to closed isothermal curves called fringes and the outer fringes temperature is measured by thermocouples. The experiments are carried out for Reynolds numbers and equivalence ratios ranging from 100 to 400 and 1.0 to 1.4, respectively. Results show that by increasing the equivalence ratio or Reynolds number, the flame height increases. The maximum flame temperature decreases by increasing the equivalence ratio but does not change considerably by changing the Reynolds number.

Keywords: landfill gas, Mach-Zehender interferometry, premix flame, slot burner, temperature filed

Procedia PDF Downloads 128
13900 Inverse Scattering of Two-Dimensional Objects Using an Enhancement Method

Authors: A.R. Eskandari, M.R. Eskandari

Abstract:

A 2D complete identification algorithm for dielectric and multiple objects immersed in air is presented. The employed technique consists of initially retrieving the shape and position of the scattering object using a linear sampling method and then determining the electric permittivity and conductivity of the scatterer using adjoint sensitivity analysis. This inversion algorithm results in high computational speed and efficiency, and it can be generalized for any scatterer structure. Also, this method is robust with respect to noise. The numerical results clearly show that this hybrid approach provides accurate reconstructions of various objects.

Keywords: inverse scattering, microwave imaging, two-dimensional objects, Linear Sampling Method (LSM)

Procedia PDF Downloads 360
13899 Digital Game Fostering Spatial Abilities for Children with Special Needs

Authors: Pedro Barros, Ana Breda, Eugenio Rocha, M. Isabel Santos

Abstract:

As visual and spatial awareness develops, children apprehension of the concept of direction, (relative) distance and (relative) location materializes. Here we present the educational inclusive digital game ORIESPA, under development by the Thematic Line Geometrix, for children aged between 6 and 10 years old, aiming the improvement of their visual and spatial awareness. Visual-spatial abilities are of crucial importance to succeed in many everyday life tasks. Unavoidable in the technological age we are living in, they are essential in many fields of study as, for instance, mathematics.The game, set on a 2D/3D environment, focusses in tasks/challenges on the following categories (1) static orientation of the subject and object, requiring an understanding of the notions of up–down, left–right, front–back, higher-lower or nearer-farther; (2) interpretation of perspectives of three-dimensional objects, requiring the understanding of 2D and 3D representations of three-dimensional objects; and (3) orientation of the subject in real space, requiring the reading and interpreting of itineraries. In ORIESPA, simpler tasks are based on a quadrangular grid, where the front-back and left-right directions and the rotations of 90º, 180º and 270º play the main requirements. The more complex ones are produced on a cubic grid adding the up and down movements. In the first levels, the game's mechanics regarding the reading and interpreting maps (from point A to point B) is based on map routes, following a given set of instructions. In higher levels, the player must produce a list of instructions taking the game character to the desired destination, avoiding obstacles. Being an inclusive game the user has the possibility to interact through the mouse (point and click with a single button), the keyboard (small set of well recognized keys) or a Kinect device (using simple gesture moves). The character control requires the action on buttons corresponding to movements in 2D and 3D environments. Buttons and instructions are also complemented with text, sound and sign language.

Keywords: digital game, inclusion, itinerary, spatial ability

Procedia PDF Downloads 151
13898 Infodemic Detection on Social Media with a Multi-Dimensional Deep Learning Framework

Authors: Raymond Xu, Cindy Jingru Wang

Abstract:

Social media has become a globally connected and influencing platform. Social media data, such as tweets, can help predict the spread of pandemics and provide individuals and healthcare providers early warnings. Public psychological reactions and opinions can be efficiently monitored by AI models on the progression of dominant topics on Twitter. However, statistics show that as the coronavirus spreads, so does an infodemic of misinformation due to pandemic-related factors such as unemployment and lockdowns. Social media algorithms are often biased toward outrage by promoting content that people have an emotional reaction to and are likely to engage with. This can influence users’ attitudes and cause confusion. Therefore, social media is a double-edged sword. Combating fake news and biased content has become one of the essential tasks. This research analyzes the variety of methods used for fake news detection covering random forest, logistic regression, support vector machines, decision tree, naive Bayes, BoW, TF-IDF, LDA, CNN, RNN, LSTM, DeepFake, and hierarchical attention network. The performance of each method is analyzed. Based on these models’ achievements and limitations, a multi-dimensional AI framework is proposed to achieve higher accuracy in infodemic detection, especially pandemic-related news. The model is trained on contextual content, images, and news metadata.

Keywords: artificial intelligence, fake news detection, infodemic detection, image recognition, sentiment analysis

Procedia PDF Downloads 205
13897 The Effect of Three-Dimensional Morphology on Vulnerability Assessment of Atherosclerotic Plaque

Authors: M. Zareh, H. Mohammadi, B. Naser

Abstract:

Atherosclerotic plaque rupture is the main trigger of heart attack and brain stroke which are the leading cause of death in developed countries. Better understanding of rupture-prone plaque can help clinicians detect vulnerable plaques- rupture prone or instable plaques- and apply immediate medical treatment to prevent these life-threatening cardiovascular events. Therefore, there are plenty of studies addressing disclosure of vulnerable plaques properties. Necrotic core and fibrous tissue are two major tissues constituting atherosclerotic plaque; using histopathological and numerical approaches, many studies have demonstrated that plaque rupture is strongly associated with a large necrotic core and a thin fibrous cap, two morphological characteristic which can be acquired by two-dimensional imaging of atherosclerotic plaque present in coronary and carotid arteries. Plaque rupture is widely considered as a mechanical failure inside plaque tissue; this failure occurs when the stress within plaque excesses the strength of tissue material; hence, finite element method, a strong numerical approach, has been extensively applied to estimate stress distribution within plaques with different compositions which is then used for assessment of various vulnerability characteristics including plaque morphology, material properties and blood pressure. This study aims to evaluate significance of three-dimensional morphology on vulnerability degree of atherosclerotic plaque. To reach this end, different two-dimensional geometrical models of atherosclerotic plaques are considered based on available data and named Main 2D Models (M2M). Then, for each of these M2Ms, two three-dimensional idealistic models are created. These two 3D models represent two possible three-dimensional morphologies which might exist for a plaque with similar 2D morphology to one of M2Ms. Finite element method is employed to estimate stress, von-Mises stress, within each 3D models. Results indicate that for each M2Ms stress can significantly varies due to possible 3D morphological changes in that plaque. Also, our results show that an atherosclerotic plaque with thick cap may experience rupture if it has a critical 3D morphology. This study highlights the effect of 3D geometry of plaque on its instability degree and suggests that 3D morphology of plaque might be necessary to more effectively and accurately assess atherosclerotic plaque vulnerability.

Keywords: atherosclerotic plaque, plaque rupture, finite element method, 3D model

Procedia PDF Downloads 287
13896 Carbohydrate Intake Estimation in Type I Diabetic Patients Described by UVA/Padova Model

Authors: David A. Padilla, Rodolfo Villamizar

Abstract:

In recent years, closed loop control strategies have been developed in order to establish a healthy glucose profile in type 1 diabetic mellitus (T1DM) patients. However, the controller itself is unable to define a suitable reference trajectory for glucose. In this paper, a control strategy Is proposed where the shape of the reference trajectory is generated bases in the amount of carbohydrates present during the digestive process, due to the effect of carbohydrate intake. Since there no exists a sensor to measure the amount of carbohydrates consumed, an estimator is proposed. Thus this paper presents the entire process of designing a carbohydrate estimator, which allows estimate disturbance for a predictive controller (MPC) in a T1MD patient, the estimation will be used to establish a profile of reference and improve the response of the controller by providing the estimated information of ingested carbohydrates. The dynamics of the diabetic model used are due to the equations described by the UVA/Padova model of the T1DMS simulator, the system was developed and simulated in Simulink, taking into account the noise and limitations of the glucose control system actuators.

Keywords: estimation, glucose control, predictive controller, MPC, UVA/Padova

Procedia PDF Downloads 232
13895 High-Fidelity 1D Dynamic Model of a Hydraulic Servo Valve Using 3D Computational Fluid Dynamics and Electromagnetic Finite Element Analysis

Authors: D. Henninger, A. Zopey, T. Ihde, C. Mehring

Abstract:

The dynamic performance of a 4-way solenoid operated hydraulic spool valve has been analyzed by means of a one-dimensional modeling approach capturing flow, magnetic and fluid forces, valve inertia forces, fluid compressibility, and damping. Increased model accuracy was achieved by analyzing the detailed three-dimensional electromagnetic behavior of the solenoids and flow behavior through the spool valve body for a set of relevant operating conditions, thereby allowing the accurate mapping of flow and magnetic forces on the moving valve body, in lieu of representing the respective forces by lower-order models or by means of simplistic textbook correlations. The resulting high-fidelity one-dimensional model provided the basis for specific and timely design modification eliminating experimentally observed valve oscillations.

Keywords: dynamic performance model, high-fidelity model, 1D-3D decoupled analysis, solenoid-operated hydraulic servo valve, CFD and electromagnetic FEA

Procedia PDF Downloads 151
13894 Estimation of Missing Values in Aggregate Level Spatial Data

Authors: Amitha Puranik, V. S. Binu, Seena Biju

Abstract:

Missing data is a common problem in spatial analysis especially at the aggregate level. Missing can either occur in covariate or in response variable or in both in a given location. Many missing data techniques are available to estimate the missing data values but not all of these methods can be applied on spatial data since the data are autocorrelated. Hence there is a need to develop a method that estimates the missing values in both response variable and covariates in spatial data by taking account of the spatial autocorrelation. The present study aims to develop a model to estimate the missing data points at the aggregate level in spatial data by accounting for (a) Spatial autocorrelation of the response variable (b) Spatial autocorrelation of covariates and (c) Correlation between covariates and the response variable. Estimating the missing values of spatial data requires a model that explicitly account for the spatial autocorrelation. The proposed model not only accounts for spatial autocorrelation but also utilizes the correlation that exists between covariates, within covariates and between a response variable and covariates. The precise estimation of the missing data points in spatial data will result in an increased precision of the estimated effects of independent variables on the response variable in spatial regression analysis.

Keywords: spatial regression, missing data estimation, spatial autocorrelation, simulation analysis

Procedia PDF Downloads 346
13893 Astronomical Object Classification

Authors: Alina Muradyan, Lina Babayan, Arsen Nanyan, Gohar Galstyan, Vigen Khachatryan

Abstract:

We present a photometric method for identifying stars, galaxies and quasars in multi-color surveys, which uses a library of ∼> 65000 color templates for comparison with observed objects. The method aims for extracting the information content of object colors in a statistically correct way, and performs a classification as well as a redshift estimation for galaxies and quasars in a unified approach based on the same probability density functions. For the redshift estimation, we employ an advanced version of the Minimum Error Variance estimator which determines the redshift error from the redshift dependent probability density function itself. The method was originally developed for the Calar Alto Deep Imaging Survey (CADIS), but is now used in a wide variety of survey projects. We checked its performance by spectroscopy of CADIS objects, where the method provides high reliability (6 errors among 151 objects with R < 24), especially for the quasar selection, and redshifts accurate within σz ≈ 0.03 for galaxies and σz ≈ 0.1 for quasars. For an optimization of future survey efforts, a few model surveys are compared, which are designed to use the same total amount of telescope time but different sets of broad-band and medium-band filters. Their performance is investigated by Monte-Carlo simulations as well as by analytic evaluation in terms of classification and redshift estimation. If photon noise were the only error source, broad-band surveys and medium-band surveys should perform equally well, as long as they provide the same spectral coverage. In practice, medium-band surveys show superior performance due to their higher tolerance for calibration errors and cosmic variance. Finally, we discuss the relevance of color calibration and derive important conclusions for the issues of library design and choice of filters. The calibration accuracy poses strong constraints on an accurate classification, which are most critical for surveys with few, broad and deeply exposed filters, but less severe for surveys with many, narrow and less deep filters.

Keywords: VO, ArVO, DFBS, FITS, image processing, data analysis

Procedia PDF Downloads 43
13892 An Atomistic Approach to Define Continuum Mechanical Quantities in One Dimensional Nanostructures at Finite Temperature

Authors: Smriti, Ajeet Kumar

Abstract:

We present a variant of the Irving-Kirkwood procedure to obtain the microscopic expressions of the cross-section averaged continuum fields such as internal force and moment in one-dimensional nanostructures in the non-equilibrium setting. In one-dimensional continuum theories for slender bodies, we deal with quantities such as mass, linear momentum, angular momentum, and strain energy densities, all defined per unit length. These quantities are obtained by integrating the corresponding pointwise (per unit volume) quantities over the cross-section of the slender body. However, no well-defined cross-section exists for these nanostructures at finite temperature. We thus define the cross-section of a nanorod to be an infinite plane which is fixed in space even when time progresses and defines the above continuum quantities by integrating the pointwise microscopic quantities over this infinite plane. The method yields explicit expressions of both the potential and kinetic parts of the above quantities. We further specialize in these expressions for helically repeating one-dimensional nanostructures in order to use them in molecular dynamics study of extension, torsion, and bending of such nanostructures. As, the Irving-Kirkwood procedure does not yield expressions of stiffnesses, we resort to a thermodynamic equilibrium approach to obtain the expressions of axial force, twisting moment, bending moment, and the associated stiffnesses by taking the first and second derivatives of the Helmholtz free energy with respect to conjugate strain measures. The equilibrium approach yields expressions independent of kinetic terms. We then establish the equivalence of the expressions obtained using the two approaches. The derived expressions are used to understand the extension, torsion, and bending of single-walled carbon nanotubes at non-zero temperatures.

Keywords: thermoelasticity, molecular dynamics, one dimensional nanostructures, nanotube buckling

Procedia PDF Downloads 106
13891 Generalized Approach to Linear Data Transformation

Authors: Abhijith Asok

Abstract:

This paper presents a generalized approach for the simple linear data transformation, Y=bX, through an integration of multidimensional coordinate geometry, vector space theory and polygonal geometry. The scaling is performed by adding an additional ’Dummy Dimension’ to the n-dimensional data, which helps plot two dimensional component-wise straight lines on pairs of dimensions. The end result is a set of scaled extensions of observations in any of the 2n spatial divisions, where n is the total number of applicable dimensions/dataset variables, created by shifting the n-dimensional plane along the ’Dummy Axis’. The derived scaling factor was found to be dependent on the coordinates of the common point of origin for diverging straight lines and the plane of extension, chosen on and perpendicular to the ’Dummy Axis’, respectively. This result indicates the geometrical interpretation of a linear data transformation and hence, opportunities for a more informed choice of the factor ’b’, based on a better choice of these coordinate values. The paper follows on to identify the effect of this transformation on certain popular distance metrics, wherein for many, the distance metric retained the same scaling factor as that of the features.

Keywords: data transformation, dummy dimension, linear transformation, scaling

Procedia PDF Downloads 279
13890 The Algerian Experience in Developing Higher Education in the Country in Light of Modern Technology: Challenges and Prospects

Authors: Mohammed Messaoudi

Abstract:

The higher education sector in Algeria has witnessed in recent years a remarkable transformation, as it witnessed the integration of institutions within the modern technological environment and harnessing all appropriate mechanisms to raise the level of education and the level of training. Observers and those interested that it is necessary for the Algerian university to enter this field, especially with the efforts that seek to employ modern technology in the sector and encourage investment in this field, in addition to the state’s keenness to move towards building a path to benefit from modern technology, and to encourage energies in light of a reality that carries many Aspirations and challenges by achieving openness to the new digital environment and keeping pace with the ranks of international universities. Higher education is one of the engines of development for societies, as it is a vital field for the transfer of knowledge and scientific expertise, and the university is at the top of the comprehensive educational system for various disciplines in light of the achievement of a multi-dimensional educational system, and amid the integration of three basic axes that establish the sound educational process (teaching, research, relevant outputs efficiency), and according to a clear strategy that monitors the advancement of academic work, and works on developing its future directions to achieve development in this field. The Algerian University is considered one of the service institutions that seeks to find the optimal mechanisms to keep pace with the changes of the times, as it has become necessary for the university to enter the technological space and thus ensure the quality of education in it and achieve the required empowerment by dedicating a structure that matches the requirements of the challenges on which the sector is based, amid unremitting efforts to develop the capabilities. He sought to harness the mechanisms of communication and information technology and achieve transformation at the level of the higher education sector with what is called higher education technology. The conceptual framework of information and communication technology at the level of higher education institutions in Algeria is determined through the factors of organization, factors of higher education institutions, characteristics of the professor, characteristics of students, the outcomes of the educational process, and there is a relentless pursuit to achieve a positive interaction between these axes as they are basic components on which the success and achievement of higher education are based for his goals.

Keywords: Information and communication technology, Algerian university, scientific and cognitive development, challenges

Procedia PDF Downloads 59
13889 Improved Multi-Channel Separation Algorithm for Satellite-Based Automatic Identification System Signals Based on Artificial Bee Colony and Adaptive Moment Estimation

Authors: Peng Li, Luan Wang, Haifeng Fei, Renhong Xie, Yibin Rui, Shanhong Guo

Abstract:

The applications of satellite-based automatic identification system (S-AIS) pave the road for wide-range maritime traffic monitoring and management. But the coverage of satellite’s view includes multiple AIS self-organizing networks, which leads to the collision of AIS signals from different cells. The contribution of this work is to propose an improved multi-channel blind source separation algorithm based on Artificial Bee Colony (ABC) and advanced stochastic optimization to perform separation of the mixed AIS signals. The proposed approach adopts modified ABC algorithm to get an optimized initial separating matrix, which can expedite the initialization bias correction, and utilizes the Adaptive Moment Estimation (Adam) to update the separating matrix by adjusting the learning rate for each parameter dynamically. Simulation results show that the algorithm can speed up convergence and lead to better performance in separation accuracy.

Keywords: satellite-based automatic identification system, blind source separation, artificial bee colony, adaptive moment estimation

Procedia PDF Downloads 159
13888 A Theorem Related to Sample Moments and Two Types of Moment-Based Density Estimates

Authors: Serge B. Provost

Abstract:

Numerous statistical inference and modeling methodologies are based on sample moments rather than the actual observations. A result justifying the validity of this approach is introduced. More specifically, it will be established that given the first n moments of a sample of size n, one can recover the original n sample points. This implies that a sample of size n and its first associated n moments contain precisely the same amount of information. However, it is efficient to make use of a limited number of initial moments as most of the relevant distributional information is included in them. Two types of density estimation techniques that rely on such moments will be discussed. The first one expresses a density estimate as the product of a suitable base density and a polynomial adjustment whose coefficients are determined by equating the moments of the density estimate to the sample moments. The second one assumes that the derivative of the logarithm of a density function can be represented as a rational function. This gives rise to a system of linear equations involving sample moments, the density estimate is then obtained by solving a differential equation. Unlike kernel density estimation, these methodologies are ideally suited to model ‘big data’ as they only require a limited number of moments, irrespective of the sample size. What is more, they produce simple closed form expressions that are amenable to algebraic manipulations. They also turn out to be more accurate as will be shown in several illustrative examples.

Keywords: density estimation, log-density, polynomial adjustments, sample moments

Procedia PDF Downloads 130