Search results for: Generalized autoregressive score model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7933

Search results for: Generalized autoregressive score model

7663 Assessing the Effect of the Shift of Rural Labor towards Non-Agricultural Sectors on Rice Cultivation in the African Environment: Evidence from Sierra Leone

Authors: Alhaji M. H Conteh, Xiangbin Yan, Alfred V Gborie

Abstract:

The crop rice is the staple food of most Sierra Leone with no close substitute. However, its cultivation has been on its last legs over the years. The decline in the domestic rice cultivation has had vicious socio-economic implications such as hiking consumer prices, balance of payment dilemmas with debt burden. The objective of this study is thus, to assess the effect of the shift of rural labour towards non-agricultural sectors on rice cultivation. The tools utilized for analyzing the problem under consideration involved a thorough descriptive statistics and generalized linear model using OLS technique. Increased rural population was established positive and significant in affecting rice cultivation. Fertilizer utilization was insignificant in rice cultivation. For reducing the shift of rural labor force towards nonagricultural sectors, the government should make the agricultural sector very lucrative.

Keywords: Regression Model, Rice Cultivation, Rural Migration, Sierra Leone.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1748
7662 Transient Thermal Stresses of Functionally Graded Thick Hollow Cylinder under the Green-Lindsay Model

Authors: Tariq T. Darabseh

Abstract:

The transient thermoelastic response of thick hollow cylinder made of functionally graded material under thermal loading is studied. The generalized coupled thermoelasticity based on the Green-Lindsay model is used. The thermal and mechanical properties of the functionally graded material are assumed to be varied in the radial direction according to a power law variation as a function of the volume fractions of the constituents. The thermal and elastic governing equations are solved by using Galerkin finite element method. All the finite element calculations were done by using commercial finite element program FlexPDE. The transient temperature, radial displacement, and thermal stresses distribution through the radial direction of the cylinder are plotted.

Keywords: Finite element method, thermal stresses, Green-Lindsay theory, functionally graded material.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1963
7661 An Approach for Transient Response Calculation of large Nonproportionally Damped Structures using Component Mode Synthesis

Authors: Alexander A. Muravyov

Abstract:

A minimal complexity version of component mode synthesis is presented that requires simplified computer programming, but still provides adequate accuracy for modeling lower eigenproperties of large structures and their transient responses. The novelty is that a structural separation into components is done along a plane/surface that exhibits rigid-like behavior, thus only normal modes of each component is sufficient to use, without computing any constraint, attachment, or residual-attachment modes. The approach requires only such input information as a few (lower) natural frequencies and corresponding undamped normal modes of each component. A novel technique is shown for formulation of equations of motion, where a double transformation to generalized coordinates is employed and formulation of nonproportional damping matrix in generalized coordinates is shown.

Keywords: component mode synthesis, finite element models, transient response, nonproportional damping

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1772
7660 Using Simulation Modeling Approach to Predict USMLE Steps 1 and 2 Performances

Authors: Chau-Kuang Chen, John Hughes, Jr., A. Dexter Samuels

Abstract:

The prediction models for the United States Medical Licensure Examination (USMLE) Steps 1 and 2 performances were constructed by the Monte Carlo simulation modeling approach via linear regression. The purpose of this study was to build robust simulation models to accurately identify the most important predictors and yield the valid range estimations of the Steps 1 and 2 scores. The application of simulation modeling approach was deemed an effective way in predicting student performances on licensure examinations. Also, sensitivity analysis (a/k/a what-if analysis) in the simulation models was used to predict the magnitudes of Steps 1 and 2 affected by changes in the National Board of Medical Examiners (NBME) Basic Science Subject Board scores. In addition, the study results indicated that the Medical College Admission Test (MCAT) Verbal Reasoning score and Step 1 score were significant predictors of the Step 2 performance. Hence, institutions could screen qualified student applicants for interviews and document the effectiveness of basic science education program based on the simulation results.

Keywords: Prediction Model, Sensitivity Analysis, Simulation Method, USMLE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1424
7659 Model Updating-Based Approach for Damage Prognosis in Frames via Modal Residual Force

Authors: Gholamreza Ghodrati Amiri, Mojtaba Jafarian Abyaneh, Ali Zare Hosseinzadeh

Abstract:

This paper presents an effective model updating strategy for damage localization and quantification in frames by defining damage detection problem as an optimization issue. A generalized version of the Modal Residual Force (MRF) is employed for presenting a new damage-sensitive cost function. Then, Grey Wolf Optimization (GWO) algorithm is utilized for solving suggested inverse problem and the global extremums are reported as damage detection results. The applicability of the presented method is investigated by studying different damage patterns on the benchmark problem of the IASC-ASCE, as well as a planar shear frame structure. The obtained results emphasize good performance of the method not only in free-noise cases, but also when the input data are contaminated with different levels of noises.

Keywords: Frame, grey wolf optimization algorithm, modal residual force, structural damage detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1454
7658 Bridging the Mental Gap between Convolution Approach and Compartmental Modeling in Functional Imaging: Typical Embedding of an Open Two-Compartment Model into the Systems Theory Approach of Indicator Dilution Theory

Authors: Gesine Hellwig

Abstract:

Functional imaging procedures for the non-invasive assessment of tissue microcirculation are highly requested, but require a mathematical approach describing the trans- and intercapillary passage of tracer particles. Up to now, two theoretical, for the moment different concepts have been established for tracer kinetic modeling of contrast agent transport in tissues: pharmacokinetic compartment models, which are usually written as coupled differential equations, and the indicator dilution theory, which can be generalized in accordance with the theory of lineartime- invariant (LTI) systems by using a convolution approach. Based on mathematical considerations, it can be shown that also in the case of an open two-compartment model well-known from functional imaging, the concentration-time course in tissue is given by a convolution, which allows a separation of the arterial input function from a system function being the impulse response function, summarizing the available information on tissue microcirculation. Due to this reason, it is possible to integrate the open two-compartment model into the system-theoretic concept of indicator dilution theory (IDT) and thus results known from IDT remain valid for the compartment approach. According to the long number of applications of compartmental analysis, even for a more general context similar solutions of the so-called forward problem can already be found in the extensively available appropriate literature of the seventies and early eighties. Nevertheless, to this day, within the field of biomedical imaging – not from the mathematical point of view – there seems to be a trench between both approaches, which the author would like to get over by exemplary analysis of the well-known model.

Keywords: Functional imaging, Tracer kinetic modeling, LTIsystem, Indicator dilution theory / convolution approach, Two-Compartment model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1374
7657 Biaxial Buckling of Single Layer Graphene Sheet Based on Nonlocal Plate Model and Molecular Dynamics Simulation

Authors: R. Pilafkan, M. Kaffash Irzarahimi, S. F. Asbaghian Namin

Abstract:

The biaxial buckling behavior of single-layered graphene sheets (SLGSs) is studied in the present work. To consider the size-effects in the analysis, Eringen’s nonlocal elasticity equations are incorporated into classical plate theory (CLPT). A Generalized Differential Quadrature Method (GDQM) approach is utilized and numerical solutions for the critical buckling loads are obtained. Then, molecular dynamics (MD) simulations are performed for a series of zigzag SLGSs with different side-lengths and with various boundary conditions, the results of which are matched with those obtained by the nonlocal plate model to numerical the appropriate values of nonlocal parameter relevant to each type of boundary conditions.

Keywords: Biaxial buckling, single-layered graphene sheets, nonlocal elasticity, molecular dynamics simulation, classical plate theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1054
7656 Mining of Interesting Prediction Rules with Uniform Two-Level Genetic Algorithm

Authors: Bilal Alatas, Ahmet Arslan

Abstract:

The main goal of data mining is to extract accurate, comprehensible and interesting knowledge from databases that may be considered as large search spaces. In this paper, a new, efficient type of Genetic Algorithm (GA) called uniform two-level GA is proposed as a search strategy to discover truly interesting, high-level prediction rules, a difficult problem and relatively little researched, rather than discovering classification knowledge as usual in the literatures. The proposed method uses the advantage of uniform population method and addresses the task of generalized rule induction that can be regarded as a generalization of the task of classification. Although the task of generalized rule induction requires a lot of computations, which is usually not satisfied with the normal algorithms, it was demonstrated that this method increased the performance of GAs and rapidly found interesting rules.

Keywords: Classification rule mining, data mining, genetic algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1560
7655 A Semi-Implicit Phase Field Model for Droplet Evolution

Authors: M. H. Kazemi, D. Salac

Abstract:

A semi-implicit phase field method for droplet evolution is proposed. Using the phase field Cahn-Hilliard equation, we are able to track the interface in multiphase flow. The idea of a semi-implicit finite difference scheme is reviewed and employed to solve two nonlinear equations, including the Navier-Stokes and the Cahn-Hilliard equations. The use of a semi-implicit method allows us to have larger time steps compared to explicit schemes. The governing equations are coupled and then solved by a GMRES solver (generalized minimal residual method) using modified Gram-Schmidt orthogonalization. To show the validity of the method, we apply the method to the simulation of a rising droplet, a leaky dielectric drop and the coalescence of drops. The numerical solutions to the phase field model match well with existing solutions over a defined range of variables.

Keywords: Coalescence, leaky dielectric, numerical method, phase field, rising droplet, semi-implicit method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 840
7654 A Boundary Backstepping Control Design for 2-D, 3-D and N-D Heat Equation

Authors: Aziz Sezgin

Abstract:

We consider the problem of stabilization of an unstable heat equation in a 2-D, 3-D and generally n-D domain by deriving a generalized backstepping boundary control design methodology. To stabilize the systems, we design boundary backstepping controllers inspired by the 1-D unstable heat equation stabilization procedure. We assume that one side of the boundary is hinged and the other side is controlled for each direction of the domain. Thus, controllers act on two boundaries for 2-D domain, three boundaries for 3-D domain and ”n” boundaries for n-D domain. The main idea of the design is to derive ”n” controllers for each of the dimensions by using ”n” kernel functions. Thus, we obtain ”n” controllers for the ”n” dimensional case. We use a transformation to change the system into an exponentially stable ”n” dimensional heat equation. The transformation used in this paper is a generalized Volterra/Fredholm type with ”n” kernel functions for n-D domain instead of the one kernel function of 1-D design.

Keywords: Backstepping, boundary control, 2-D, 3-D, n-D heat equation, distributed parameter systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1621
7653 A Survey of Model Comparison Strategies and Techniques in Model Driven Engineering

Authors: Junaid Rashid, Waqar Mehmood, Muhammad Wasif Nisar

Abstract:

This survey paper shows the recent state of model comparison as it’s applies to Model Driven engineering. In Model Driven Engineering to calculate the difference between the models is a very important and challenging task. There are number of tasks involved in model differencing that firstly starts with identifying and matching the elements of the model. In this paper, we discuss how model matching is accomplished, the strategies, techniques and the types of the model. We also discuss the future direction. We found out that many of the latest model comparison strategies are geared near enabling Meta model and similarity based matching. Therefore model versioning is the most dominant application of the model comparison. Recently to work on comparison for versioning has begun to deteriorate, giving way to different applications. Ultimately there is wide change among the tools in the measure of client exertion needed to perform model comparisons, as some require more push to encourage more sweeping statement and expressive force.

Keywords: Model comparison, model clone detection, model versioning, EMF Model, model diff.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2130
7652 Bond Graph Modeling of Mechanical Dynamics of an Excavator for Hydraulic System Analysis and Design

Authors: Mutuku Muvengei, John Kihiu

Abstract:

This paper focuses on the development of bond graph dynamic model of the mechanical dynamics of an excavating mechanism previously designed to be used with small tractors, which are fabricated in the Engineering Workshops of Jomo Kenyatta University of Agriculture and Technology. To develop a mechanical dynamics model of the manipulator, forward recursive equations similar to those applied in iterative Newton-Euler method were used to obtain kinematic relationships between the time rates of joint variables and the generalized cartesian velocities for the centroids of the links. Representing the obtained kinematic relationships in bondgraphic form, while considering the link weights and momenta as the elements led to a detailed bond graph model of the manipulator. The bond graph method was found to reduce significantly the number of recursive computations performed on a 3 DOF manipulator for a mechanical dynamic model to result, hence indicating that bond graph method is more computationally efficient than the Newton-Euler method in developing dynamic models of 3 DOF planar manipulators. The model was verified by comparing the joint torque expressions of a two link planar manipulator to those obtained using Newton- Euler and Lagrangian methods as analyzed in robotic textbooks. The expressions were found to agree indicating that the model captures the aspects of rigid body dynamics of the manipulator. Based on the model developed, actuator sizing and valve sizing methodologies were developed and used to obtain the optimal sizes of the pistons and spool valve ports respectively. It was found that using the pump with the sized flow rate capacity, the engine of the tractor is able to power the excavating mechanism in digging a sandy-loom soil.

Keywords: Actuators, bond graphs, inverse dynamics, recursive equations, quintic polynomial trajectory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2846
7651 BIDENS: Iterative Density Based Biclustering Algorithm With Application to Gene Expression Analysis

Authors: Mohamed A. Mahfouz, M. A. Ismail

Abstract:

Biclustering is a very useful data mining technique for identifying patterns where different genes are co-related based on a subset of conditions in gene expression analysis. Association rules mining is an efficient approach to achieve biclustering as in BIMODULE algorithm but it is sensitive to the value given to its input parameters and the discretization procedure used in the preprocessing step, also when noise is present, classical association rules miners discover multiple small fragments of the true bicluster, but miss the true bicluster itself. This paper formally presents a generalized noise tolerant bicluster model, termed as μBicluster. An iterative algorithm termed as BIDENS based on the proposed model is introduced that can discover a set of k possibly overlapping biclusters simultaneously. Our model uses a more flexible method to partition the dimensions to preserve meaningful and significant biclusters. The proposed algorithm allows discovering biclusters that hard to be discovered by BIMODULE. Experimental study on yeast, human gene expression data and several artificial datasets shows that our algorithm offers substantial improvements over several previously proposed biclustering algorithms.

Keywords: Machine learning, biclustering, bi-dimensional clustering, gene expression analysis, data mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1915
7650 Generalized Rough Sets Applied to Graphs Related to Urban Problems

Authors: Mihai Rebenciuc, Simona Mihaela Bibic

Abstract:

Branch of modern mathematics, graphs represent instruments for optimization and solving practical applications in various fields such as economic networks, engineering, network optimization, the geometry of social action, generally, complex systems including contemporary urban problems (path or transport efficiencies, biourbanism, & c.). In this paper is studied the interconnection of some urban network, which can lead to a simulation problem of a digraph through another digraph. The simulation is made univoc or more general multivoc. The concepts of fragment and atom are very useful in the study of connectivity in the digraph that is simulation - including an alternative evaluation of k- connectivity. Rough set approach in (bi)digraph which is proposed in premier in this paper contribute to improved significantly the evaluation of k-connectivity. This rough set approach is based on generalized rough sets - basic facts are presented in this paper.

Keywords: (Bi)digraphs, rough set theory, systems of interacting agents, complex systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1145
7649 The European Union’s Generalized System of Preferences (GSP) and the Prospect of a Unified Database

Authors: Iasha Meskhia, Rusudan Seturidze

Abstract:

Free access for Georgian goods to the EU markets is one of the important factors for Georgia’s economic development, attraction of investments and raising the standard of living. The European Union is the most important trade partner for Georgia. Great experience has been accumulated with respect to removing trade barriers between Georgia and the European Union. Despite it, certain problems still persist.

In the present article, we have reviewed the systems of preferences with the European Union, the EU’s Generalized System of Preferences (GSP) and the essence of ongoing reform; we have assessed weak and strong sides of relations established between the European Union and Georgia in this regard; analyzed Georgia’s export and import over the past years; also reviewed the prospect of a unified database; established existing and anticipated positive and negative factors. Based on the analysis, we have provided the relevant recommendations. 

Keywords: EU-Georgia trade, EU’s GSP reform, Georgia’s export-import, REX system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2174
7648 Determining Optimal Production Plan by Revised Surrogate Worth Trade-off Method

Authors: Tunjo Peric, Zoran Babic

Abstract:

The authors of this work indicate by means of a concrete example that it is possible to apply efficaciously the method of multiple criteria programming in dealing with the problem of determining the optimal production plan for a certain period of time. The work presents: (1) the selection of optimization criteria, (2) the setting of the problem of determining an optimal production plan, (3) the setting of the model of multiple criteria programming in finding a solution to a given problem, (4) the revised surrogate trade-off method, (5) generalized multicriteria model for solving production planning problem and problem of choosing technological variants in the metal manufacturing industry. In the final part of this work the authors reflect on the application of the method of multiple criteria programming while determining the optimal production plan in manufacturing enterprises.

Keywords: multi-criteria programming, production planning, technological variant, Surrogate Worth Trade-off Method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1819
7647 A Medical Resource Forecasting Model for Emergency Room Patients with Acute Hepatitis

Authors: R. J. Kuo, W. C. Cheng, W. C. Lien, T. J. Yang

Abstract:

Taiwan is a hyper endemic area for the Hepatitis B virus (HBV). The estimated total number of HBsAg carriers in the general population who are more than 20 years old is more than 3 million. Therefore, a case record review is conducted from January 2003 to June 2007 for all patients with a diagnosis of acute hepatitis who were admitted to the Emergency Department (ED) of a well-known teaching hospital. The cost for the use of medical resources is defined as the total medical fee. In this study, principal component analysis (PCA) is firstly employed to reduce the number of dimensions. Support vector regression (SVR) and artificial neural network (ANN) are then used to develop the forecasting model. A total of 117 patients meet the inclusion criteria. 61% patients involved in this study are hepatitis B related. The computational result shows that the proposed PCA-SVR model has superior performance than other compared algorithms. In conclusion, the Child-Pugh score and echogram can both be used to predict the cost of medical resources for patients with acute hepatitis in the ED.

Keywords: Acute hepatitis, Medical resource cost, Artificial neural network, Support vector regression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1880
7646 Dynamic Analysis of Composite Doubly Curved Panels with Variable Thickness

Authors: I. Algul, G. Akgun, H. Kurtaran

Abstract:

Dynamic analysis of composite doubly curved panels with variable thickness subjected to different pulse types using Generalized Differential Quadrature method (GDQ) is presented in this study. Panels with variable thickness are used in the construction of aerospace and marine industry. Giving variable thickness to panels can allow the designer to get optimum structural efficiency. For this reason, estimating the response of variable thickness panels is very important to design more reliable structures under dynamic loads. Dynamic equations for composite panels with variable thickness are obtained using virtual work principle. Partial derivatives in the equation of motion are expressed with GDQ and Newmark average acceleration scheme is used for temporal discretization. Several examples are used to highlight the effectiveness of the proposed method. Results are compared with finite element method. Effects of taper ratios, boundary conditions and loading type on the response of composite panel are investigated.

Keywords: Generalized differential quadrature method, doubly curved panels, laminated composite materials, small displacement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 897
7645 On Solution of Interval Valued Intuitionistic Fuzzy Assignment Problem Using Similarity Measure and Score Function

Authors: Gaurav Kumar, Rakesh Kumar Bajaj

Abstract:

The primary objective of the paper is to propose a new method for solving assignment problem under uncertain situation. In the classical assignment problem (AP), zpqdenotes the cost for assigning the qth job to the pth person which is deterministic in nature. Here in some uncertain situation, we have assigned a cost in the form of composite relative degree Fpq instead of  and this replaced cost is in the maximization form. In this paper, it has been solved and validated by the two proposed algorithms, a new mathematical formulation of IVIF assignment problem has been presented where the cost has been considered to be an IVIFN and the membership of elements in the set can be explained by positive and negative evidences. To determine the composite relative degree of similarity of IVIFS the concept of similarity measure and the score function is used for validating the solution which is obtained by Composite relative similarity degree method. Further, hypothetical numeric illusion is conducted to clarify the method’s effectiveness and feasibility developed in the study. Finally, conclusion and suggestion for future work are also proposed.

Keywords: Assignment problem, Interval-valued Intuitionistic Fuzzy Sets, Similarity Measures, score function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2911
7644 Observer Design for Chaos Synchronization of Time-delayed Power Systems

Authors: Jui-Sheng Lin, Yi-Sung Yang, Meei-Ling Hung, Teh-Lu Liao, Jun-Juh Yan

Abstract:

The global chaos synchronization for a class of time-delayed power systems is investigated via observer-based approach. By employing the concepts of quadratic stability theory and generalized system model, a new sufficient criterion for constructing an observer is deduced. In contrast to the previous works, this paper proposes a theoretical and systematic design procedure to realize chaos synchronization for master-slave power systems. Finally, an illustrative example is given to show the applicability of the obtained scheme.

Keywords: Chaos, Synchronization, Quadratic stability theory, Observer

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1673
7643 Using Speech Emotion Recognition as a Longitudinal Biomarker for Alzheimer’s Disease

Authors: Yishu Gong, Liangliang Yang, Jianyu Zhang, Zhengyu Chen, Sihong He, Xusheng Zhang, Wei Zhang

Abstract:

Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that affects millions of people worldwide and is characterized by cognitive decline and behavioral changes. People living with Alzheimer’s disease often find it hard to complete routine tasks. However, there are limited objective assessments that aim to quantify the difficulty of certain tasks for AD patients compared to non-AD people. In this study, we propose to use speech emotion recognition (SER), especially the frustration level as a potential biomarker for quantifying the difficulty patients experience when describing a picture. We build an SER model using data from the IEMOCAP dataset and apply the model to the DementiaBank data to detect the AD/non-AD group difference and perform longitudinal analysis to track the AD disease progression. Our results show that the frustration level detected from the SER model can possibly be used as a cost-effective tool for objective tracking of AD progression in addition to the Mini-Mental State Examination (MMSE) score.

Keywords: Alzheimer’s disease, Speech Emotion Recognition, longitudinal biomarker, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 175
7642 Improving Detection of Illegitimate Scores and Assessment in Most Advantageous Tenders

Authors: Hao-Hsi Tseng, Hsin-Yun Lee

Abstract:

Adopting Most Advantageous Tender (MAT) for the government procurement projects has become popular in Taiwan. As time pass by, the problems of MAT has appeared gradually. People condemn two points that are the result might be manipulated by a single committee member’s partiality and how to make a fair decision when the winner has two or more. Arrow’s Impossibility Theorem proposed that the best scoring method should meet the four reasonable criteria. According to these four criteria this paper constructed an “Illegitimate Scores Checking Scheme” for a scoring method and used the scheme to find out the illegitimate of the current evaluation method of MAT. This paper also proposed a new scoring method that is called the “Standardizing Overall Evaluated Score Method”. This method makes each committee member’s influence tend to be identical. Thus, the committee members can scoring freely according to their partiality without losing the fairness. Finally, it was examined by a large-scale simulation, and the experiment revealed that the it improved the problem of dictatorship and perfectly avoided the situation of cyclical majorities, simultaneously. This result verified that the Standardizing Overall Evaluated Score Method is better than any current evaluation method of MAT.

Keywords: Arrow’s impossibility theorem, most advantageous tender, illegitimate scores checking scheme, standard score.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1434
7641 Sentiment Analysis of Fake Health News Using Naive Bayes Classification Models

Authors: Danielle Shackley, Yetunde Folajimi

Abstract:

As more people turn to the internet seeking health related information, there is more risk of finding false, inaccurate, or dangerous information. Sentiment analysis is a natural language processing technique that assigns polarity scores of text, ranging from positive, neutral and negative. In this research, we evaluate the weight of a sentiment analysis feature added to fake health news classification models. The dataset consists of existing reliably labeled health article headlines that were supplemented with health information collected about COVID-19 from social media sources. We started with data preprocessing, tested out various vectorization methods such as Count and TFIDF vectorization. We implemented 3 Naive Bayes classifier models, including Bernoulli, Multinomial and Complement. To test the weight of the sentiment analysis feature on the dataset, we created benchmark Naive Bayes classification models without sentiment analysis, and those same models were reproduced and the feature was added. We evaluated using the precision and accuracy scores. The Bernoulli initial model performed with 90% precision and 75.2% accuracy, while the model supplemented with sentiment labels performed with 90.4% precision and stayed constant at 75.2% accuracy. Our results show that the addition of sentiment analysis did not improve model precision by a wide margin; while there was no evidence of improvement in accuracy, we had a 1.9% improvement margin of the precision score with the Complement model. Future expansion of this work could include replicating the experiment process, and substituting the Naive Bayes for a deep learning neural network model.

Keywords: Sentiment analysis, Naive Bayes model, natural language processing, topic analysis, fake health news classification model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 381
7640 A Generalized Sparse Bayesian Learning Algorithm for Near-Field Synthetic Aperture Radar Imaging: By Exploiting Impropriety and Noncircularity

Authors: Pan Long, Bi Dongjie, Li Xifeng, Xie Yongle

Abstract:

The near-field synthetic aperture radar (SAR) imaging is an advanced nondestructive testing and evaluation (NDT&E) technique. This paper investigates the complex-valued signal processing related to the near-field SAR imaging system, where the measurement data turns out to be noncircular and improper, meaning that the complex-valued data is correlated to its complex conjugate. Furthermore, we discover that the degree of impropriety of the measurement data and that of the target image can be highly correlated in near-field SAR imaging. Based on these observations, A modified generalized sparse Bayesian learning algorithm is proposed, taking impropriety and noncircularity into account. Numerical results show that the proposed algorithm provides performance gain, with the help of noncircular assumption on the signals.

Keywords: Complex-valued signal processing, synthetic aperture radar (SAR), 2-D radar imaging, compressive sensing, Sparse Bayesian learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1438
7639 Sensory Characterization of Cookies with Chestnut Flour

Authors: Ljubica Dokić, Ivana Nikolić, Dragana Šoronja–Simović, Biljana Pajin, Nils Juul

Abstract:

In this work sensory characteristics of cookies with different amount of chestnut flour were determined by sensory and instrumental methods. The wheat flour for cookies was substituted with chestnut flour in three different levels (20, 40 and 60%) and the dough moisture was 22%. The control sample was with 100% of wheat flour. Sensory quality of the cookies was described using quantity descriptive method (QDA) by six trained members of descriptive panel. Instrumental evaluation included texture characterization by texture analyzer, the color measurements (CIE L*a*b* system) and determination by videometer.

The samples with 20% of chestnut flour were with highest ponderated score for overall sensory impression (17.6), which is very close to score for control sample (18). Increase in amount of chestnut flour caused decrease in scores for all sensory properties, thus overall sensory score decreased also. Compared to control sample and with increase in amount of chestnut flour, instrumental determination of the samples confirmed the sensory analysis results. The hardness of the cookies increased, as well as the values of red a* and yellow (b*) component coordinate, but the values for lightness (L*) decreased. Also the values, evaluated by videometer at defined wavelength, were the highest for control cookies and decreased with increase in amount of chestnut flour.

Keywords: Cookies, chestnut flour, sensory characteristics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2776
7638 Optimization of Reaction Rate Parameters in Modeling of Heavy Paraffins Dehydrogenation

Authors: Leila Vafajoo, Farhad Khorasheh, Mehrnoosh Hamzezadeh Nakhjavani, Moslem Fattahi

Abstract:

In the present study, a procedure was developed to determine the optimum reaction rate constants in generalized Arrhenius form and optimized through the Nelder-Mead method. For this purpose, a comprehensive mathematical model of a fixed bed reactor for dehydrogenation of heavy paraffins over Pt–Sn/Al2O3 catalyst was developed. Utilizing appropriate kinetic rate expressions for the main dehydrogenation reaction as well as side reactions and catalyst deactivation, a detailed model for the radial flow reactor was obtained. The reactor model composed of a set of partial differential equations (PDE), ordinary differential equations (ODE) as well as algebraic equations all of which were solved numerically to determine variations in components- concentrations in term of mole percents as a function of time and reactor radius. It was demonstrated that most significant variations observed at the entrance of the bed and the initial olefin production obtained was rather high. The aforementioned method utilized a direct-search optimization algorithm along with the numerical solution of the governing differential equations. The usefulness and validity of the method was demonstrated by comparing the predicted values of the kinetic constants using the proposed method with a series of experimental values reported in the literature for different systems.

Keywords: Dehydrogenation, Pt-Sn/Al2O3 Catalyst, Modeling, Nelder-Mead, Optimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2695
7637 Effect of Different Lactic Acid Bacteria on Phytic Acid Content and Quality of whole Wheat Toast Bread

Authors: Z. Didar, A. Pourfarzad, M. H. Haddad Khodaparast

Abstract:

Nowadays, consumption of whole flours and flours with high extraction rate is recommended, because of their high amount of fibers, vitamins and minerals. Despite nutritional benefits of whole flour, concentration of some undesirable components such as phytic acid is higher than white flour. In this study, effect of several lactic acid bacteria sourdough on Toast bread is investigated. Sourdough from lactic acid bacteria (Lb. plantarum, Lb. reuteri) with different dough yield (250 and 300) is made and incubated at 30°C for 20 hour, then added to dough in the ratio of 10, 20 and 30% replacement. Breads that supplemented with Lb. plantarum sourdough had lower phytic acid. Higher replacement of sourdough and higher DY cause higher decrease in phytic acid content. Sourdough from Lb. plantarum, DY = 300 and 30% replacement cause the highest decrease in phytic acid content (49.63 mg/100g). As indicated by panelists, Lb. reuteri sourdough can present the greatest effect on overall quality score of the breads. DY reduction cause a decrease in bread quality score. Sensory score of Toast bread is 81.71 in the samples that treated with Lb. reuteri sourdough with DY = 250 and 20% replacement.

Keywords: Phytic Acid, Sourdough, Toast Bread, Whole Wheat Flour, Lactic Acid Bacteria.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2901
7636 Lineup Optimization Model of Basketball Players Based on the Prediction of Recursive Neural Networks

Authors: Wang Yichen, Haruka Yamashita

Abstract:

In recent years, in the field of sports, decision making such as member in the game and strategy of the game based on then analysis of the accumulated sports data are widely attempted. In fact, in the NBA basketball league where the world's highest level players gather, to win the games, teams analyze the data using various statistical techniques. However, it is difficult to analyze the game data for each play such as the ball tracking or motion of the players in the game, because the situation of the game changes rapidly, and the structure of the data should be complicated. Therefore, it is considered that the analysis method for real time game play data is proposed. In this research, we propose an analytical model for "determining the optimal lineup composition" using the real time play data, which is considered to be difficult for all coaches. In this study, because replacing the entire lineup is too complicated, and the actual question for the replacement of players is "whether or not the lineup should be changed", and “whether or not Small Ball lineup is adopted”. Therefore, we propose an analytical model for the optimal player selection problem based on Small Ball lineups. In basketball, we can accumulate scoring data for each play, which indicates a player's contribution to the game, and the scoring data can be considered as a time series data. In order to compare the importance of players in different situations and lineups, we combine RNN (Recurrent Neural Network) model, which can analyze time series data, and NN (Neural Network) model, which can analyze the situation on the field, to build the prediction model of score. This model is capable to identify the current optimal lineup for different situations. In this research, we collected all the data of accumulated data of NBA from 2019-2020. Then we apply the method to the actual basketball play data to verify the reliability of the proposed model.

Keywords: Recurrent Neural Network, players lineup, basketball data, decision making model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 747
7635 Experimental Results about the Dynamics of the Generalized Belief Propagation Used on LDPC Codes

Authors: Jean-Christophe Sibel, Sylvain Reynal, David Declercq

Abstract:

In the context of channel coding, the Generalized Belief Propagation (GBP) is an iterative algorithm used to recover the transmission bits sent through a noisy channel. To ensure a reliable transmission, we apply a map on the bits, that is called a code. This code induces artificial correlations between the bits to send, and it can be modeled by a graph whose nodes are the bits and the edges are the correlations. This graph, called Tanner graph, is used for most of the decoding algorithms like Belief Propagation or Gallager-B. The GBP is based on a non unic transformation of the Tanner graph into a so called region-graph. A clear advantage of the GBP over the other algorithms is the freedom in the construction of this graph. In this article, we explain a particular construction for specific graph topologies that involves relevant performance of the GBP. Moreover, we investigate the behavior of the GBP considered as a dynamic system in order to understand the way it evolves in terms of the time and in terms of the noise power of the channel. To this end we make use of classical measures and we introduce a new measure called the hyperspheres method that enables to know the size of the attractors.

Keywords: iterative decoder, LDPC, region-graph, chaos.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1597
7634 Categorization and Estimation of Relative Connectivity of Genes from Meta-OFTEN Network

Authors: U. Kairov, T. Karpenyuk, E. Ramanculov, A. Zinovyev

Abstract:

The most common result of analysis of highthroughput data in molecular biology represents a global list of genes, ranked accordingly to a certain score. The score can be a measure of differential expression. Recent work proposed a new method for selecting a number of genes in a ranked gene list from microarray gene expression data such that this set forms the Optimally Functionally Enriched Network (OFTEN), formed by known physical interactions between genes or their products. Here we present calculation results of relative connectivity of genes from META-OFTEN network and tentative biological interpretation of the most reproducible signal. The relative connectivity and inbetweenness values of genes from META-OFTEN network were estimated.

Keywords: Microarray, META-OFTEN, gene network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1579