Search results for: Vector Error Correction Model (VECM)
16249 Tackling the Value-Action-Gap: Improving Civic Participation Using a Holistic Behavioral Model Approach
Authors: Long Pham, Julia Blanke
Abstract:
An increasingly popular way of establishing citizen engagement within communities is through ‘city apps’. Currently, most of these mobile applications seem to be extensions of the existing communication media, sometimes merely replicating the information available on the classical city web sites, and therefore provide minimal additional impact on citizen behavior and engagement. In order to overcome this challenge, we propose to use a holistic behavioral model to generate dynamic and contextualized app content based on optimizing well defined city-related performance goals constrained by the proposed behavioral model. In this paper, we will show how the data collected by the CorkCitiEngage project in the Irish city of Cork can be utilized to calibrate aspects of the proposed model enabling the design of a personalized citizen engagement app aiming at positively influencing people’s behavior towards more active participation in their communities. We will focus on the important aspect of intentions to act, which is essential for understanding the reasons behind the common value-action-gap being responsible for the mismatch between good intentions and actual observable behavior, and will discuss how customized app design can be based on a rigorous model of behavior optimized towards maximizing well defined city-related performance goals.Keywords: city apps, holistic behaviour model, intention to act, value-action-gap, citizen engagement
Procedia PDF Downloads 22916248 A Fuzzy Mathematical Model for Order Acceptance and Scheduling Problem
Authors: E. Koyuncu
Abstract:
The problem of Order Acceptance and Scheduling (OAS) is defined as a joint decision of which orders to accept for processing and how to schedule them. Any linear programming model representing real-world situation involves the parameters defined by the decision maker in an uncertain way or by means of language statement. Fuzzy data can be used to incorporate vagueness in the real-life situation. In this study, a fuzzy mathematical model is proposed for a single machine OAS problem, where the orders are defined by their fuzzy due dates, fuzzy processing times, and fuzzy sequence dependent setup times. The signed distance method, one of the fuzzy ranking methods, is used to handle the fuzzy constraints in the model.Keywords: fuzzy mathematical programming, fuzzy ranking, order acceptance, single machine scheduling
Procedia PDF Downloads 34116247 Optimization Process for Ride Quality of a Nonlinear Suspension Model Based on Newton-Euler’ Augmented Formulation
Authors: Mohamed Belhorma, Aboubakar S. Bouchikhi, Belkacem Bounab
Abstract:
This paper addresses modeling a Double A-Arm suspension, a three-dimensional nonlinear model has been developed using the multibody systems formalism. Dynamical study of the different components responses was done, particularly for the wheel assembly. To validate those results, the system was constructed and simulated by RecurDyn, a professional multibody dynamics simulation software. The model has been used as the Objectif function in an optimization algorithm for ride quality improvement.Keywords: double A-Arm suspension, multibody systems, ride quality optimization, dynamic simulation
Procedia PDF Downloads 13916246 Embedded System of Signal Processing on FPGA: Underwater Application Architecture
Authors: Abdelkader Elhanaoui, Mhamed Hadji, Rachid Skouri, Said Agounad
Abstract:
The purpose of this paper is to study the phenomenon of acoustic scattering by using a new method. The signal processing (Fast Fourier Transform FFT Inverse Fast Fourier Transform iFFT and BESSEL functions) is widely applied to obtain information with high precision accuracy. Signal processing has a wider implementation in general-purpose pro-cessors. Our interest was focused on the use of FPGAs (Field-Programmable Gate Ar-rays) in order to minimize the computational complexity in single processor architecture, then be accelerated on FPGA and meet real-time and energy efficiency requirements. Gen-eral-purpose processors are not efficient for signal processing. We implemented the acous-tic backscattered signal processing model on the Altera DE-SOC board and compared it to Odroid xu4. By comparison, the computing latency of Odroid xu4 and FPGA is 60 sec-onds and 3 seconds, respectively. The detailed SoC FPGA-based system has shown that acoustic spectra are performed up to 20 times faster than the Odroid xu4 implementation. FPGA-based system of processing algorithms is realized with an absolute error of about 10⁻³. This study underlines the increasing importance of embedded systems in underwater acoustics, especially in non-destructive testing. It is possible to obtain information related to the detection and characterization of submerged cells. So we have achieved good exper-imental results in real-time and energy efficiency.Keywords: DE1 FPGA, acoustic scattering, form function, signal processing, non-destructive testing
Procedia PDF Downloads 8016245 A Comprehensive Approach to Scour Depth Estimation Through HEC-RAS 2D and Physical Modeling
Authors: Ashvinie Thembiliyagoda, Kasun De Silva, Nimal Wijayaratna
Abstract:
The lowering of the riverbed level as a result of water erosion is termed as scouring. This phenomenon remarkably undermines the potential stability of the bridge pier, causing a threat of failure or collapse. The formation of vortices in the vicinity of bridges due to the obstruction caused by river flow is the main reason behind this pursuit. Scouring is aggravated by factors including high flow rates, bridge pier geometry, sediment configuration etc. Tackling scour-related problems when they become severe is more costly and disruptive compared to implementing preventive measures based on predicted scour depths. This paper presents a comprehensive investigation of the development of a numerical model that could reproduce the scouring effect around bridge piers and estimate the scour depth. The numerical model was developed for one selected bridge in Sri Lanka, the Kelanisiri Bridge. HEC-RAS two-dimensional (2D) modeling approach was utilized for the development of the model and was calibrated and validated with field data. To further enhance the reliability of the model, a physical model was developed, allowing for additional validation. Results from the numerical model were compared with those obtained from the physical model, revealing a strong correlation between the two methods and confirming the numerical model's accuracy in predicting scour depths. The findings from this study underscore the ability of the HEC-RAS two-dimensional modeling approach for the estimation of scour depth around bridge piers. The developed model is able to estimate the scour depth under varying flow conditions, and its flexibility allows it to be adapted for application to other bridges with similar hydraulic and geomorphological conditions, providing a robust tool for widespread use in scour estimation. The developed two-dimensional model not only offers reliable predictions for the case study bridge but also holds significant potential for broader implementation, contributing to the improved design and maintenance of bridge structures in diverse environments.Keywords: piers, scouring, HEC-RAS, physical model
Procedia PDF Downloads 1816244 Study and Analysis of the Factors Affecting Road Safety Using Decision Tree Algorithms
Authors: Naina Mahajan, Bikram Pal Kaur
Abstract:
The purpose of traffic accident analysis is to find the possible causes of an accident. Road accidents cannot be totally prevented but by suitable traffic engineering and management the accident rate can be reduced to a certain extent. This paper discusses the classification techniques C4.5 and ID3 using the WEKA Data mining tool. These techniques use on the NH (National highway) dataset. With the C4.5 and ID3 technique it gives best results and high accuracy with less computation time and error rate.Keywords: C4.5, ID3, NH(National highway), WEKA data mining tool
Procedia PDF Downloads 34016243 Transmit Power Optimization for Cooperative Beamforming in Reverse-Link MIMO Ad-Hoc Networks
Authors: Younghyun Jeon, Seungjoo Maeng
Abstract:
In the Ad-hoc network, the great interests regarding MIMO scheme leads to their combination, which is also utilized into its applicable network. We manage the field of the problem into Reverse-link MIMO Ad-hoc Network (RMAN) and propose the methodology to maximize the data rate with its power consumption using Node-Cooperative beamforming technique. Based on the result of mathematical optimization formulation, we design the algorithm to construct optimal orthogonal weight vector according to channel feedback and control its transmission power according to QoS-pricing value level. In simulation results, we show the validity of the proposed mathematical optimization result and algorithm which mean that the sum-rate of each link is converged into some point.Keywords: ad-hoc network, MIMO, cooperative beamforming, transmit power
Procedia PDF Downloads 39916242 An Advanced Exponential Model for Seismic Isolators Having Hardening or Softening Behavior at Large Displacements
Authors: Nicolò Vaiana, Giorgio Serino
Abstract:
In this paper, an advanced Nonlinear Exponential Model (NEM), able to simulate the uniaxial dynamic behavior of seismic isolators having a continuously decreasing tangent stiffness with increasing displacement in the relatively large displacements range and a hardening or softening behavior at large displacements, is presented. The mathematical model is validated by comparing the experimental force-displacement hysteresis loops obtained during cyclic tests, conducted on a helical wire rope isolator and a recycled rubber-fiber reinforced bearing, with those predicted analytically. Good agreement between the experimental and simulated results shows that the proposed model can be an effective numerical tool to predict the force-displacement relationship of seismic isolation devices within the large displacements range. Compared to the widely used Bouc-Wen model, unable to simulate the response of seismic isolators at large displacements, the proposed one allows to avoid the numerical solution of a first order nonlinear ordinary differential equation for each time step of a nonlinear time history analysis, thus reducing the computation effort. Furthermore, the proposed model can simulate the smooth transition of the hysteresis loops from small to large displacements by adopting only one set of five parameters determined from the experimental hysteresis loops having the largest amplitude.Keywords: base isolation, hardening behavior, nonlinear exponential model, seismic isolators, softening behavior
Procedia PDF Downloads 33116241 Quality Assessment of New Zealand Mānuka Honeys Using Hyperspectral Imaging Combined with Deep 1D-Convolutional Neural Networks
Authors: Hien Thi Dieu Truong, Mahmoud Al-Sarayreh, Pullanagari Reddy, Marlon M. Reis, Richard Archer
Abstract:
New Zealand mānuka honey is a honeybee product derived mainly from Leptospermum scoparium nectar. The potent antibacterial activity of mānuka honey derives principally from methylglyoxal (MGO), in addition to the hydrogen peroxide and other lesser activities present in all honey. MGO is formed from dihydroxyacetone (DHA) unique to L. scoparium nectar. Mānuka honey also has an idiosyncratic phenolic profile that is useful as a chemical maker. Authentic mānuka honey is highly valuable, but almost all honey is formed from natural mixtures of nectars harvested by a hive over a time period. Once diluted by other nectars, mānuka honey irrevocably loses value. We aimed to apply hyperspectral imaging to honey frames before bulk extraction to minimise the dilution of genuine mānuka by other honey and ensure authenticity at the source. This technology is non-destructive and suitable for an industrial setting. Chemometrics using linear Partial Least Squares (PLS) and Support Vector Machine (SVM) showed limited efficacy in interpreting chemical footprints due to large non-linear relationships between predictor and predictand in a large sample set, likely due to honey quality variability across geographic regions. Therefore, an advanced modelling approach, one-dimensional convolutional neural networks (1D-CNN), was investigated for analysing hyperspectral data for extraction of biochemical information from honey. The 1D-CNN model showed superior prediction of honey quality (R² = 0.73, RMSE = 2.346, RPD= 2.56) to PLS (R² = 0.66, RMSE = 2.607, RPD= 1.91) and SVM (R² = 0.67, RMSE = 2.559, RPD=1.98). Classification of mono-floral manuka honey from multi-floral and non-manuka honey exceeded 90% accuracy for all models tried. Overall, this study reveals the potential of HSI and deep learning modelling for automating the evaluation of honey quality in frames.Keywords: mānuka honey, quality, purity, potency, deep learning, 1D-CNN, chemometrics
Procedia PDF Downloads 14116240 Prevalence of Near Visual Impairment and Associated Factors among School Teachers in Gondar City, North West Ethiopia, 2022
Authors: Bersufekad Wubie
Abstract:
Introduction: Near visual impairment is presenting near visual acuity of the eye worse than N6 at a 40 cm distance. Teachers' regular duties, such as reading books, writing on the blackboard, and recognizing students' faces, need good near vision. If a teacher has near-visual impairment, the work output is unsatisfactory. Objective: The study was aimed to assess the prevalence and associated factors near vision impairment among school teachers at Gondar city Northwest Ethiopia, August 2022. Methods: To select 567 teachers in Gondar city schools, an institutional-based cross-sectional study design with a multistage sampling technique were used. The study was conducted in selected schools from May 1 to May 30, 2022. Trained data collectors used well-structured Amharic and English language questionnaires and ophthalmic instruments for examination. The collected data were checked for completeness and entered into Epi data version 4.6, then exported to SPSS version 26 for further analysis. A binary and multivariate logistic regression model was fitted. And associated factors of the outcome variable. Result: The prevalence of near visual impairment was 64.6%, with a confidence interval of 60.3%–68.4%. Near visual impairment was significantly associated with age >= 35 years (AOR: 4.90 at 95% CI: 3.15, 7.65), having prolonged years of teaching experience (AOR: 3.29 at 95% CI: 1.70, 4.62), having a history of ocular surgery (AOR: 1.96 at 95% CI: 1.10, 4.62), smokers (AOR: 2.21 at 95% CI: 1.22, 4.07), history of ocular trauma (AOR : 1.80 at 95%CI:1.11,3.18 and uncorrected refractive error (AOR:2.01 at 95%CI:1.13,4.03). Conclusion and recommendations: This study showed the prevalence of near vision impairment among school teachers was high, and it is not a problem of the presbyopia age group alone; it also happens at a young age. So teachers' ocular health should be well accommodated in the school's eye health.Keywords: Gondar, near visual impairment, school, teachers
Procedia PDF Downloads 14016239 A Scalable Model of Fair Socioeconomic Relations Based on Blockchain and Machine Learning Algorithms-1: On Hyperinteraction and Intuition
Authors: Merey M. Sarsengeldin, Alexandr S. Kolokhmatov, Galiya Seidaliyeva, Alexandr Ozerov, Sanim T. Imatayeva
Abstract:
This series of interdisciplinary studies is an attempt to investigate and develop a scalable model of fair socioeconomic relations on the base of blockchain using positive psychology techniques and Machine Learning algorithms for data analytics. In this particular study, we use hyperinteraction approach and intuition to investigate their influence on 'wisdom of crowds' via created mobile application which was created for the purpose of this research. Along with the public blockchain and private Decentralized Autonomous Organization (DAO) which were elaborated by us on the base of Ethereum blockchain, a model of fair financial relations of members of DAO was developed. We developed a smart contract, so-called, Fair Price Protocol and use it for implementation of model. The data obtained from mobile application was analyzed by ML algorithms. A model was tested on football matches.Keywords: blockchain, Naïve Bayes algorithm, hyperinteraction, intuition, wisdom of crowd, decentralized autonomous organization
Procedia PDF Downloads 17116238 Reduced Model Investigations Supported by Fuzzy Cognitive Map to Foster Circular Economy
Authors: A. Buruzs, M. F. Hatwágner, L. T. Kóczy
Abstract:
The aim of the present paper is to develop an integrated method that may provide assistance to decision makers during system planning, design, operation and evaluation. In order to support the realization of Circular Economy (CE), it is essential to evaluate local needs and conditions which help to select the most appropriate system components and resource needs. Each of these activities requires careful planning, however, the model of CE offers a comprehensive interdisciplinary framework. The aim of this research was to develop and to introduce a practical methodology for evaluation of local and regional opportunities to promote CE.Keywords: circular economy, factors, fuzzy cognitive map, model reduction, sustainability
Procedia PDF Downloads 24616237 The Use of Geographically Weighted Regression for Deforestation Analysis: Case Study in Brazilian Cerrado
Authors: Ana Paula Camelo, Keila Sanches
Abstract:
The Geographically Weighted Regression (GWR) was proposed in geography literature to allow relationship in a regression model to vary over space. In Brazil, the agricultural exploitation of the Cerrado Biome is the main cause of deforestation. In this study, we propose a methodology using geostatistical methods to characterize the spatial dependence of deforestation in the Cerrado based on agricultural production indicators. Therefore, it was used the set of exploratory spatial data analysis tools (ESDA) and confirmatory analysis using GWR. It was made the calibration a non-spatial model, evaluation the nature of the regression curve, election of the variables by stepwise process and multicollinearity analysis. After the evaluation of the non-spatial model was processed the spatial-regression model, statistic evaluation of the intercept and verification of its effect on calibration. In an analysis of Spearman’s correlation the results between deforestation and livestock was +0.783 and with soybeans +0.405. The model presented R²=0.936 and showed a strong spatial dependence of agricultural activity of soybeans associated to maize and cotton crops. The GWR is a very effective tool presenting results closer to the reality of deforestation in the Cerrado when compared with other analysis.Keywords: deforestation, geographically weighted regression, land use, spatial analysis
Procedia PDF Downloads 36516236 Dislocation Density-Based Modeling of the Grain Refinement in Surface Mechanical Attrition Treatment
Authors: Reza Miresmaeili, Asghar Heydari Astaraee, Fereshteh Dolati
Abstract:
In the present study, an analytical model based on dislocation density model was developed to simulate grain refinement in surface mechanical attrition treatment (SMAT). The correlation between SMAT time and development in plastic strain on one hand, and dislocation density evolution, on the other hand, was established to simulate the grain refinement in SMAT. A dislocation density-based constitutive material law was implemented using VUHARD subroutine. A random sequence of shots is taken into consideration for multiple impacts model using Python programming language by utilizing a random function. The simulation technique was to model each impact in a separate run and then transferring the results of each run as initial conditions for the next run (impact). The developed Finite Element (FE) model of multiple impacts describes the coverage evolution in SMAT. Simulations were run to coverage levels as high as 4500%. It is shown that the coverage implemented in the FE model is equal to the experimental coverage. It is depicted that numerical SMAT coverage parameter is adequately conforming to the well-known Avrami model. Comparison between numerical results and experimental measurements for residual stresses and depth of deformation layers confirms the performance of the established FE model for surface engineering evaluations in SMA treatment. X-ray diffraction (XRD) studies of grain refinement, including resultant grain size and dislocation density, were conducted to validate the established model. The full width at half-maximum in XRD profiles can be used to measure the grain size. Numerical results and experimental measurements of grain refinement illustrate good agreement and show the capability of established FE model to predict the gradient microstructure in SMA treatment.Keywords: dislocation density, grain refinement, severe plastic deformation, simulation, surface mechanical attrition treatment
Procedia PDF Downloads 13816235 Observation of the Orthodontic Tooth's Long-Term Movement Using Stereovision System
Authors: Hao-Yuan Tseng, Chuan-Yang Chang, Ying-Hui Chen, Sheng-Che Chen, Chih-Han Chang
Abstract:
Orthodontic tooth treatment has demonstrated a high success rate in clinical studies. It has been agreed upon that orthodontic tooth movement is based on the ability of surrounding bone and periodontal ligament (PDL) to react to a mechanical stimulus with remodeling processes. However, the mechanism of the tooth movement is still unclear. Recent studies focus on the simple principle compression-tension theory while rare studies directly measure tooth movement. Therefore, tracking tooth movement information during orthodontic treatment is very important in clinical practice. The aim of this study is to investigate the mechanism responses of the tooth movement during the orthodontic treatments. A stereovision system applied to track the tooth movement of the patient with the stamp brackets. The system was established by two cameras with their relative position calibrate. And the orthodontic force measured by 3D printing model with the six-axis load cell to determine the initial force application. The result shows that the stereovision system accuracy revealed the measurement presents a maximum error less than 2%. For the study on patient tracking, the incisor moved about 0.9 mm during 60 days tracking, and half of movement occurred in the first few hours. After removing the orthodontic force in 100 hours, the distance between before and after position incisor tooth decrease 0.5 mm consisted with the release of the phenomenon. Using the stereovision system can accurately locate the three-dimensional position of the teeth and superposition of 3D coordinate system for all the data to integrate the complex tooth movement.Keywords: orthodontic treatment, tooth movement, stereovision system, long-term tracking
Procedia PDF Downloads 42416234 The Grand Unified Theory of Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow Model
Authors: Tory Erickson
Abstract:
The "Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow" (BST-SCWPDF) Model introduces a framework aimed at unifying general relativity (GR) and quantum mechanics (QM). By proposing a concept of bidirectional spacetime, this model suggests that time can flow in more than one direction, thus offering a perspective on temporal dynamics. Integrated with spatial covariance and wave-particle duality in spacetime flow, the BST-SCWPDF Model resolves long-standing discrepancies between GR and QM. This unified theory has profound implications for quantum gravity, potentially offering insights into quantum entanglement, the collapse of the wave function, and the fabric of spacetime itself. The Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow" (BST-SCWPDF) Model offers researchers a framework for a better understanding of theoretical physics.Keywords: astrophysics, quantum mechanics, general relativity, unification theory, theoretical physics
Procedia PDF Downloads 9016233 Assessment of Students Skills in Error Detection in SQL Classes using Rubric Framework - An Empirical Study
Authors: Dirson Santos De Campos, Deller James Ferreira, Anderson Cavalcante Gonçalves, Uyara Ferreira Silva
Abstract:
Rubrics to learning research provide many evaluation criteria and expected performance standards linked to defined student activity for learning and pedagogical objectives. Despite the rubric being used in education at all levels, academic literature on rubrics as a tool to support research in SQL Education is quite rare. There is a large class of SQL queries is syntactically correct, but certainly, not all are semantically correct. Detecting and correcting errors is a recurring problem in SQL education. In this paper, we usthe Rubric Abstract Framework (RAF), which consists of steps, that allows us to map the information to measure student performance guided by didactic objectives defined by the teacher as long as it is contextualized domain modeling by rubric. An empirical study was done that demonstrates how rubrics can mitigate student difficulties in finding logical errors and easing teacher workload in SQL education. Detecting and correcting logical errors is an important skill for students. Researchers have proposed several ways to improve SQL education because understanding this paradigm skills are crucial in software engineering and computer science. The RAF instantiation was using in an empirical study developed during the COVID-19 pandemic in database course. The pandemic transformed face-to-face and remote education, without presential classes. The lab activities were conducted remotely, which hinders the teaching-learning process, in particular for this research, in verifying the evidence or statements of knowledge, skills, and abilities (KSAs) of students. Various research in academia and industry involved databases. The innovation proposed in this paper is the approach used where the results obtained when using rubrics to map logical errors in query formulation have been analyzed with gains obtained by students empirically verified. The research approach can be used in the post-pandemic period in both classroom and distance learning.Keywords: rubric, logical error, structured query language (SQL), empirical study, SQL education
Procedia PDF Downloads 19216232 Generalized Additive Model for Estimating Propensity Score
Authors: Tahmidul Islam
Abstract:
Propensity Score Matching (PSM) technique has been widely used for estimating causal effect of treatment in observational studies. One major step of implementing PSM is estimating the propensity score (PS). Logistic regression model with additive linear terms of covariates is most used technique in many studies. Logistics regression model is also used with cubic splines for retaining flexibility in the model. However, choosing the functional form of the logistic regression model has been a question since the effectiveness of PSM depends on how accurately the PS been estimated. In many situations, the linearity assumption of linear logistic regression may not hold and non-linear relation between the logit and the covariates may be appropriate. One can estimate PS using machine learning techniques such as random forest, neural network etc for more accuracy in non-linear situation. In this study, an attempt has been made to compare the efficacy of Generalized Additive Model (GAM) in various linear and non-linear settings and compare its performance with usual logistic regression. GAM is a non-parametric technique where functional form of the covariates can be unspecified and a flexible regression model can be fitted. In this study various simple and complex models have been considered for treatment under several situations (small/large sample, low/high number of treatment units) and examined which method leads to more covariate balance in the matched dataset. It is found that logistic regression model is impressively robust against inclusion quadratic and interaction terms and reduces mean difference in treatment and control set equally efficiently as GAM does. GAM provided no significantly better covariate balance than logistic regression in both simple and complex models. The analysis also suggests that larger proportion of controls than treatment units leads to better balance for both of the methods.Keywords: accuracy, covariate balances, generalized additive model, logistic regression, non-linearity, propensity score matching
Procedia PDF Downloads 36916231 Systematic and Simple Guidance for Feed Forward Design in Model Predictive Control
Authors: Shukri Dughman, Anthony Rossiter
Abstract:
This paper builds on earlier work which demonstrated that Model Predictive Control (MPC) may give a poor choice of default feed forward compensator. By first demonstrating the impact of future information of target changes on the performance, this paper proposes a pragmatic method for identifying the amount of future information on the target that can be utilised effectively in both finite and infinite horizon algorithms. Numerical illustrations in MATLAB give evidence of the efficacy of the proposal.Keywords: model predictive control, tracking control, advance knowledge, feed forward
Procedia PDF Downloads 54916230 Research on Detection of Web Page Visual Salience Region Based on Eye Tracker and Spectral Residual Model
Authors: Xiaoying Guo, Xiangyun Wang, Chunhua Jia
Abstract:
Web page has been one of the most important way of knowing the world. Humans catch a lot of information from it everyday. Thus, understanding where human looks when they surfing the web pages is rather important. In normal scenes, the down-top features and top-down tasks significantly affect humans’ eye movement. In this paper, we investigated if the conventional visual salience algorithm can properly predict humans’ visual attractive region when they viewing the web pages. First, we obtained the eye movement data when the participants viewing the web pages using an eye tracker. By the analysis of eye movement data, we studied the influence of visual saliency and thinking way on eye-movement pattern. The analysis result showed that thinking way affect human’ eye-movement pattern much more than visual saliency. Second, we compared the results of web page visual salience region extracted by Itti model and Spectral Residual (SR) model. The results showed that Spectral Residual (SR) model performs superior than Itti model by comparison with the heat map from eye movements. Considering the influence of mind habit on humans’ visual region of interest, we introduced one of the most important cue in mind habit-fixation position to improved the SR model. The result showed that the improved SR model can better predict the human visual region of interest in web pages.Keywords: web page salience region, eye-tracker, spectral residual, visual salience
Procedia PDF Downloads 27716229 A Dynamical Study of Fractional Order Obesity Model by a Combined Legendre Wavelet Method
Authors: Hakiki Kheira, Belhamiti Omar
Abstract:
In this paper, we propose a new compartmental fractional order model for the simulation of epidemic obesity dynamics. Using the Legendre wavelet method combined with the decoupling and quasi-linearization technique, we demonstrate the validity and applicability of our model. We also present some fractional differential illustrative examples to demonstrate the applicability and efficiency of the method. The fractional derivative is described in the Caputo sense.Keywords: Caputo derivative, epidemiology, Legendre wavelet method, obesity
Procedia PDF Downloads 42216228 An Optimal Approach for Full-Detailed Friction Model Identification of Reaction Wheel
Authors: Ghasem Sharifi, Hamed Shahmohamadi Ousaloo, Milad Azimi, Mehran Mirshams
Abstract:
The ever-increasing use of satellites demands a search for increasingly accurate and reliable pointing systems. Reaction wheels are rotating devices used commonly for the attitude control of the spacecraft since provide a wide range of torque magnitude and high reliability. The numerical modeling of this device can significantly enhance the accuracy of the satellite control in space. Modeling the wheel rotation in the presence of the various frictions is one of the critical parts of this approach. This paper presents a Dynamic Model Control of a Reaction Wheel (DMCR) in the current control mode. In current-mode, the required current is delivered to the coils in order to achieve the desired torque. During this research, all the friction parameters as viscous and coulomb, motor coefficient, resistance and voltage constant are identified. In order to model identification of a reaction wheel, numerous varying current commands apply on the particular wheel to verify the estimated model. All the parameters of DMCR are identified by classical Levenberg-Marquardt (CLM) optimization method. The experimental results demonstrate that the developed model has an appropriate precise and can be used in the satellite control simulation.Keywords: experimental modeling, friction parameters, model identification, reaction wheel
Procedia PDF Downloads 23516227 Performance Study of ZigBee-Based Wireless Sensor Networks
Authors: Afif Saleh Abugharsa
Abstract:
The IEEE 802.15.4 standard is designed for low-rate wireless personal area networks (LR-WPAN) with focus on enabling wireless sensor networks. It aims to give a low data rate, low power consumption, and low cost wireless networking on the device-level communication. The objective of this study is to investigate the performance of IEEE 802.15.4 based networks using simulation tool. In this project the network simulator 2 NS2 was used to several performance measures of wireless sensor networks. Three scenarios were considered, multi hop network with a single coordinator, star topology, and an ad hoc on demand distance vector AODV. Results such as packet delivery ratio, hop delay, and number of collisions are obtained from these scenarios.Keywords: ZigBee, wireless sensor networks, IEEE 802.15.4, low power, low data rate
Procedia PDF Downloads 43516226 Numerical Simulation and Experimental Validation of the Tire-Road Separation in Quarter-car Model
Authors: Quy Dang Nguyen, Reza Nakhaie Jazar
Abstract:
The paper investigates vibration dynamics of tire-road separation for a quarter-car model; this separation model is developed to be close to the real situation considering the tire is able to separate from the ground plane. A set of piecewise linear mathematical models is developed and matches the in-contact and no-contact states to be considered as mother models for further investigations. The bound dynamics are numerically simulated in the time response and phase portraits. The separation analysis may determine which values of suspension parameters can delay and avoid the no-contact phenomenon, which results in improving ride comfort and eliminating the potentially dangerous oscillation. Finally, model verification is carried out in the MSC-ADAMS environment.Keywords: quarter-car vibrations, tire-road separation, separation analysis, separation dynamics, ride comfort, ADAMS validation
Procedia PDF Downloads 9316225 On Reliability of a Credit Default Swap Contract during the EMU Debt Crisis
Authors: Petra Buzkova, Milos Kopa
Abstract:
Reliability of the credit default swap market had been questioned repeatedly during the EMU debt crisis. The article examines whether this development influenced sovereign EMU CDS prices in general. We regress the CDS market price on a model risk neutral CDS price obtained from an adopted reduced form valuation model in the 2009-2013 period. We look for a break point in the single-equation and multi-equation econometric models in order to show the changes in relations between CDS market and model prices. Our results differ according to the risk profile of a country. We find that in the case of riskier countries, the relationship between the market and model price changed when market participants started to question the ability of CDS contracts to protect their buyers. Specifically, it weakened after the change. In the case of less risky countries, the change happened earlier and the effect of a weakened relationship is not observed.Keywords: chow stability test, credit default swap, debt crisis, reduced form valuation model, seemingly unrelated regression
Procedia PDF Downloads 26516224 Application of a Generalized Additive Model to Reveal the Relations between the Density of Zooplankton with Other Variables in the West Daya Bay, China
Authors: Weiwen Li, Hao Huang, Chengmao You, Jianji Liao, Lei Wang, Lina An
Abstract:
Zooplankton are a central issue in the ecology which makes a great contribution to maintaining the balance of an ecosystem. It is critical in promoting the material cycle and energy flow within the ecosystems. A generalized additive model (GAM) was applied to analyze the relationships between the density (individuals per m³) of zooplankton and other variables in West Daya Bay. All data used in this analysis (the survey month, survey station (longitude and latitude), the depth of the water column, the superficial concentration of chlorophyll a, the benthonic concentration of chlorophyll a, the number of zooplankton species and the number of zooplankton species) were collected through monthly scientific surveys during January to December 2016. GLM model (generalized linear model) was used to choose the significant variables’ impact on the density of zooplankton, and the GAM was employed to analyze the relationship between the density of zooplankton and the significant variables. The results showed that the density of zooplankton increased with an increase of the benthonic concentration of chlorophyll a, but decreased with a decrease in the depth of the water column. Both high numbers of zooplankton species and the overall total number of zooplankton individuals led to a higher density of zooplankton.Keywords: density, generalized linear model, generalized additive model, the West Daya Bay, zooplankton
Procedia PDF Downloads 15416223 Design and Development of an Algorithm to Predict Fluctuations of Currency Rates
Authors: Nuwan Kuruwitaarachchi, M. K. M. Peiris, C. N. Madawala, K. M. A. R. Perera, V. U. N Perera
Abstract:
Dealing with businesses with the foreign market always took a special place in a country’s economy. Political and social factors came into play making currency rate changes fluctuate rapidly. Currency rate prediction has become an important factor for larger international businesses since large amounts of money exchanged between countries. This research focuses on comparing the accuracy of mainly three models; Autoregressive Integrated Moving Average (ARIMA), Artificial Neural Networks(ANN) and Support Vector Machines(SVM). series of data import, export, USD currency exchange rate respect to LKR has been selected for training using above mentioned algorithms. After training the data set and comparing each algorithm, it was able to see that prediction in SVM performed better than other models. It was improved more by combining SVM and SVR models together.Keywords: ARIMA, ANN, FFNN, RMSE, SVM, SVR
Procedia PDF Downloads 21416222 Effect of Different Porous Media Models on Drug Delivery to Solid Tumors: Mathematical Approach
Authors: Mostafa Sefidgar, Sohrab Zendehboudi, Hossein Bazmara, Madjid Soltani
Abstract:
Based on findings from clinical applications, most drug treatments fail to eliminate malignant tumors completely even though drug delivery through systemic administration may inhibit their growth. Therefore, better understanding of tumor formation is crucial in developing more effective therapeutics. For this purpose, nowadays, solid tumor modeling and simulation results are used to predict how therapeutic drugs are transported to tumor cells by blood flow through capillaries and tissues. A solid tumor is investigated as a porous media for fluid flow simulation. Most of the studies use Darcy model for porous media. In Darcy model, the fluid friction is neglected and a few simplified assumptions are implemented. In this study, the effect of these assumptions is studied by considering Brinkman model. A multi scale mathematical method which calculates fluid flow to a solid tumor is used in this study to investigate how neglecting fluid friction affects the solid tumor simulation. In this work, the mathematical model in our previous studies is developed by considering two model of momentum equation for porous media: Darcy and Brinkman. The mathematical method involves processes such as fluid flow through solid tumor as porous media, extravasation of blood flow from vessels, blood flow through vessels and solute diffusion, convective transport in extracellular matrix. The sprouting angiogenesis model is used for generating capillary network and then fluid flow governing equations are implemented to calculate blood flow through the tumor-induced capillary network. Finally, the two models of porous media are used for modeling fluid flow in normal and tumor tissues in three different shapes of tumors. Simulations of interstitial fluid transport in a solid tumor demonstrate that the simplifications used in Darcy model affect the interstitial velocity and Brinkman model predicts a lower value for interstitial velocity than the values that Darcy model does.Keywords: solid tumor, porous media, Darcy model, Brinkman model, drug delivery
Procedia PDF Downloads 30816221 Synthesis and Characterization of Model Amines for Corrosion Applications
Authors: John Vergara, Giuseppe Palmese
Abstract:
Fundamental studies aimed at elucidating the key contributions to corrosion performance are needed to make progress toward effective and environmentally compliant corrosion control. Epoxy/amine systems are typically employed as barrier coatings for corrosion control. However, the hardening agents used for coating applications can be very complex, making fundamental studies of water and oxygen permeability challenging to carry out. Creating model building blocks for epoxy/amine coatings is the first step in carrying out these studies. We will demonstrate the synthesis and characterization of model amine building blocks from saturated fatty acids and simple amines such as diethylenetriamine (DETA) and Bis(3-aminopropyl)amine. The structure-property relationship of thermosets made from these model amines and Diglycidyl ether of bisphenol A (DGBEA) will be discussed.Keywords: building block, amine, synthesis, characterization
Procedia PDF Downloads 54416220 Estimation of Transition and Emission Probabilities
Authors: Aakansha Gupta, Neha Vadnere, Tapasvi Soni, M. Anbarsi
Abstract:
Protein secondary structure prediction is one of the most important goals pursued by bioinformatics and theoretical chemistry; it is highly important in medicine and biotechnology. Some aspects of protein functions and genome analysis can be predicted by secondary structure prediction. This is used to help annotate sequences, classify proteins, identify domains, and recognize functional motifs. In this paper, we represent protein secondary structure as a mathematical model. To extract and predict the protein secondary structure from the primary structure, we require a set of parameters. Any constants appearing in the model are specified by these parameters, which also provide a mechanism for efficient and accurate use of data. To estimate these model parameters there are many algorithms out of which the most popular one is the EM algorithm or called the Expectation Maximization Algorithm. These model parameters are estimated with the use of protein datasets like RS126 by using the Bayesian Probabilistic method (data set being categorical). This paper can then be extended into comparing the efficiency of EM algorithm to the other algorithms for estimating the model parameters, which will in turn lead to an efficient component for the Protein Secondary Structure Prediction. Further this paper provides a scope to use these parameters for predicting secondary structure of proteins using machine learning techniques like neural networks and fuzzy logic. The ultimate objective will be to obtain greater accuracy better than the previously achieved.Keywords: model parameters, expectation maximization algorithm, protein secondary structure prediction, bioinformatics
Procedia PDF Downloads 483