Search results for: statistical modeling
6875 Reliability Modeling on Drivers’ Decision during Yellow Phase
Authors: Sabyasachi Biswas, Indrajit Ghosh
Abstract:
The random and heterogeneous behavior of vehicles in India puts up a greater challenge for researchers. Stop-and-go modeling at signalized intersections under heterogeneous traffic conditions has remained one of the most sought-after fields. Vehicles are often caught up in the dilemma zone and are unable to take quick decisions whether to stop or cross the intersection. This hampers the traffic movement and may lead to accidents. The purpose of this work is to develop a stop and go prediction model that depicts the drivers’ decision during the yellow time at signalised intersections. To accomplish this, certain traffic parameters were taken into account to develop surrogate model. This research investigated the Stop and Go behavior of the drivers by collecting data from 4-signalized intersections located in two major Indian cities. Model was developed to predict the drivers’ decision making during the yellow phase of the traffic signal. The parameters used for modeling included distance to stop line, time to stop line, speed, and length of the vehicle. A Kriging base surrogate model has been developed to investigate the drivers’ decision-making behavior in amber phase. It is observed that the proposed approach yields a highly accurate result (97.4 percent) by Gaussian function. It was observed that the accuracy for the crossing probability was 95.45, 90.9 and 86.36.11 percent respectively as predicted by the Kriging models with Gaussian, Exponential and Linear functions.Keywords: decision-making decision, dilemma zone, surrogate model, Kriging
Procedia PDF Downloads 3096874 Catalytic Thermodynamics of Nanocluster Adsorbates from Informational Statistical Mechanics
Authors: Forrest Kaatz, Adhemar Bultheel
Abstract:
We use an informational statistical mechanics approach to study the catalytic thermodynamics of platinum and palladium cuboctahedral nanoclusters. Nanoclusters and their adatoms are viewed as chemical graphs with a nearest neighbor adjacency matrix. We use the Morse potential to determine bond energies between cluster atoms in a coordination type calculation. We use adsorbate energies calculated from density functional theory (DFT) to study the adatom effects on the thermodynamic quantities, which are derived from a Hamiltonian. Oxygen radical and molecular adsorbates are studied on platinum clusters and hydrogen on palladium clusters. We calculate the entropy, free energy, and total energy as the coverage of adsorbates increases from bridge and hollow sites on the surface. Thermodynamic behavior versus adatom coverage is related to the structural distribution of adatoms on the nanocluster surfaces. The thermodynamic functions are characterized using a simple adsorption model, with linear trends as the coverage of adatoms increases. The data exhibits size effects for the measured thermodynamic properties with cluster diameters between 2 and 5 nm. Entropy and enthalpy calculations of Pt-O2 compare well with previous theoretical data for Pt(111)-O2, and our Pd-H results show similar trends as experimental measurements for Pd-H2 nanoclusters. Our methods are general and may be applied to wide variety of nanocluster adsorbate systems.Keywords: catalytic thermodynamics, palladium nanocluster absorbates, platinum nanocluster absorbates, statistical mechanics
Procedia PDF Downloads 1666873 Analyzing the Evolution and Maturation of Bitcoin Improvement Proposals
Authors: Rodrigo Costa, Thomas Mazzuchi, Shahram Sarkani
Abstract:
This study analyzes the evolution of Bitcoin Improvement Proposals (BIPs), the self-governing mechanism that enables updates to the Bitcoin protocol. By modeling BIP submission frequencies with a Negative Binomial distribution and detecting change points with the Pelt Rupture model, we identify three distinct intervals of proposal activity, suggesting shifts in development priorities over time. Long-term growth patterns, captured by Gompertz and Weibull models, indicate an S-shaped trend in cumulative BIP counts, pointing toward a maturation phase in Bitcoin’s protocol. Our findings suggest that Bitcoin may be entering a stable stage, with fewer fundamental changes and more incremental enhancements. This trend highlights the need for further research into BIP content and more studies into its dynamics to better understand decentralized protocol governance and maturation.Keywords: bitcoin improvement proposals, innovation management, change point detection, systems modeling, simulation
Procedia PDF Downloads 96872 Therapeutic Effect of 12 Weeks of Sensorimotor Exercise on Pain, Functionality and Quality of Life in Non-athlete Women With Patellofemoral Pain Syndrome
Authors: Kasbparast Mehdi, Hassani Zainab
Abstract:
Aim: The purpose of this research was to investigate the effectiveness of therapeutical sensorimotor exercise. The statistical population of women who were diagnosed with patellofemoral pain syndrome by a doctor and were between the ages of 35 and 45 and registered for the first time in a sports club in the 4th district of Tehran, 30 people by random sampling and according to The include and exclude criteria were selected and divided into 2 equal control and experimental and homogeneous groups (in terms of height, weight and BMI).In both control and experimental groups, the pain was measured using a Visual Analog Scale(VAS) functionality was measured using the step-down test and quality of life was measured using a World Health Organization Quality of Life Scale (WHOQOL-BREF) (pre-test). Then, only the experimental group performed sensorimotor exercises for 12 weeks and 3 sessions each week, a total of 24 sessions and each session for 1 hour, and during this period, the control group only continued their daily activities. After the end of the training period, the desired factors were evaluated again (post-test) in the same way as the pre-test was done for them (experimental group and control group), with the same quality. Findings: The statistical results showed that in the experimental group, the amount of pain, function and quality of life had a statistical improvement (P≤0.05). Conclusion: In general conclusion, it can be stated that using sensorimotor exercises not only improved functionality and quality of life but also reduced the amount of pain in people with patellofemoral pain syndrome.Keywords: pain, PFPS, sensori motor training, functionality
Procedia PDF Downloads 756871 Analysis and Modeling of Photovoltaic System with Different Research Methods of Maximum Power Point Tracking
Authors: Mehdi Ameur, Ahmed Essakdi, Tamou Nasser
Abstract:
The purpose of this paper is the analysis and modeling of the photovoltaic system with MPPT techniques. This system is developed by combining the models of established solar module and DC-DC converter with the algorithms of perturb and observe (P&O), incremental conductance (INC) and fuzzy logic controller(FLC). The system is simulated under different climate conditions and MPPT algorithms to determine the influence of these conditions on characteristic power-voltage of PV system. According to the comparisons of the simulation results, the photovoltaic system can extract the maximum power with precision and rapidity using the MPPT algorithms discussed in this paper.Keywords: photovoltaic array, maximum power point tracking, MPPT, perturb and observe, P&O, incremental conductance, INC, hill climbing, HC, fuzzy logic controller, FLC
Procedia PDF Downloads 4296870 An Implementation of Meshless Method for Modeling an Elastoplasticity Coupled to Damage
Authors: Sendi Zohra, Belhadjsalah Hedi, Labergere Carl, Saanouni Khemais
Abstract:
The modeling of mechanical problems including both material and geometric nonlinearities with Finite Element Method (FEM) remains challenging. Meshless methods offer special properties to get rid of well-known drawbacks of the FEM. The main objective of Meshless Methods is to eliminate the difficulty of meshing and remeshing the entire structure by simply insertion or deletion of nodes, and alleviate other problems associated with the FEM, such as element distortion, locking and others. In this study, a robust numerical implementation of an Element Free Galerkin Method for an elastoplastic coupled to damage problem is presented. Several results issued from the numerical simulations by a DynamicExplicit resolution scheme are analyzed and critically compared with Element Finite Method results. Finally, different numerical examples are carried out to demonstrate the efficiency of this method.Keywords: damage, dynamic explicit, elastoplasticity, isotropic hardening, meshless
Procedia PDF Downloads 2956869 A Formal Verification Approach for Linux Kernel Designing
Authors: Zi Wang, Xinlei He, Jianghua Lv, Yuqing Lan
Abstract:
Kernel though widely used, is complicated. Errors caused by some bugs are often costly. Statically, more than half of the mistakes occur in the design phase. Thus, we introduce a modeling method, KMVM (Linux Kernel Modeling and verification Method), based on type theory for proper designation and correct exploitation of the Kernel. In the model, the Kernel is separated into six levels: subsystem, dentry, file, struct, func, and base. Each level is treated as a type. The types are specified in the structure and relationship. At the same time, we use a demanding path to express the function to be implemented. The correctness of the design is verified by recursively checking the type relationship and type existence. The method has been applied to verify the OPEN business of VFS (virtual file system) in Linux Kernel. Also, we have designed and developed a set of security communication mechanisms in the Kernel with verification.Keywords: formal approach, type theory, Linux Kernel, software program
Procedia PDF Downloads 1376868 Pitfalls and Drawbacks in Visual Modelling of Learning Knowledge by Students
Authors: Tatyana Gavrilova, Vadim Onufriev
Abstract:
Knowledge-based systems’ design requires the developer’s owning the advanced analytical skills. The efficient development of that skills within university courses needs a deep understanding of main pitfalls and drawbacks, which students usually make during their analytical work in form of visual modeling. Thus, it was necessary to hold an analysis of 5-th year students’ learning exercises within courses of 'Intelligent systems' and 'Knowledge engineering' in Saint-Petersburg Polytechnic University. The analysis shows that both lack of system thinking skills and methodological mistakes in course design cause the errors that are discussed in the paper. The conclusion contains an exploration of the issues and topics necessary and sufficient for the implementation of the improved practices in educational design for future curricula of teaching programs.Keywords: knowledge based systems, knowledge engineering, students’ errors, visual modeling
Procedia PDF Downloads 3116867 Secured Embedding of Patient’s Confidential Data in Electrocardiogram Using Chaotic Maps
Authors: Butta Singh
Abstract:
This paper presents a chaotic map based approach for secured embedding of patient’s confidential data in electrocardiogram (ECG) signal. The chaotic map generates predefined locations through the use of selective control parameters. The sample value difference method effectually hides the confidential data in ECG sample pairs at these predefined locations. Evaluation of proposed method on all 48 records of MIT-BIH arrhythmia ECG database demonstrates that the embedding does not alter the diagnostic features of cover ECG. The secret data imperceptibility in stego-ECG is evident through various statistical and clinical performance measures. Statistical metrics comprise of Percentage Root Mean Square Difference (PRD) and Peak Signal to Noise Ratio (PSNR). Further, a comparative analysis between proposed method and existing approaches was also performed. The results clearly demonstrated the superiority of proposed method.Keywords: chaotic maps, ECG steganography, data embedding, electrocardiogram
Procedia PDF Downloads 1966866 Modeling of Combustion Process in the Piston Aircraft Engine Using a MCFM-3Z Model
Authors: Marcin Szlachetka, Konrad Pietrykowski
Abstract:
Modeling of a combustion process in a 9-cylinder aircraft engine is presented. The simulations of the combustion process in the IC engine have provided the information on the spatial and time distributions of selected quantities within the combustion chamber of the engine. The numerical analysis results have been compared with the results of indication process of the engine on the test stand. Modeling of combustion process an auto-ignited IC engine in the AVL Fire was carried out within the study. For the calculations, a ECFM-3Z model was used. Verification of simulation results was carried out by comparison of the pressure in the cylinder. The courses of indicated pressure, obtained from the simulations and during the engine tests mounted on a test stand were compared. The engine was braked by the propeller, which results in an adequate external power characteristics. The test object is a modified ASz-62IR engine with the injection system. The engine was running at take-off power. To check the optimum ignition timing regarding power, calculations, tests were performed for 7 different moments of ignition. The analyses of temperature distribution in the cylinder depending on the moments of ignition were carried out. Additional the course of pressure in the cylinder at different angles of ignition delays of the second spark plug were examined. The swirling of the mixture in the combustion chamber was also analysed. It has been shown that the largest vortexes occur in the middle of the chamber, and gets smaller, closer to the combustion chamber walls. This work has been financed by the Polish National Centre for Research and Development, INNOLOT, under Grant Agreement No. INNOLOT/I/1/NCBR/2013.Keywords: CFD, combustion, internal combustion engine, aircraft engine
Procedia PDF Downloads 3726865 Factors that Predict Pre-Service Teachers' Decision to Integrate E-Learning: A Structural Equation Modeling (SEM) Approach
Authors: Mohd Khairezan Rahmat
Abstract:
Since the impetus of becoming a develop country by the year 2020, the Malaysian government have been proactive in strengthening the integration of ICT into the national educational system. Teacher-education programs have the responsibility to prepare the nation future teachers by instilling in them the desire, confidence, and ability to fully utilized the potential of ICT into their instruction process. In an effort to fulfill this responsibility, teacher-education program are beginning to create alternatives means for preparing cutting-edge teachers. One of the alternatives is the student’s learning portal. In line with this mission, this study investigates the Faculty of Education, University Teknologi MARA (UiTM) pre-service teachers’ perception of usefulness, attitude, and ability toward the usage of the university learning portal, known as iLearn. The study also aimed to predict factors that might hinder the pre-service teachers’ decision to used iLearn as their platform in learning. The Structural Equation Modeling (SEM), was employed in analyzed the survey data. The suggested findings informed that pre-service teacher’s successful integration of the iLearn was highly influenced by their perception of usefulness of the system. The findings also suggested that the more familiar the pre-service teacher with the iLearn, the more possibility they will use the system. In light of similar study, the present findings hope to highlight the important to understand the user’s perception toward any proposed technology.Keywords: e-learning, prediction factors, pre-service teacher, structural equation modeling (SEM)
Procedia PDF Downloads 3396864 A Stochastic Analytic Hierarchy Process Based Weighting Model for Sustainability Measurement in an Organization
Authors: Faramarz Khosravi, Gokhan Izbirak
Abstract:
A weighted statistical stochastic based Analytical Hierarchy Process (AHP) model for modeling the potential barriers and enablers of sustainability for measuring and assessing the sustainability level is proposed. For context-dependent potential barriers and enablers, the proposed model takes the basis of the properties of the variables describing the sustainability functions and was developed into a realistic analytical model for the sustainable behavior of an organization. This thus serves as a means for measuring the sustainability of the organization. The main focus of this paper was the application of the AHP tool in a statistically-based model for measuring sustainability. Hence a strong weighted stochastic AHP based procedure was achieved. A case study scenario of a widely reported major Canadian electric utility was adopted to demonstrate the applicability of the developed model and comparatively examined its results with those of an equal-weighted model method. Variations in the sustainability of a company, as fluctuations, were figured out during the time. In the results obtained, sustainability index for successive years changed form 73.12%, 79.02%, 74.31%, 76.65%, 80.49%, 79.81%, 79.83% to more exact values 73.32%, 77.72%, 76.76%, 79.41%, 81.93%, 79.72%, and 80,45% according to priorities of factors that have found by expert views, respectively. By obtaining relatively necessary informative measurement indicators, the model can practically and effectively evaluate the sustainability extent of any organization and also to determine fluctuations in the organization over time.Keywords: AHP, sustainability fluctuation, environmental indicators, performance measurement
Procedia PDF Downloads 1216863 Modeling the Effects of Leachate-Impacted Groundwater on the Water Quality of a Large Tidal River
Authors: Emery Coppola Jr., Marwan Sadat, Il Kim, Diane Trube, Richard Kurisko
Abstract:
Contamination sites like landfills often pose significant risks to receptors like surface water bodies. Surface water bodies are often a source of recreation, including fishing and swimming, which not only enhances their value but also serves as a direct exposure pathway to humans, increasing their need for protection from water quality degradation. In this paper, a case study presents the potential effects of leachate-impacted groundwater from a large closed sanitary landfill on the surface water quality of the nearby Raritan River, situated in New Jersey. The study, performed over a two year period, included in-depth field evaluation of both the groundwater and surface water systems, and was supplemented by computer modeling. The analysis required delineation of a representative average daily groundwater discharge from the Landfill shoreline into the large, highly tidal Raritan River, with a corresponding estimate of daily mass loading of potential contaminants of concern. The average daily groundwater discharge into the river was estimated from a high-resolution water level study and a 24-hour constant-rate aquifer pumping test. The significant tidal effects induced on groundwater levels during the aquifer pumping test were filtered out using an advanced algorithm, from which aquifer parameter values were estimated using conventional curve match techniques. The estimated hydraulic conductivity values obtained from individual observation wells closely agree with tidally-derived values for the same wells. Numerous models were developed and used to simulate groundwater contaminant transport and surface water quality impacts. MODFLOW with MT3DMS was used to simulate the transport of potential contaminants of concern from the down-gradient edge of the Landfill to the Raritan River shoreline. A surface water dispersion model based upon a bathymetric and flow study of the river was used to simulate the contaminant concentrations over space within the river. The modeling results helped demonstrate that because of natural attenuation, the Landfill does not have a measurable impact on the river, which was confirmed by an extensive surface water quality study.Keywords: groundwater flow and contaminant transport modeling, groundwater/surface water interaction, landfill leachate, surface water quality modeling
Procedia PDF Downloads 2616862 Laser-Dicing Modeling: Implementation of a High Accuracy Tool for Laser-Grooving and Cutting Application
Authors: Jeff Moussodji, Dominique Drouin
Abstract:
The highly complex technology requirements of today’s integrated circuits (ICs), lead to the increased use of several materials types such as metal structures, brittle and porous low-k materials which are used in both front end of line (FEOL) and back end of line (BEOL) process for wafer manufacturing. In order to singulate chip from wafer, a critical laser-grooving process, prior to blade dicing, is used to remove these layers of materials out of the dicing street. The combination of laser-grooving and blade dicing allows to reduce the potential risk of induced mechanical defects such micro-cracks, chipping, on the wafer top surface where circuitry is located. It seems, therefore, essential to have a fundamental understanding of the physics involving laser-dicing in order to maximize control of these critical process and reduce their undesirable effects on process efficiency, quality, and reliability. In this paper, the study was based on the convergence of two approaches, numerical and experimental studies which allowed us to investigate the interaction of a nanosecond pulsed laser and BEOL wafer materials. To evaluate this interaction, several laser grooved samples were compared with finite element modeling, in which three different aspects; phase change, thermo-mechanical and optic sensitive parameters were considered. The mathematical model makes it possible to highlight a groove profile (depth, width, etc.) of a single pulse or multi-pulses on BEOL wafer material. Moreover, the heat affected zone, and thermo-mechanical stress can be also predicted as a function of laser operating parameters (power, frequency, spot size, defocus, speed, etc.). After modeling validation and calibration, a satisfying correlation between experiment and modeling, results have been observed in terms of groove depth, width and heat affected zone. The study proposed in this work is a first step toward implementing a quick assessment tool for design and debug of multiple laser grooving conditions with limited experiments on hardware in industrial application. More correlations and validation tests are in progress and will be included in the full paper.Keywords: laser-dicing, nano-second pulsed laser, wafer multi-stack, multiphysics modeling
Procedia PDF Downloads 2096861 Statistical Analysis of Rainfall Change over the Blue Nile Basin
Authors: Hany Mustafa, Mahmoud Roushdi, Khaled Kheireldin
Abstract:
Rainfall variability is an important feature of semi-arid climates. Climate change is very likely to increase the frequency, magnitude, and variability of extreme weather events such as droughts, floods, and storms. The Blue Nile Basin is facing extreme climate change-related events such as floods and droughts and its possible impacts on ecosystem, livelihood, agriculture, livestock, and biodiversity are expected. Rainfall variability is a threat to food production in the Blue Nile Basin countries. This study investigates the long-term variations and trends of seasonal and annual precipitation over the Blue Nile Basin for 102-year period (1901-2002). Six statistical trend analysis of precipitation was performed with nonparametric Mann-Kendall test and Sen's slope estimator. On the other hands, four statistical absolute homogeneity tests: Standard Normal Homogeneity Test, Buishand Range test, Pettitt test and the Von Neumann ratio test were applied to test the homogeneity of the rainfall data, using XLSTAT software, which results of p-valueless than alpha=0.05, were significant. The percentages of significant trends obtained for each parameter in the different seasons are presented. The study recommends adaptation strategies to be streamlined to relevant policies, enhancing local farmers’ adaptive capacity for facing future climate change effects.Keywords: Blue Nile basin, climate change, Mann-Kendall test, trend analysis
Procedia PDF Downloads 5506860 Various Advanced Statistical Analyses of Index Values Extracted from Outdoor Agricultural Workers Motion Data
Authors: Shinji Kawakura, Ryosuke Shibasaki
Abstract:
We have been grouping and developing various kinds of practical, promising sensing applied systems concerning agricultural advancement and technical tradition (guidance). These include advanced devices to secure real-time data related to worker motion, and we analyze by methods of various advanced statistics and human dynamics (e.g. primary component analysis, Ward system based cluster analysis, and mapping). What is more, we have been considering worker daily health and safety issues. Targeted fields are mainly common farms, meadows, and gardens. After then, we observed and discussed time-line style, changing data. And, we made some suggestions. The entire plan makes it possible to improve both the aforementioned applied systems and farms.Keywords: advanced statistical analysis, wearable sensing system, tradition of skill, supporting for workers, detecting crisis
Procedia PDF Downloads 3946859 Chemometric QSRR Evaluation of Behavior of s-Triazine Pesticides in Liquid Chromatography
Authors: Lidija R. Jevrić, Sanja O. Podunavac-Kuzmanović, Strahinja Z. Kovačević
Abstract:
This study considers the selection of the most suitable in silico molecular descriptors that could be used for s-triazine pesticides characterization. Suitable descriptors among topological, geometrical and physicochemical are used for quantitative structure-retention relationships (QSRR) model establishment. Established models were obtained using linear regression (LR) and multiple linear regression (MLR) analysis. In this paper, MLR models were established avoiding multicollinearity among the selected molecular descriptors. Statistical quality of established models was evaluated by standard and cross-validation statistical parameters. For detection of similarity or dissimilarity among investigated s-triazine pesticides and their classification, principal component analysis (PCA) and hierarchical cluster analysis (HCA) were used and gave similar grouping. This study is financially supported by COST action TD1305.Keywords: chemometrics, classification analysis, molecular descriptors, pesticides, regression analysis
Procedia PDF Downloads 3936858 Modeling of Age Hardening Process Using Adaptive Neuro-Fuzzy Inference System: Results from Aluminum Alloy A356/Cow Horn Particulate Composite
Authors: Chidozie C. Nwobi-Okoye, Basil Q. Ochieze, Stanley Okiy
Abstract:
This research reports on the modeling of age hardening process using adaptive neuro-fuzzy inference system (ANFIS). The age hardening output (Hardness) was predicted using ANFIS. The input parameters were ageing time, temperature and percentage composition of cow horn particles (CHp%). The results show the correlation coefficient (R) of the predicted hardness values versus the measured values was of 0.9985. Subsequently, values outside the experimental data points were predicted. When the temperature was kept constant, and other input parameters were varied, the average relative error of the predicted values was 0.0931%. When the temperature was varied, and other input parameters kept constant, the average relative error of the hardness values predictions was 80%. The results show that ANFIS with coarse experimental data points for learning is not very effective in predicting process outputs in the age hardening operation of A356 alloy/CHp particulate composite. The fine experimental data requirements by ANFIS make it more expensive in modeling and optimization of age hardening operations of A356 alloy/CHp particulate composite.Keywords: adaptive neuro-fuzzy inference system (ANFIS), age hardening, aluminum alloy, metal matrix composite
Procedia PDF Downloads 1546857 A Fuzzy Inference Tool for Assessing Cancer Risk from Radiation Exposure
Authors: Bouharati Lokman, Bouharati Imen, Bouharati Khaoula, Bouharati Oussama, Bouharati Saddek
Abstract:
Ionizing radiation exposure is an established cancer risk factor. Compared to other common environmental carcinogens, it is relatively easy to determine organ-specific radiation dose and, as a result, radiation dose-response relationships tend to be highly quantified. Nevertheless, there can be considerable uncertainty about questions of radiation-related cancer risk as they apply to risk protection and public policy, and the interpretations of interested parties can differ from one person to another. Examples of tools used in the analysis of the risk of developing cancer due to radiation are characterized by uncertainty. These uncertainties are related to the history of exposure and different assumptions involved in the calculation. We believe that the results of statistical calculations are characterized by uncertainty and imprecision. Having regard to the physiological variation from one person to another. In this study, we develop a tool based on fuzzy logic inference. As fuzzy logic deals with imprecise and uncertain, its application in this area is adequate. We propose a fuzzy system with three input variables (age, sex and body attainable cancer). The output variable expresses the risk of infringement rate of each organ. A base rule is established from recorded actual data. After successful simulation, this will instantly predict the risk of infringement rate of each body following chronic exposure to 0.1 Gy.Keywords: radiation exposure, cancer, modeling, fuzzy logic
Procedia PDF Downloads 3116856 A Robust Software for Advanced Analysis of Space Steel Frames
Authors: Viet-Hung Truong, Seung-Eock Kim
Abstract:
This paper presents a robust software package for practical advanced analysis of space steel framed structures. The pre- and post-processors of the presented software package are coded in the C++ programming language while the solver is written by using the FORTRAN programming language. A user-friendly graphical interface of the presented software is developed to facilitate the modeling process and result interpretation of the problem. The solver employs the stability functions for capturing the second-order effects to minimize modeling and computational time. Both the plastic-hinge and fiber-hinge beam-column elements are available in the presented software. The generalized displacement control method is adopted to solve the nonlinear equilibrium equations.Keywords: advanced analysis, beam-column, fiber-hinge, plastic hinge, steel frame
Procedia PDF Downloads 3076855 The Methodology of System Modeling of Mechatronic Systems
Authors: Lakhoua Najeh
Abstract:
Aims of the work: After a presentation of the functionality of an example of a mechatronic system which is a paint mixer system, we present the concepts of modeling and safe operation. This paper briefly discusses how to model and protect the functioning of a mechatronic system relying mainly on functional analysis and safe operation techniques. Methods: For the study of an example of a mechatronic system, we use methods for external functional analysis that illustrate the relationships between a mechatronic system and its external environment. Thus, we present the Safe-Structured Analysis Design Technique method (Safe-SADT) which allows the representation of a mechatronic system. A model of operating safety and automation is proposed. This model enables us to use a functional analysis technique of the mechatronic system based on the GRAFCET (Graphe Fonctionnel de Commande des Etapes et Transitions: Step Transition Function Chart) method; study of the safe operation of the mechatronic system based on the Safe-SADT method; automation of the mechatronic system based on a software tool. Results: The expected results are to propose a model and safe operation of a mechatronic system. This methodology enables us to analyze the relevance of the different models based on Safe-SADT and GRAFCET in relation to the control and monitoring functions and to study the means allowing exploiting their synergy. Conclusion: In order to propose a general model of a mechatronic system, a model of analysis, safety operation and automation of a mechatronic system has been developed. This is how we propose to validate this methodology through a case study of a paint mixer system.Keywords: mechatronic systems, system modeling, safe operation, Safe-SADT
Procedia PDF Downloads 2456854 A Cross-Dialect Statistical Analysis of Final Declarative Intonation in Tuvinian
Authors: D. Beziakina, E. Bulgakova
Abstract:
This study continues the research on Tuvinian intonation and presents a general cross-dialect analysis of intonation of Tuvinian declarative utterances, specifically the character of the tone movement in order to test the hypothesis about the prevalence of level tone in some Tuvinian dialects. The results of the analysis of basic pitch characteristics of Tuvinian speech (in general and in comparison with two other Turkic languages - Uzbek and Azerbaijani) are also given in this paper. The goal of our work was to obtain the ranges of pitch parameter values typical for Tuvinian speech. Such language-specific values can be used in speaker identification systems in order to get more accurate results of ethnic speech analysis. We also present the results of a cross-dialect analysis of declarative intonation in the poorly studied Tuvinian language.Keywords: speech analysis, statistical analysis, speaker recognition, identification of person
Procedia PDF Downloads 4706853 Multi-Source Data Fusion for Urban Comprehensive Management
Authors: Bolin Hua
Abstract:
In city governance, various data are involved, including city component data, demographic data, housing data and all kinds of business data. These data reflects different aspects of people, events and activities. Data generated from various systems are different in form and data source are different because they may come from different sectors. In order to reflect one or several facets of an event or rule, data from multiple sources need fusion together. Data from different sources using different ways of collection raised several issues which need to be resolved. Problem of data fusion include data update and synchronization, data exchange and sharing, file parsing and entry, duplicate data and its comparison, resource catalogue construction. Governments adopt statistical analysis, time series analysis, extrapolation, monitoring analysis, value mining, scenario prediction in order to achieve pattern discovery, law verification, root cause analysis and public opinion monitoring. The result of Multi-source data fusion is to form a uniform central database, which includes people data, location data, object data, and institution data, business data and space data. We need to use meta data to be referred to and read when application needs to access, manipulate and display the data. A uniform meta data management ensures effectiveness and consistency of data in the process of data exchange, data modeling, data cleansing, data loading, data storing, data analysis, data search and data delivery.Keywords: multi-source data fusion, urban comprehensive management, information fusion, government data
Procedia PDF Downloads 3936852 The Use of Creativity to Nudge Students Into Heutagogy: An Implementation in Graduate Business Education
Authors: Ricardo Bragança, Tom Vinaimont
Abstract:
This paper discusses the introduction of processes of self-determined learning (heutagogy) into a graduate course on financial modeling, using elements of entangled pedagogy and Biggs’ constructive alignment. To encourage learners to take control of their own learning journey and develop critical thinking and problem-solving skills, each session in the course receives tailor-made media-enhanced pedagogical assets. The design of those assets specifically supports entangled pedagogy, which opposes technological or pedagogical determinism in support of the collaborative integration of pedagogy and technology. Media assets for each of the ten sessions in this course consist of three components. The first component in this three-pronged approach is a game-cut-like cinematographic representation that introduces the context of the session. The second component represents a character from an open-source-styled community that encourages self-determined learning. The third component consists of a character, which refers to the in-person instructor and also aligns learning outcomes and assessment tasks, using Biggs’ constructive alignment, to the cinematographic and open-source-styled component. In essence, the course's metamorphosis helps students apply the concepts they've studied to actual financial modeling issues. The audio-visual media assets create a storyline throughout the course based on gamified and real-world applications, thus encouraging student engagement and interaction. The structured entanglement of pedagogy and technology also guides the instructor in the design of the in-class interactions and directs the focus on outcomes and assessments. The transformation process of this graduate course in financial modeling led to an institutional teaching award in 2021. The transformation of this course may be used as a model for other courses and programs in many disciplines to help with intended learning outcomes integration, constructive alignment, and Assurance of Learning.Keywords: innovative education, active learning, entangled pedagogy, heutagogy, constructive alignment, project based learning, financial modeling, graduate business education
Procedia PDF Downloads 726851 Discrete Choice Modeling in Education: Evaluating Early Childhood Educators’ Practices
Authors: Michalis Linardakis, Vasilis Grammatikopoulos, Athanasios Gregoriadis, Kalliopi Trouli
Abstract:
Discrete choice models belong to the family of Conjoint analysis that are applied on the preferences of the respondents towards a set of scenarios that describe alternative choices. The scenarios have been pre-designed to cover all the attributes of the alternatives that may affect the choices. In this study, we examine how preschool educators integrate physical activities into their everyday teaching practices through the use of discrete choice models. One of the advantages of discrete choice models compared to other more traditional data collection methods (e.g. questionnaires and interviews that use ratings) is that the respondent is called to select among competitive and realistic alternatives, rather than objectively rate each attribute that the alternatives may have. We present the effort to construct and choose representative attributes that would cover all possible choices of the respondents, and the scenarios that have arisen. For the purposes of the study, we used a sample of 50 preschool educators in Greece that responded to 4 scenarios (from the total of 16 scenarios that the orthogonal design resulted), with each scenario having three alternative teaching practices. Seven attributes of the alternatives were used in the scenarios. For the analysis of the data, we used multinomial logit model with random effects, multinomial probit model and generalized mixed logit model. The conclusions drawn from the estimated parameters of the models are discussed.Keywords: conjoint analysis, discrete choice models, educational data, multivariate statistical analysis
Procedia PDF Downloads 4656850 A Statistical Model for the Dynamics of Single Cathode Spot in Vacuum Cylindrical Cathode
Authors: Po-Wen Chen, Jin-Yu Wu, Md. Manirul Ali, Yang Peng, Chen-Te Chang, Der-Jun Jan
Abstract:
Dynamics of cathode spot has become a major part of vacuum arc discharge with its high academic interest and wide application potential. In this article, using a three-dimensional statistical model, we simulate the distribution of the ignition probability of a new cathode spot occurring in different magnetic pressure on old cathode spot surface and at different arcing time. This model for the ignition probability of a new cathode spot was proposed in two typical situations, one by the pure isotropic random walk in the absence of an external magnetic field, other by the retrograde motion in external magnetic field, in parallel with the cathode surface. We mainly focus on developed relationship between the ignition probability density distribution of a new cathode spot and the external magnetic field.Keywords: cathode spot, vacuum arc discharge, transverse magnetic field, random walk
Procedia PDF Downloads 4346849 Application of Watershed Modeling System for Urbanization Management in Tabuk Area, Saudi Arabia
Authors: Abd-Alrahman Embaby, Ayman Abu Halawa, Medhat Ramadan
Abstract:
The infiltrated water into the subsurface activates expansive soil in localized manner, leading to the differential heaving and destructive of the construction. The Watershed Modeling System (WMS) and Hydrologic Engineering Center (HEC-1) are used to delineate and identify the drainage system and basin morphometry in Tabuk area, where flash floods and accumulation of water may take place. Eight drainage basins effect on Tabuk city. Three of them are expected to be high. The flash floods and surface runoff behavior in these basins are important for any protection projects. It was found that the risky areas that contain Tabuk shale could be expanded when exposed to flash floods and/or surface runoff. The resident neighborhoods in the middle of Tabuk city and affected by surface runoff of the tributaries of the basin of Wadi Abu Nishayfah, Na'am and Atanah outlet, represent high-risk zones. These high-risk neighborhoods are Al Qadsiyah, Al Maseif, Arrwdah, Al Nakhil and Al Rajhi. It can be avoided new constructions on these districts. The low or very low-risk zones include the western and the eastern districts. The western side of the city is lying in the upstream of the small basin. It is suitable for a future urban extension. The direction of surface runoff flow or storm water drain discharge should be away from Tabuk city. The quicker the water can flow out, the better it is.Keywords: digital elevation model (DEM), flash floods, Saudi Arabia, Tabuk City, watershed modeling system (WMS)
Procedia PDF Downloads 2646848 Design and Preliminary Evaluation of Benzoxazolone-Based Agents for Targeting Mitochondrial-Located Translocator Protein
Authors: Nidhi Chadha, A. K. Tiwari, Marilyn D. Milton, Anil K. Mishra
Abstract:
Translocator protein (18 kDa) TSPO is highly expressed during microglia activation in neuroinflammation. Although a number of PET ligands have been developed for the visualization of activated microglia, one of the advantageous approaches is to develop potential optical imaging (OI) probe. Our study involves computational screening, synthesis and evaluation of TSPO ligand through various imaging modalities namely PET/SPECT/Optical. The initial computational screening involves pharmacophore modeling from the library designing having oxo-benzooxazol-3-yl-N-phenyl-acetamide groups and synthesis for visualization of efficacy of these compounds as multimodal imaging probes. Structure modeling of monomer, Ala147Thr mutated, parallel and anti-parallel TSPO dimers was performed and docking analysis was performed for distinct binding sites. Computational analysis showed pattern of variable binding profile of known diagnostic ligands and NBMP via interactions with conserved residues along with TSPO’s natural polymorphism of Ala147→Thr, which showed alteration in the binding affinity due to considerable changes in tertiary structure. Preliminary in vitro binding studies shows binding affinity in the range of 1-5 nm and selectivity was also certified by blocking studies. In summary, this skeleton was found to be potential probe for TSPO imaging due to ease in synthesis, appropriate lipophilicity and reach to specific region of brain.Keywords: TSPO, molecular modeling, imaging, docking
Procedia PDF Downloads 4626847 A Kernel-Based Method for MicroRNA Precursor Identification
Authors: Bin Liu
Abstract:
MicroRNAs (miRNAs) are small non-coding RNA molecules, functioning in transcriptional and post-transcriptional regulation of gene expression. The discrimination of the real pre-miRNAs from the false ones (such as hairpin sequences with similar stem-loops) is necessary for the understanding of miRNAs’ role in the control of cell life and death. Since both their small size and sequence specificity, it cannot be based on sequence information alone but requires structure information about the miRNA precursor to get satisfactory performance. Kmers are convenient and widely used features for modeling the properties of miRNAs and other biological sequences. However, Kmers suffer from the inherent limitation that if the parameter K is increased to incorporate long range effects, some certain Kmer will appear rarely or even not appear, as a consequence, most Kmers absent and a few present once. Thus, the statistical learning approaches using Kmers as features become susceptible to noisy data once K becomes large. In this study, we proposed a Gapped k-mer approach to overcome the disadvantages of Kmers, and applied this method to the field of miRNA prediction. Combined with the structure status composition, a classifier called imiRNA-GSSC was proposed. We show that compared to the original imiRNA-kmer and alternative approaches. Trained on human miRNA precursors, this predictor can achieve an accuracy of 82.34 for predicting 4022 pre-miRNA precursors from eleven species.Keywords: gapped k-mer, imiRNA-GSSC, microRNA precursor, support vector machine
Procedia PDF Downloads 1626846 A Model Architecture Transformation with Approach by Modeling: From UML to Multidimensional Schemas of Data Warehouses
Authors: Ouzayr Rabhi, Ibtissam Arrassen
Abstract:
To provide a complete analysis of the organization and to help decision-making, leaders need to have relevant data; Data Warehouses (DW) are designed to meet such needs. However, designing DW is not trivial and there is no formal method to derive a multidimensional schema from heterogeneous databases. In this article, we present a Model-Driven based approach concerning the design of data warehouses. We describe a multidimensional meta-model and also specify a set of transformations starting from a Unified Modeling Language (UML) metamodel. In this approach, the UML metamodel and the multidimensional one are both considered as a platform-independent model (PIM). The first meta-model is mapped into the second one through transformation rules carried out by the Query View Transformation (QVT) language. This proposal is validated through the application of our approach to generating a multidimensional schema of a Balanced Scorecard (BSC) DW. We are interested in the BSC perspectives, which are highly linked to the vision and the strategies of an organization.Keywords: data warehouse, meta-model, model-driven architecture, transformation, UML
Procedia PDF Downloads 160