Search results for: cluster model approach
25360 Deep Learning Framework for Predicting Bus Travel Times with Multiple Bus Routes: A Single-Step Multi-Station Forecasting Approach
Authors: Muhammad Ahnaf Zahin, Yaw Adu-Gyamfi
Abstract:
Bus transit is a crucial component of transportation networks, especially in urban areas. Any intelligent transportation system must have accurate real-time information on bus travel times since it minimizes waiting times for passengers at different stations along a route, improves service reliability, and significantly optimizes travel patterns. Bus agencies must enhance the quality of their information service to serve their passengers better and draw in more travelers since people waiting at bus stops are frequently anxious about when the bus will arrive at their starting point and when it will reach their destination. For solving this issue, different models have been developed for predicting bus travel times recently, but most of them are focused on smaller road networks due to their relatively subpar performance in high-density urban areas on a vast network. This paper develops a deep learning-based architecture using a single-step multi-station forecasting approach to predict average bus travel times for numerous routes, stops, and trips on a large-scale network using heterogeneous bus transit data collected from the GTFS database. Over one week, data was gathered from multiple bus routes in Saint Louis, Missouri. In this study, Gated Recurrent Unit (GRU) neural network was followed to predict the mean vehicle travel times for different hours of the day for multiple stations along multiple routes. Historical time steps and prediction horizon were set up to 5 and 1, respectively, which means that five hours of historical average travel time data were used to predict average travel time for the following hour. The spatial and temporal information and the historical average travel times were captured from the dataset for model input parameters. As adjacency matrices for the spatial input parameters, the station distances and sequence numbers were used, and the time of day (hour) was considered for the temporal inputs. Other inputs, including volatility information such as standard deviation and variance of journey durations, were also included in the model to make it more robust. The model's performance was evaluated based on a metric called mean absolute percentage error (MAPE). The observed prediction errors for various routes, trips, and stations remained consistent throughout the day. The results showed that the developed model could predict travel times more accurately during peak traffic hours, having a MAPE of around 14%, and performed less accurately during the latter part of the day. In the context of a complicated transportation network in high-density urban areas, the model showed its applicability for real-time travel time prediction of public transportation and ensured the high quality of the predictions generated by the model.Keywords: gated recurrent unit, mean absolute percentage error, single-step forecasting, travel time prediction.
Procedia PDF Downloads 7225359 Use of Technology Based Intervention for Continuous Professional Development of Teachers in Pakistan
Authors: Rabia Aslam
Abstract:
Overwhelming evidence from all around the world suggests that high-quality teacher professional development facilitates the improvement of teaching practices which in turn could improve student learning outcomes. The new Continuous Professional Development (CPD) model for primary school teachers in Punjab uses a blended approach in which pedagogical content knowledge is delivered through technology (high-quality instructional videos and lesson plans delivered to school tablets or mobile phones) with face-to-face support by Assistant Education Officers (AEOs). The model also develops Communities of Practice operationalized through formal meetings led by the AEOs and informal interactions through social media groups to provide opportunities for teachers to engage with each other and share their ideas, reflect on learning, and come up with solutions to issues they experience. Using Kirkpatrick’s 4 levels of the learning evaluation model, this paper investigates how school tablets and teacher mobile phones may act as transformational cultural tools to potentially expand perceptions and access to teaching and learning resources and explore some of the affordances of social media (Facebook, WhatsApp groups) in learning in an informal context. The results will be used to inform policy-level decisions on what shape could CPD of all teachers take in the context of a developing country like Pakistan.Keywords: CPD, teaching & learning, blended learning, learning technologies
Procedia PDF Downloads 8425358 A Data-Driven Optimal Control Model for the Dynamics of Monkeypox in a Variable Population with a Comprehensive Cost-Effectiveness Analysis
Authors: Martins Onyekwelu Onuorah, Jnr Dahiru Usman
Abstract:
Introduction: In the realm of public health, the threat posed by Monkeypox continues to elicit concern, prompting rigorous studies to understand its dynamics and devise effective containment strategies. Particularly significant is its recurrence in variable populations, such as the observed outbreak in Nigeria in 2022. In light of this, our study undertakes a meticulous analysis, employing a data-driven approach to explore, validate, and propose optimized intervention strategies tailored to the distinct dynamics of Monkeypox within varying demographic structures. Utilizing a deterministic mathematical model, we delved into the intricate dynamics of Monkeypox, with a particular focus on a variable population context. Our qualitative analysis provided insights into the disease-free equilibrium, revealing its stability when R0 is less than one and discounting the possibility of backward bifurcation, as substantiated by the presence of a single stable endemic equilibrium. The model was rigorously validated using real-time data from the Nigerian 2022 recorded cases for Epi weeks 1 – 52. Transitioning from qualitative to quantitative, we augmented our deterministic model with optimal control, introducing three time-dependent interventions to scrutinize their efficacy and influence on the epidemic's trajectory. Numerical simulations unveiled a pronounced impact of the interventions, offering a data-supported blueprint for informed decision-making in containing the disease. A comprehensive cost-effectiveness analysis employing the Infection Averted Ratio (IAR), Average Cost-Effectiveness Ratio (ACER), and Incremental Cost-Effectiveness Ratio (ICER) facilitated a balanced evaluation of the interventions’ economic and health impacts. In essence, our study epitomizes a holistic approach to understanding and mitigating Monkeypox, intertwining rigorous mathematical modeling, empirical validation, and economic evaluation. The insights derived not only bolster our comprehension of Monkeypox's intricate dynamics but also unveil optimized, cost-effective interventions. This integration of methodologies and findings underscores a pivotal stride towards aligning public health imperatives with economic sustainability, marking a significant contribution to global efforts in combating infectious diseases.Keywords: monkeypox, equilibrium states, stability, bifurcation, optimal control, cost-effectiveness
Procedia PDF Downloads 8625357 Scientific Recommender Systems Based on Neural Topic Model
Authors: Smail Boussaadi, Hassina Aliane
Abstract:
With the rapid growth of scientific literature, it is becoming increasingly challenging for researchers to keep up with the latest findings in their fields. Academic, professional networks play an essential role in connecting researchers and disseminating knowledge. To improve the user experience within these networks, we need effective article recommendation systems that provide personalized content.Current recommendation systems often rely on collaborative filtering or content-based techniques. However, these methods have limitations, such as the cold start problem and difficulty in capturing semantic relationships between articles. To overcome these challenges, we propose a new approach that combines BERTopic (Bidirectional Encoder Representations from Transformers), a state-of-the-art topic modeling technique, with community detection algorithms in a academic, professional network. Experiences confirm our performance expectations by showing good relevance and objectivity in the results.Keywords: scientific articles, community detection, academic social network, recommender systems, neural topic model
Procedia PDF Downloads 9725356 A Basic Metric Model: Foundation for an Evidence-Based HRM System
Authors: K. M. Anusha, R. Krishnaveni
Abstract:
Crossing a decade of the 21st century, the paradigm of human resources can be seen evolving with the strategic gene induced into it. There seems to be a radical shift descending as the corporate sector calls on its HR team to become strategic rather than administrative. This transferal eventually requires the metrics employed by these HR teams not to be just operationally reactive but to be aligned to an evidence-based strategic thinking. Realizing the growing need for a prescriptive metric model for effective HR analytics, this study has designed a conceptual framework for a basic metric model that can assist IT-HRM professionals to transition to a practice of evidence-based decision-making to enhance organizational performance.Keywords: metric model, evidence based HR, HR analytics, strategic HR practices, IT sector
Procedia PDF Downloads 40325355 A Fully Coupled Thermo-Hydraulic Mechanical Elastoplastic Damage Constitutive Model for Porous Fractured Medium during CO₂ Injection
Authors: Nikolaos Reppas, Yilin Gui
Abstract:
A dual-porosity finite element-code will be presented for the stability analysis of the wellbore during CO₂ injection. An elastoplastic damage response will be considered to the model. The Finite Element Method (FEM) will be validated using experimental results from literature or from experiments that are planned to be undertaken at Newcastle University. The main target of the research paper is to present a constitutive model that can help industries to safely store CO₂ in geological rock formations and forecast any changes on the surrounding rock of the wellbore. The fully coupled elastoplastic damage Thermo-Hydraulic-Mechanical (THM) model will determine the pressure and temperature of the injected CO₂ as well as the size of the radius of the wellbore that can make the Carbon Capture and Storage (CCS) procedure more efficient.Keywords: carbon capture and storage, Wellbore stability, elastoplastic damage response for rock, constitutive THM model, fully coupled thermo-hydraulic-mechanical model
Procedia PDF Downloads 17425354 Model Updating Based on Modal Parameters Using Hybrid Pattern Search Technique
Authors: N. Guo, C. Xu, Z. C. Yang
Abstract:
In order to ensure the high reliability of an aircraft, the accurate structural dynamics analysis has become an indispensable part in the design of an aircraft structure. Therefore, the structural finite element model which can be used to accurately calculate the structural dynamics and their transfer relations is the prerequisite in structural dynamic design. A dynamic finite element model updating method is presented to correct the uncertain parameters of the finite element model of a structure using measured modal parameters. The coordinate modal assurance criterion is used to evaluate the correlation level at each coordinate over the experimental and the analytical mode shapes. Then, the weighted summation of the natural frequency residual and the coordinate modal assurance criterion residual is used as the objective function. Moreover, the hybrid pattern search (HPS) optimization technique, which synthesizes the advantages of pattern search (PS) optimization technique and genetic algorithm (GA), is introduced to solve the dynamic FE model updating problem. A numerical simulation and a model updating experiment for GARTEUR aircraft model are performed to validate the feasibility and effectiveness of the present dynamic model updating method, respectively. The updated results show that the proposed method can be successfully used to modify the incorrect parameters with good robustness.Keywords: model updating, modal parameter, coordinate modal assurance criterion, hybrid genetic/pattern search
Procedia PDF Downloads 16125353 Efficiency of the Slovak Commercial Banks Applying the DEA Window Analysis
Authors: Iveta Řepková
Abstract:
The aim of this paper is to estimate the efficiency of the Slovak commercial banks employing the Data Envelopment Analysis (DEA) window analysis approach during the period 2003-2012. The research is based on unbalanced panel data of the Slovak commercial banks. Undesirable output was included into analysis of banking efficiency. It was found that most efficient banks were Postovabanka, UniCredit Bank and Istrobanka in CCR model and the most efficient banks were Slovenskasporitelna, Istrobanka and UniCredit Bank in BCC model. On contrary, the lowest efficient banks were found Privatbanka and CitiBank. We found that the largest banks in the Slovak banking market were lower efficient than medium-size and small banks. Results of the paper is that during the period 2003-2008 the average efficiency was increasing and then during the period 2010-2011 the average efficiency decreased as a result of financial crisis.Keywords: data envelopment analysis, efficiency, Slovak banking sector, window analysis
Procedia PDF Downloads 35725352 Analytical Response Characterization of High Mobility Transistor Channels
Authors: F. Z. Mahi, H. Marinchio, C. Palermo, L. Varani
Abstract:
We propose an analytical approach for the admittance response calculation of the high mobility InGaAs channel transistors. The development of the small-signal admittance takes into account the longitudinal and transverse electric fields through a pseudo two-dimensional approximation of the Poisson equation. The total currents and the potentials matrix relation between the gate and the drain terminals determine the frequency-dependent small-signal admittance response. The analytical results show that the admittance spectrum exhibits a series of resonant peaks corresponding to the excitation of plasma waves. The appearance of the resonance is discussed and analyzed as functions of the channel length and the temperature. The model can be used, on one hand, to control the appearance of plasma resonances, and on the other hand, can give significant information about the admittance phase frequency dependence.Keywords: small-signal admittance, Poisson equation, currents and potentials matrix, the drain and the gate terminals, analytical model
Procedia PDF Downloads 54025351 Process Modeling in an Aeronautics Context
Authors: Sophie Lemoussu, Jean-Charles Chaudemar, Robertus A. Vingerhoeds
Abstract:
Many innovative projects exist in the field of aeronautics, each addressing specific areas so to reduce weight, increase autonomy, reduction of CO2, etc. In many cases, such innovative developments are being carried out by very small enterprises (VSE’s) or small and medium sized-enterprises (SME’s). A good example concerns airships that are being studied as a real alternative to passenger and cargo transportation. Today, no international regulations propose a precise and sufficiently detailed framework for the development and certification of airships. The absence of such a regulatory framework requires a very close contact with regulatory instances. However, VSE’s/SME’s do not always have sufficient resources and internal knowledge to handle this complexity and to discuss these issues. This poses an additional challenge for those VSE’s/SME’s, in particular those that have system integration responsibilities and that must provide all the necessary evidence to demonstrate their ability to design, produce, and operate airships with the expected level of safety and reliability. The main objective of this research is to provide a methodological framework enabling VSE’s/SME’s with limited resources to organize the development of airships while taking into account the constraints of safety, cost, time and performance. This paper proposes to provide a contribution to this problematic by proposing a Model-Based Systems Engineering approach. Through a comprehensive process modeling approach applied to the development processes, the regulatory constraints, existing best practices, etc., a good image can be obtained as to the process landscape that may influence the development of airships. To this effect, not only the necessary regulatory information is taken on board, also other international standards and norms on systems engineering and project management are being modeled and taken into account. In a next step, the model can be used for analysis of the specific situation for given developments, derive critical paths for the development, identify eventual conflicting aspects between the norms, standards, and regulatory expectations, or also identify those areas where not enough information is available. Once critical paths are known, optimization approaches can be used and decision support techniques can be applied so to better support VSE’s/SME’s in their innovative developments. This paper reports on the adopted modeling approach, the retained modeling languages, and how they all fit together.Keywords: aeronautics, certification, process modeling, project management, regulation, SME, systems engineering, VSE
Procedia PDF Downloads 16125350 Risk Prioritization in Tunneling Construction Projects
Authors: David Nantes, George Gilbert
Abstract:
There are a lot of risks that might crop up as a tunneling project develops, and it's crucial to be aware of them. Due to the unexpected nature of tunneling projects and the interconnectedness of risk occurrences, the risk assessment approach presents a significant challenge. The purpose of this study is to provide a hybrid FDEMATEL-ANP model to help prioritize risks during tunnel construction projects. The ambiguity in expert judgments and the relative severity of interdependencies across risk occurrences are both taken into consideration by this model, thanks to the Fuzzy Decision-Making Trial and Evaluation Laboratory (FDEMATEL). The Analytic Network Process (ANP) method is used to rank priorities and assess project risks. The authors provide a case study of a subway tunneling construction project to back up the validity of their methodology. The results showed that the proposed method successfully isolated key risk factors and elucidated their interplay in the case study. The proposed method has the potential to become a helpful resource for evaluating dangers associated with tunnel construction projects.Keywords: risk, prioritization, FDEMATEL, ANP, tunneling construction projects
Procedia PDF Downloads 9225349 New Dynamic Constitutive Model for OFHC Copper Film
Authors: Jin Sung Kim, Hoon Huh
Abstract:
The material properties of OFHC copper film was investigated with the High-Speed Material Micro Testing Machine (HSMMTM) at the high strain rates. The rate-dependent stress-strain curves from the experiment and the Johnson-Cook curve fitting showed large discrepancies as the plastic strain increases since the constitutive model implies no rate-dependent strain hardening effect. A new constitutive model was proposed in consideration of rate-dependent strain hardening effect. The strain rate hardening term in the new constitutive model consists of the strain rate sensitivity coefficients of the yield strength and strain hardening.Keywords: rate dependent material properties, dynamic constitutive model, OFHC copper film, strain rate
Procedia PDF Downloads 48625348 Exploring the Situational Approach to Decision Making: User eConsent on a Health Social Network
Authors: W. Rowan, Y. O’Connor, L. Lynch, C. Heavin
Abstract:
Situation Awareness can offer the potential for conscious dynamic reflection. In an era of online health data sharing, it is becoming increasingly important that users of health social networks (HSNs) have the information necessary to make informed decisions as part of the registration process and in the provision of eConsent. This research aims to leverage an adapted Situation Awareness (SA) model to explore users’ decision making processes in the provision of eConsent. A HSN platform was used to investigate these behaviours. A mixed methods approach was taken. This involved the observation of registration behaviours followed by a questionnaire and focus group/s. Early results suggest that users are apt to automatically accept eConsent, and only later consider the long-term implications of sharing their personal health information. Further steps are required to continue developing knowledge and understanding of this important eConsent process. The next step in this research will be to develop a set of guidelines for the improved presentation of eConsent on the HSN platform.Keywords: eConsent, health social network, mixed methods, situation awareness
Procedia PDF Downloads 29225347 Low-Cost, Portable Optical Sensor with Regression Algorithm Models for Accurate Monitoring of Nitrites in Environments
Authors: David X. Dong, Qingming Zhang, Meng Lu
Abstract:
Nitrites enter waterways as runoff from croplands and are discharged from many industrial sites. Excessive nitrite inputs to water bodies lead to eutrophication. On-site rapid detection of nitrite is of increasing interest for managing fertilizer application and monitoring water source quality. Existing methods for detecting nitrites use spectrophotometry, ion chromatography, electrochemical sensors, ion-selective electrodes, chemiluminescence, and colorimetric methods. However, these methods either suffer from high cost or provide low measurement accuracy due to their poor selectivity to nitrites. Therefore, it is desired to develop an accurate and economical method to monitor nitrites in environments. We report a low-cost optical sensor, in conjunction with a machine learning (ML) approach to enable high-accuracy detection of nitrites in water sources. The sensor works under the principle of measuring molecular absorptions of nitrites at three narrowband wavelengths (295 nm, 310 nm, and 357 nm) in the ultraviolet (UV) region. These wavelengths are chosen because they have relatively high sensitivity to nitrites; low-cost light-emitting devices (LEDs) and photodetectors are also available at these wavelengths. A regression model is built, trained, and utilized to minimize cross-sensitivities of these wavelengths to the same analyte, thus achieving precise and reliable measurements with various interference ions. The measured absorbance data is input to the trained model that can provide nitrite concentration prediction for the sample. The sensor is built with i) a miniature quartz cuvette as the test cell that contains a liquid sample under test, ii) three low-cost UV LEDs placed on one side of the cell as light sources, with each LED providing a narrowband light, and iii) a photodetector with a built-in amplifier and an analog-to-digital converter placed on the other side of the test cell to measure the power of transmitted light. This simple optical design allows measuring the absorbance data of the sample at the three wavelengths. To train the regression model, absorbances of nitrite ions and their combination with various interference ions are first obtained at the three UV wavelengths using a conventional spectrophotometer. Then, the spectrophotometric data are inputs to different regression algorithm models for training and evaluating high-accuracy nitrite concentration prediction. Our experimental results show that the proposed approach enables instantaneous nitrite detection within several seconds. The sensor hardware costs about one hundred dollars, which is much cheaper than a commercial spectrophotometer. The ML algorithm helps to reduce the average relative errors to below 3.5% over a concentration range from 0.1 ppm to 100 ppm of nitrites. The sensor has been validated to measure nitrites at three sites in Ames, Iowa, USA. This work demonstrates an economical and effective approach to the rapid, reagent-free determination of nitrites with high accuracy. The integration of the low-cost optical sensor and ML data processing can find a wide range of applications in environmental monitoring and management.Keywords: optical sensor, regression model, nitrites, water quality
Procedia PDF Downloads 7225346 Improving the Quantification Model of Internal Control Impact on Banking Risks
Authors: M. Ndaw, G. Mendy, S. Ouya
Abstract:
Risk management in banking sector is a key issue linked to financial system stability and its importance has been elevated by technological developments and emergence of new financial instruments. In this paper, we improve the model previously defined for quantifying internal control impact on banking risks by automatizing the residual criticality estimation step of FMECA. For this, we defined three equations and a maturity coefficient to obtain a mathematical model which is tested on all banking processes and type of risks. The new model allows an optimal assessment of residual criticality and improves the correlation rate that has become 98%.Keywords: risk, control, banking, FMECA, criticality
Procedia PDF Downloads 33425345 Formulation of Extended-Release Gliclazide Tablet Using a Mathematical Model for Estimation of Hypromellose
Authors: Farzad Khajavi, Farzaneh Jalilfar, Faranak Jafari, Leila Shokrani
Abstract:
Formulation of gliclazide in the form of extended-release tablet in 30 and 60 mg dosage forms was performed using hypromellose (HPMC K4M) as a retarding agent. Drug-release profiles were investigated in comparison with references Diamicron MR 30 and 60 mg tablets. The effect of size of powder particles, the amount of hypromellose in formulation, hardness of tablets, and also the effect of halving the tablets were investigated on drug release profile. A mathematical model which describes hypromellose behavior in initial times of drug release was proposed for the estimation of hypromellose content in modified-release gliclazide 60 mg tablet. This model is based on erosion of hypromellose in dissolution media. The model is applicable to describe release profiles of insoluble drugs. Therefore, by using dissolved amount of drug in initial times of dissolution and the model, the amount of hypromellose in formulation can be predictable. The model was used to predict the HPMC K4M content in modified-release gliclazide 30 mg and extended-release quetiapine 200 mg tablets.Keywords: Gliclazide, hypromellose, drug release, modified-release tablet, mathematical model
Procedia PDF Downloads 22325344 Elemental Graph Data Model: A Semantic and Topological Representation of Building Elements
Authors: Yasmeen A. S. Essawy, Khaled Nassar
Abstract:
With the rapid increase of complexity in the building industry, professionals in the A/E/C industry were forced to adopt Building Information Modeling (BIM) in order to enhance the communication between the different project stakeholders throughout the project life cycle and create a semantic object-oriented building model that can support geometric-topological analysis of building elements during design and construction. This paper presents a model that extracts topological relationships and geometrical properties of building elements from an existing fully designed BIM, and maps this information into a directed acyclic Elemental Graph Data Model (EGDM). The model incorporates BIM-based search algorithms for automatic deduction of geometrical data and topological relationships for each building element type. Using graph search algorithms, such as Depth First Search (DFS) and topological sortings, all possible construction sequences can be generated and compared against production and construction rules to generate an optimized construction sequence and its associated schedule. The model is implemented in a C# platform.Keywords: building information modeling (BIM), elemental graph data model (EGDM), geometric and topological data models, graph theory
Procedia PDF Downloads 38225343 Forming Simulation of Thermoplastic Pre-Impregnated Textile Composite
Authors: Masato Nishi, Tetsushi Kaburagi, Masashi Kurose, Tei Hirashima, Tetsusei Kurasiki
Abstract:
The process of thermoforming a carbon fiber reinforced thermoplastic (CFRTP) has increased its presence in the automotive industry for its wide applicability to the mass production car. A non-isothermal forming for CFRTP can shorten its cycle time to less than 1 minute. In this paper, the textile reinforcement FE model which the authors proposed in a previous work is extended to the CFRTP model for non-isothermal forming simulation. The effect of thermoplastic is given by adding shell elements which consider thermal effect to the textile reinforcement model. By applying Reuss model to the stress calculation of thermoplastic, the proposed model can accurately predict in-plane shear behavior, which is the key deformation mode during forming, in the range of the process temperature. Using the proposed model, thermoforming simulation was conducted and the results are in good agreement with the experimental results.Keywords: carbon fiber reinforced thermoplastic, finite element analysis, pre-impregnated textile composite, non-isothermal forming
Procedia PDF Downloads 42925342 Bayesian Flexibility Modelling of the Conditional Autoregressive Prior in a Disease Mapping Model
Authors: Davies Obaromi, Qin Yongsong, James Ndege, Azeez Adeboye, Akinwumi Odeyemi
Abstract:
The basic model usually used in disease mapping, is the Besag, York and Mollie (BYM) model and which combines the spatially structured and spatially unstructured priors as random effects. Bayesian Conditional Autoregressive (CAR) model is a disease mapping method that is commonly used for smoothening the relative risk of any disease as used in the Besag, York and Mollie (BYM) model. This model (CAR), which is also usually assigned as a prior to one of the spatial random effects in the BYM model, successfully uses information from adjacent sites to improve estimates for individual sites. To our knowledge, there are some unrealistic or counter-intuitive consequences on the posterior covariance matrix of the CAR prior for the spatial random effects. In the conventional BYM (Besag, York and Mollie) model, the spatially structured and the unstructured random components cannot be seen independently, and which challenges the prior definitions for the hyperparameters of the two random effects. Therefore, the main objective of this study is to construct and utilize an extended Bayesian spatial CAR model for studying tuberculosis patterns in the Eastern Cape Province of South Africa, and then compare for flexibility with some existing CAR models. The results of the study revealed the flexibility and robustness of this alternative extended CAR to the commonly used CAR models by comparison, using the deviance information criteria. The extended Bayesian spatial CAR model is proved to be a useful and robust tool for disease modeling and as a prior for the structured spatial random effects because of the inclusion of an extra hyperparameter.Keywords: Besag2, CAR models, disease mapping, INLA, spatial models
Procedia PDF Downloads 28125341 An Inviscid Compressible Flow Solver Based on Unstructured OpenFOAM Mesh Format
Authors: Utkan Caliskan
Abstract:
Two types of numerical codes based on finite volume method are developed in order to solve compressible Euler equations to simulate the flow through forward facing step channel. Both algorithms have AUSM+- up (Advection Upstream Splitting Method) scheme for flux splitting and two-stage Runge-Kutta scheme for time stepping. In this study, the flux calculations differentiate between the algorithm based on OpenFOAM mesh format which is called 'face-based' algorithm and the basic algorithm which is called 'element-based' algorithm. The face-based algorithm avoids redundant flux computations and also is more flexible with hybrid grids. Moreover, some of OpenFOAM’s preprocessing utilities can be used on the mesh. Parallelization of the face based algorithm for which atomic operations are needed due to the shared memory model, is also presented. For several mesh sizes, 2.13x speed up is obtained with face-based approach over the element-based approach.Keywords: cell centered finite volume method, compressible Euler equations, OpenFOAM mesh format, OpenMP
Procedia PDF Downloads 31925340 Numerical Investigation of Wave Interaction with Double Vertical Slotted Walls
Authors: H. Ahmed, A. Schlenkhoff
Abstract:
Recently, permeable breakwaters have been suggested to overcome the disadvantages of fully protection breakwaters. These protection structures have minor impacts on the coastal environment and neighboring beaches where they provide a more economical protection from waves and currents. For regular waves, a numerical model is used (FLOW-3D, VOF) to investigate the hydraulic performance of a permeable breakwater. The model of permeable breakwater consists of a pair of identical vertical slotted walls with an impermeable upper and lower part, where the draft is a decimal multiple of the total depth. The middle part is permeable with a porosity of 50%. The second barrier is located at distant of 0.5 and 1.5 of the water depth from the first one. The numerical model is validated by comparisons with previous laboratory data and semi-analytical results of the same model. A good agreement between the numerical results and both laboratory data and semi-analytical results has been shown and the results indicate the applicability of the numerical model to reproduce most of the important features of the interaction. Through the numerical investigation, the friction factor of the model is carefully discussed.Keywords: coastal structures, permeable breakwater, slotted wall, numerical model, energy dissipation coefficient
Procedia PDF Downloads 39125339 Numerical Solution of a Mathematical Model of Vortex Using Projection Method: Applications to Tornado Dynamics
Authors: Jagdish Prasad Maurya, Sanjay Kumar Pandey
Abstract:
Inadequate understanding of the complex nature of flow features in tornado vortex is a major problem in modelling tornadoes. Tornadoes are violent atmospheric phenomenon that appear all over the world. Modelling tornadoes aim to reduce the loss of the human lives and material damage caused by the tornadoes. Dynamics of tornado is investigated by a numerical technique, the improved version of the projection method. In this paper, authors solve the problem for axisymmetric tornado vortex by the said method that uses a finite difference approach for getting an accurate and stable solution. The conclusions drawn are that large radial inflow velocity occurs near the ground that leads to increase the tangential velocity. The increased velocity phenomenon occurs close to the boundary and absolute maximum wind is obtained near the vortex core. The results validate previous numerical and theoretical models.Keywords: computational fluid dynamics, mathematical model, Navier-Stokes equations, tornado
Procedia PDF Downloads 35325338 Geographical Information System-Based Approach for Vertical Takeoff and Landing Takeoff and Landing Site Selection
Authors: Chamnan Kumsap, Somsarit Sinnung, Suriyawate Boonthalarath, Teeranai Srithamarong
Abstract:
This research paper addresses the GIS analysis approach to the investigation of suitable sites for a vertical takeoff and landing drone. The study manipulated GIS and terrain layers into a proper input before the spatial analysis that included slope, reclassify, classify, and buffer was applied to the individual layers. The output layers were weighted, and multi-criteria analyzed before those patches failing to comply with filtering out criteria were discarded. Field survey for each suitable candidate site was conducted to cross-check the proposed approach with the real world. Conclusion was extracted for the VTOL takeoff and landing sites, and discussion was provided with further study being suggested on the mission simulation of selected takeoff and landing sites.Keywords: GIS approach, site selection, VTOL, takeoff and landing
Procedia PDF Downloads 10425337 Aggregate Supply Response of Some Livestock Commodities in Algeria: Cointegration- Vector Error Correction Model Approach
Authors: Amine M. Benmehaia, Amine Oulmane
Abstract:
The supply response of agricultural commodities to changes in price incentives is an important issue for the success of any policy reform in the agricultural sector. This study aims to quantify the responsiveness of producers of some livestock commodities to price incentives in Algerian context. Time series analysis is used on annual data for a period of 52 years (1966-2018). Both co-integration and vector error correction model (VECM) are used through the Nerlove model of partial adjustment. The study attempts to determine the long-run and short-run relationships along with the magnitudes of disequilibria in the selected commodities. Results show that the short-run price elasticities are low in cow and sheep meat sectors (8.7 and 8% respectively), while their respective long-run elasticities are 16.5 and 10.5, whereas eggs and milk have very high short-run price elasticities (82 and 90% respectively) with long-run elasticities of 40 and 46 respectively. The error correction coefficient, reflecting the speed of adjustment towards the long-run equilibrium, is statistically significant and have the expected negative sign. Its estimates are 12.7 for cow meat, 33.5 for sheep meat, 46.7 for eggs and 8.4 for milk. It seems that cow meat and milk producers have a weak feedback of about 12.7% and 8.4% respectively of the previous year's disequilibrium from the long-run price elasticity, whereas sheep meat and eggs producers adjust to correct long run disequilibrium with a high speed of adjustment (33.5% and 46.7 % respectively). The implication of this is that much more in-depth research is needed to identify those factors that affect agricultural supply and to describe the effect of factors that shift supply in response to price incentives. This could provide valuable information for government in the use of appropriate policy measures.Keywords: Algeria, cointegration, livestock, supply response, vector error correction model
Procedia PDF Downloads 14125336 A New Approach to Interval Matrices and Applications
Authors: Obaid Algahtani
Abstract:
An interval may be defined as a convex combination as follows: I=[a,b]={x_α=(1-α)a+αb: α∈[0,1]}. Consequently, we may adopt interval operations by applying the scalar operation point-wise to the corresponding interval points: I ∙J={x_α∙y_α ∶ αϵ[0,1],x_α ϵI ,y_α ϵJ}, With the usual restriction 0∉J if ∙ = ÷. These operations are associative: I+( J+K)=(I+J)+ K, I*( J*K)=( I*J )* K. These two properties, which are missing in the usual interval operations, will enable the extension of the usual linear system concepts to the interval setting in a seamless manner. The arithmetic introduced here avoids such vague terms as ”interval extension”, ”inclusion function”, determinants which we encounter in the engineering literature that deal with interval linear systems. On the other hand, these definitions were motivated by our attempt to arrive at a definition of interval random variables and investigate the corresponding statistical properties. We feel that they are the natural ones to handle interval systems. We will enable the extension of many results from usual state space models to interval state space models. The interval state space model we will consider here is one of the form X_((t+1) )=AX_t+ W_t, Y_t=HX_t+ V_t, t≥0, where A∈ 〖IR〗^(k×k), H ∈ 〖IR〗^(p×k) are interval matrices and 〖W 〗_t ∈ 〖IR〗^k,V_t ∈〖IR〗^p are zero – mean Gaussian white-noise interval processes. This feeling is reassured by the numerical results we obtained in a simulation examples.Keywords: interval analysis, interval matrices, state space model, Kalman Filter
Procedia PDF Downloads 42525335 Using an Epidemiological Model to Study the Spread of Misinformation during the Black Lives Matter Movement
Authors: Maryam Maleki, Esther Mead, Mohammad Arani, Nitin Agarwal
Abstract:
The proliferation of social media platforms like Twitter has heightened the consequences of the spread of misinformation. To understand and model the spread of misinformation, in this paper, we leveraged the SEIZ (Susceptible, Exposed, Infected, Skeptics) epidemiological model to describe the underlying process that delineates the spread of misinformation on Twitter. Compared to the other epidemiological models, this model produces broader results because it includes the additional Skeptics (Z) compartment, wherein a user may be Exposed to an item of misinformation but not engage in any reaction to it, and the additional Exposed (E) compartment, wherein the user may need some time before deciding to spread a misinformation item. We analyzed misinformation regarding the unrest in Washington, D.C. in the month of March 2020, which was propagated by the use of the #DCblackout hashtag by different users across the U.S. on Twitter. Our analysis shows that misinformation can be modeled using the concept of epidemiology. To the best of our knowledge, this research is the first to attempt to apply the SEIZ epidemiological model to the spread of a specific item of misinformation, which is a category distinct from that of rumor and hoax on online social media platforms. Applying a mathematical model can help to understand the trends and dynamics of the spread of misinformation on Twitter and ultimately help to develop techniques to quickly identify and control it.Keywords: Black Lives Matter, epidemiological model, mathematical modeling, misinformation, SEIZ model, Twitter
Procedia PDF Downloads 16725334 Determining a Sustainability Business Model Using Materiality Matrices in an Electricity Bus Factory
Authors: Ozcan Yavas, Berrak Erol Nalbur, Sermin Gunarslan
Abstract:
A materiality matrix is a tool that organizations use to prioritize their activities and adapt to the increasing sustainability requirements in recent years. For the materiality index to move from business models to the sustainability business model stage, it must be done with all partners in the raw material, supply, production, product, and end-of-life product stages. Within the scope of this study, the Materiality Matrix was used to transform the business model into a sustainability business model and to create a sustainability roadmap in a factory producing electric buses. This matrix determines the necessary roadmap for all stakeholders to participate in the process, especially in sectors that produce sustainable products, such as the electric vehicle sector, and to act together with the cradle-to-cradle approach of sustainability roadmaps. Global Reporting Initiative analysis was used in the study conducted with 1150 stakeholders within the scope of the study, and 43 questions were asked to the stakeholders under the main headings of 'Legal Compliance Level,' 'Environmental Strategies,' 'Risk Management Activities,' 'Impact of Sustainability Activities on Products and Services,' 'Corporate Culture,' 'Responsible and Profitable Business Model Practices' and 'Achievements in Leading the Sector' and Economic, Governance, Environment, Social and Other. The results of the study aimed to include five 1st priority issues and four 2nd priority issues in the sustainability strategies of the organization in the short and medium term. When the studies carried out in the short term are evaluated in terms of Sustainability and Environmental Risk Management, it is seen that the studies are still limited to the level of legal legislation (60%) and individual studies in line with the strategies (20%). At the same time, the stakeholders expect the company to integrate sustainability activities into its business model within five years (35%) and to carry out projects to become the first company that comes to mind with its success leading the sector (20%). Another result obtained within the study's scope is identifying barriers to implementation. It is seen that the most critical obstacles identified by stakeholders with climate change and environmental impacts are financial deficiency and lack of infrastructure in the dissemination of sustainable products. These studies are critical for transitioning to sustainable business models for the electric vehicle sector to achieve the EU Green Deal and CBAM targets.Keywords: sustainability business model, materiality matrix, electricity bus, carbon neutrality, sustainability management
Procedia PDF Downloads 6125333 Developing a DNN Model for the Production of Biogas From a Hybrid BO-TPE System in an Anaerobic Wastewater Treatment Plant
Authors: Hadjer Sadoune, Liza Lamini, Scherazade Krim, Amel Djouadi, Rachida Rihani
Abstract:
Deep neural networks are highly regarded for their accuracy in predicting intricate fermentation processes. Their ability to learn from a large amount of datasets through artificial intelligence makes them particularly effective models. The primary obstacle in improving the performance of these models is to carefully choose the suitable hyperparameters, including the neural network architecture (number of hidden layers and hidden units), activation function, optimizer, learning rate, and other relevant factors. This study predicts biogas production from real wastewater treatment plant data using a sophisticated approach: hybrid Bayesian optimization with a tree-structured Parzen estimator (BO-TPE) for an optimised deep neural network (DNN) model. The plant utilizes an Upflow Anaerobic Sludge Blanket (UASB) digester that treats industrial wastewater from soft drinks and breweries. The digester has a working volume of 1574 m3 and a total volume of 1914 m3. Its internal diameter and height were 19 and 7.14 m, respectively. The data preprocessing was conducted with meticulous attention to preserving data quality while avoiding data reduction. Three normalization techniques were applied to the pre-processed data (MinMaxScaler, RobustScaler and StandardScaler) and compared with the Non-Normalized data. The RobustScaler approach has strong predictive ability for estimating the volume of biogas produced. The highest predicted biogas volume was 2236.105 Nm³/d, with coefficient of determination (R2), mean absolute error (MAE), and root mean square error (RMSE) values of 0.712, 164.610, and 223.429, respectively.Keywords: anaerobic digestion, biogas production, deep neural network, hybrid bo-tpe, hyperparameters tuning
Procedia PDF Downloads 3825332 Sparsity Order Selection and Denoising in Compressed Sensing Framework
Authors: Mahdi Shamsi, Tohid Yousefi Rezaii, Siavash Eftekharifar
Abstract:
Compressed sensing (CS) is a new powerful mathematical theory concentrating on sparse signals which is widely used in signal processing. The main idea is to sense sparse signals by far fewer measurements than the Nyquist sampling rate, but the reconstruction process becomes nonlinear and more complicated. Common dilemma in sparse signal recovery in CS is the lack of knowledge about sparsity order of the signal, which can be viewed as model order selection procedure. In this paper, we address the problem of sparsity order estimation in sparse signal recovery. This is of main interest in situations where the signal sparsity is unknown or the signal to be recovered is approximately sparse. It is shown that the proposed method also leads to some kind of signal denoising, where the observations are contaminated with noise. Finally, the performance of the proposed approach is evaluated in different scenarios and compared to an existing method, which shows the effectiveness of the proposed method in terms of order selection as well as denoising.Keywords: compressed sensing, data denoising, model order selection, sparse representation
Procedia PDF Downloads 48325331 Three-Stage Least Squared Models of a Station-Level Subway Ridership: Incorporating an Analysis on Integrated Transit Network Topology Measures
Authors: Jungyeol Hong, Dongjoo Park
Abstract:
The urban transit system is a critical part of a solution to the economic, energy, and environmental challenges. Furthermore, it ultimately contributes the improvement of people’s quality of lives. For taking these kinds of advantages, the city of Seoul has tried to construct an integrated transit system including both subway and buses. The effort led to the fact that approximately 6.9 million citizens use the integrated transit system every day for their trips. Diagnosing the current transit network is a significant task to provide more convenient and pleasant transit environment. Therefore, the critical objective of this study is to establish a methodological framework for the analysis of an integrated bus-subway network and to examine the relationship between subway ridership and parameters such as network topology measures, bus demand, and a variety of commercial business facilities. Regarding a statistical approach to estimate subway ridership at a station level, many previous studies relied on Ordinary Least Square regression, but there was lack of studies considering the endogeneity issues which might show in the subway ridership prediction model. This study focused on both discovering the impacts of integrated transit network topology measures and endogenous effect of bus demand on subway ridership. It could ultimately contribute to developing more accurate subway ridership estimation accounting for its statistical bias. The spatial scope of the study covers Seoul city in South Korea, and it includes 243 subway stations and 10,120 bus stops with the temporal scope set during twenty-four hours with one-hour interval time panels each. The subway and bus ridership information in detail was collected from the Seoul Smart Card data in 2015 and 2016. First, integrated subway-bus network topology measures which have characteristics regarding connectivity, centrality, transitivity, and reciprocity were estimated based on the complex network theory. The results of integrated transit network topology analysis were compared to subway-only network topology. Also, the non-recursive approach which is Three-Stage Least Square was applied to develop the daily subway ridership model as capturing the endogeneity between bus and subway demands. Independent variables included roadway geometry, commercial business characteristics, social-economic characteristics, safety index, transit facility attributes, and dummies for seasons and time zone. Consequently, it was found that network topology measures were significant size effect. Especially, centrality measures showed that the elasticity was a change of 4.88% for closeness centrality, 24.48% for betweenness centrality while the elasticity of bus ridership was 8.85%. Moreover, it was proved that bus demand and subway ridership were endogenous in a non-recursive manner as showing that predicted bus ridership and predicted subway ridership is statistically significant in OLS regression models. Therefore, it shows that three-stage least square model appears to be a plausible model for efficient subway ridership estimation. It is expected that the proposed approach provides a reliable guideline that can be used as part of the spectrum of tools for evaluating a city-wide integrated transit network.Keywords: integrated transit system, network topology measures, three-stage least squared, endogeneity, subway ridership
Procedia PDF Downloads 177