Search results for: validated model
16537 Computational Models for Accurate Estimation of Joint Forces
Authors: Ibrahim Elnour Abdelrahman Eltayeb
Abstract:
Computational modelling is a method used to investigate joint forces during a movement. It can get high accuracy in the joint forces via subject-specific models. However, the construction of subject-specific models remains time-consuming and expensive. The purpose of this paper was to identify what alterations we can make to generic computational models to get a better estimation of the joint forces. It appraised the impact of these alterations on the accuracy of the estimated joint forces. It found different strategies of alterations: joint model, muscle model, and an optimisation problem. All these alterations affected joint contact force accuracy, so showing the potential for improving the model predictions without involving costly and time-consuming medical images.Keywords: joint force, joint model, optimisation problem, validation
Procedia PDF Downloads 17016536 Investigating the Effects of Thermal and Surface Energy on the Two-Dimensional Flow Characteristics of Oil in Water Mixture between Two Parallel Plates: A Lattice Boltzmann Method Study
Abstract:
A hybrid quasi-steady thermal lattice Boltzmann model was used to study the combined effects of temperature and contact angle on the movement of slugs and droplets of oil in water (O/W) system flowing between two parallel plates. The model static contact angle due to the deposition of the O/W droplet on a flat surface with simulated hydrophilic characteristic at different fluid temperatures, matched very well the proposed theoretical calculation. Furthermore, the model was used to simulate the dynamic behavior of droplets and slugs deposited on the domain’s upper and lower surfaces, while subjected to parabolic flow conditions. The model accurately simulated the contact angle hysteresis for the dynamic droplets cases. It was also shown that at elevated temperatures the required power to transport the mixture diminished remarkably.Keywords: lattice Boltzmann method, Gunstensen model, thermal, contact angle, high viscosity ratio
Procedia PDF Downloads 37016535 Spherical Harmonic Based Monostatic Anisotropic Point Scatterer Model for RADAR Applications
Authors: Eric Huang, Coleman DeLude, Justin Romberg, Saibal Mukhopadhyay, Madhavan Swaminathan
Abstract:
High performance computing (HPC) based emulators can be used to model the scattering from multiple stationary and moving targets for RADAR applications. These emulators rely on the RADAR Cross Section (RCS) of the targets being available in complex scenarios. Representing the RCS using tables generated from electromagnetic (EM) simulations is often times cumbersome leading to large storage requirement. This paper proposed a spherical harmonic based anisotropic scatterer model to represent the RCS of complex targets. The problem of finding the locations and reflection profiles of all scatterers can be formulated as a linear least square problem with a special sparsity constraint. This paper solves this problem using a modified Orthogonal Matching Pursuit algorithm. The results show that the spherical harmonic based scatterer model can effectively represent the RCS data of complex targets.Keywords: RADAR, RCS, high performance computing, point scatterer model
Procedia PDF Downloads 19116534 Estimating Anthropometric Dimensions for Saudi Males Using Artificial Neural Networks
Authors: Waleed Basuliman
Abstract:
Anthropometric dimensions are considered one of the important factors when designing human-machine systems. In this study, the estimation of anthropometric dimensions has been improved by using Artificial Neural Network (ANN) model that is able to predict the anthropometric measurements of Saudi males in Riyadh City. A total of 1427 Saudi males aged 6 to 60 years participated in measuring 20 anthropometric dimensions. These anthropometric measurements are considered important for designing the work and life applications in Saudi Arabia. The data were collected during eight months from different locations in Riyadh City. Five of these dimensions were used as predictors variables (inputs) of the model, and the remaining 15 dimensions were set to be the measured variables (Model’s outcomes). The hidden layers varied during the structuring stage, and the best performance was achieved with the network structure 6-25-15. The results showed that the developed Neural Network model was able to estimate the body dimensions of Saudi male population in Riyadh City. The network's mean absolute percentage error (MAPE) and the root mean squared error (RMSE) were found to be 0.0348 and 3.225, respectively. These results were found less, and then better, than the errors found in the literature. Finally, the accuracy of the developed neural network was evaluated by comparing the predicted outcomes with regression model. The ANN model showed higher coefficient of determination (R2) between the predicted and actual dimensions than the regression model.Keywords: artificial neural network, anthropometric measurements, back-propagation
Procedia PDF Downloads 48716533 Dynamic Model for Forecasting Rainfall Induced Landslides
Authors: R. Premasiri, W. A. H. A. Abeygunasekara, S. M. Hewavidana, T. Jananthan, R. M. S. Madawala, K. Vaheeshan
Abstract:
Forecasting the potential for disastrous events such as landslides has become one of the major necessities in the current world. Most of all, the landslides occurred in Sri Lanka are found to be triggered mostly by intense rainfall events. The study area is the landslide near Gerandiella waterfall which is located by the 41st kilometer post on Nuwara Eliya-Gampala main road in Kotmale Division in Sri Lanka. The landslide endangers the entire Kotmale town beneath the slope. Geographic Information System (GIS) platform is very much useful when it comes to the need of emulating the real-world processes. The models are used in a wide array of applications ranging from simple evaluations to the levels of forecast future events. This project investigates the possibility of developing a dynamic model to map the spatial distribution of the slope stability. The model incorporates several theoretical models including the infinite slope model, Green Ampt infiltration model and Perched ground water flow model. A series of rainfall values can be fed to the model as the main input to simulate the dynamics of slope stability. Hydrological model developed using GIS is used to quantify the perched water table height, which is one of the most critical parameters affecting the slope stability. Infinite slope stability model is used to quantify the degree of slope stability in terms of factor of safety. DEM was built with the use of digitized contour data. Stratigraphy was modeled in Surfer using borehole data and resistivity images. Data available from rainfall gauges and piezometers were used in calibrating the model. During the calibration, the parameters were adjusted until a good fit between the simulated ground water levels and the piezometer readings was obtained. This model equipped with the predicted rainfall values can be used to forecast of the slope dynamics of the area of interest. Therefore it can be investigated the slope stability of rainfall induced landslides by adjusting temporal dimensions.Keywords: factor of safety, geographic information system, hydrological model, slope stability
Procedia PDF Downloads 42316532 A New Verification Based Congestion Control Scheme in Mobile Networks
Authors: P. K. Guha Thakurta, Shouvik Roy, Bhawana Raj
Abstract:
A congestion control scheme in mobile networks is proposed in this paper through a verification based model. The model proposed in this work is represented through performance metric like buffer Occupancy, latency and packet loss rate. Based on pre-defined values, each of the metric is introduced in terms of three different states. A Markov chain based model for the proposed work is introduced to monitor the occurrence of the corresponding state transitions. Thus, the estimation of the network status is obtained in terms of performance metric. In addition, the improved performance of our proposed model over existing works is shown with experimental results.Keywords: congestion, mobile networks, buffer, delay, call drop, markov chain
Procedia PDF Downloads 44116531 A New Mathematical Model for Scheduling Preventive Maintenance and Renewal Projects of Multi-Unit Systems; Application to Railway Track
Authors: Farzad Pargar
Abstract:
We introduce the preventive maintenance and renewal scheduling problem for a multi-unit system over a finite and discretized time horizon. Given the latest possible time for carrying out the next maintenance and renewal projects after the previous ones and considering several common set-up costs, the introduced scheduling model tries to minimize the cost of projects by grouping them and simultaneously finding the optimal balance between doing maintenance and renewal. We present a 0-1 pure integer linear programming that determines which projects should be performed together on which location and in which period (e.g., week or month). We consider railway track as a case for our study and test the performance of the proposed model on a set of test problems. The experimental results show that the proposed approach performs well.Keywords: maintenance, renewal, scheduling, mathematical programming model
Procedia PDF Downloads 68816530 WiFi Data Offloading: Bundling Method in a Canvas Business Model
Authors: Majid Mokhtarnia, Alireza Amini
Abstract:
Mobile operators deal with increasing in the data traffic as a critical issue. As a result, a vital responsibility of the operators is to deal with such a trend in order to create added values. This paper addresses a bundling method in a Canvas business model in a WiFi Data Offloading (WDO) strategy by which some elements of the model may be affected. In the proposed method, it is supposed to sell a number of data packages for subscribers in which there are some packages with a free given volume of data-offloaded WiFi complimentary. The paper on hands analyses this method in the views of attractiveness and profitability. The results demonstrate that the quality of implementation of the WDO strongly affects the final result and helps the decision maker to make the best one.Keywords: bundling, canvas business model, telecommunication, WiFi data offloading
Procedia PDF Downloads 20016529 The Effect of CPU Location in Total Immersion of Microelectronics
Authors: A. Almaneea, N. Kapur, J. L. Summers, H. M. Thompson
Abstract:
Meeting the growth in demand for digital services such as social media, telecommunications, and business and cloud services requires large scale data centres, which has led to an increase in their end use energy demand. Generally, over 30% of data centre power is consumed by the necessary cooling overhead. Thus energy can be reduced by improving the cooling efficiency. Air and liquid can both be used as cooling media for the data centre. Traditional data centre cooling systems use air, however liquid is recognised as a promising method that can handle the more densely packed data centres. Liquid cooling can be classified into three methods; rack heat exchanger, on-chip heat exchanger and full immersion of the microelectronics. This study quantifies the improvements of heat transfer specifically for the case of immersed microelectronics by varying the CPU and heat sink location. Immersion of the server is achieved by filling the gap between the microelectronics and a water jacket with a dielectric liquid which convects the heat from the CPU to the water jacket on the opposite side. Heat transfer is governed by two physical mechanisms, which is natural convection for the fixed enclosure filled with dielectric liquid and forced convection for the water that is pumped through the water jacket. The model in this study is validated with published numerical and experimental work and shows good agreement with previous work. The results show that the heat transfer performance and Nusselt number (Nu) is improved by 89% by placing the CPU and heat sink on the bottom of the microelectronics enclosure.Keywords: CPU location, data centre cooling, heat sink in enclosures, immersed microelectronics, turbulent natural convection in enclosures
Procedia PDF Downloads 27216528 Critical Success Factors Quality Requirement Change Management
Authors: Jamshed Ahmad, Abdul Wahid Khan, Javed Ali Khan
Abstract:
Managing software quality requirements change management is a difficult task in the field of software engineering. Avoiding incoming changes result in user dissatisfaction while accommodating to many requirement changes may delay product delivery. Poor requirements management is solely considered the primary cause of the software failure. It becomes more challenging in global software outsourcing. Addressing success factors in quality requirement change management is desired today due to the frequent change requests from the end-users. In this research study, success factors are recognized and scrutinized with the help of a systematic literature review (SLR). In total, 16 success factors were identified, which significantly impacted software quality requirement change management. The findings show that Proper Requirement Change Management, Rapid Delivery, Quality Software Product, Access to Market, Project Management, Skills and Methodologies, Low Cost/Effort Estimation, Clear Plan and Road Map, Agile Processes, Low Labor Cost, User Satisfaction, Communication/Close Coordination, Proper Scheduling and Time Constraints, Frequent Technological Changes, Robust Model, Geographical distribution/Cultural differences are the key factors that influence software quality requirement change. The recognized success factors and validated with the help of various research methods, i.e., case studies, interviews, surveys and experiments. These factors are then scrutinized in continents, database, company size and period of time. Based on these findings, requirement change will be implemented in a better way.Keywords: global software development, requirement engineering, systematic literature review, success factors
Procedia PDF Downloads 19716527 Modeling and Design of a Solar Thermal Open Volumetric Air Receiver
Authors: Piyush Sharma, Laltu Chandra, P. S. Ghoshdastidar, Rajiv Shekhar
Abstract:
Metals processing operations such as melting and heat treatment of metals are energy-intensive, requiring temperatures greater than 500oC. The desired temperature in these industrial furnaces is attained by circulating electrically-heated air. In most of these furnaces, electricity produced from captive coal-based thermal power plants is used. Solar thermal energy could be a viable heat source in these furnaces. A retrofitted solar convective furnace (SCF) concept, which uses solar thermal generated hot air, has been proposed. Critical to the success of a SCF is the design of an open volumetric air receiver (OVAR), which can heat air in excess of 800oC. The OVAR is placed on top of a tower and receives concentrated solar radiation from a heliostat field. Absorbers, mixer assembly, and the return air flow chamber (RAFC) are the major components of an OVAR. The absorber is a porous structure that transfers heat from concentrated solar radiation to ambient air, referred to as primary air. The mixer ensures uniform air temperature at the receiver exit. Flow of the relatively cooler return air in the RAFC ensures that the absorbers do not fail by overheating. In an earlier publication, the detailed design basis, fabrication, and characterization of a 2 kWth open volumetric air receiver (OVAR) based laboratory solar air tower simulator was presented. Development of an experimentally-validated, CFD based mathematical model which can ultimately be used for the design and scale-up of an OVAR has been the major objective of this investigation. In contrast to the published literature, where flow and heat transfer have been modeled primarily in a single absorber module, the present study has modeled the entire receiver assembly, including the RAFC. Flow and heat transfer calculations have been carried out in ANSYS using the LTNE model. The complex return air flow pattern in the RAFC requires complicated meshes and is computational and time intensive. Hence a simple, realistic 1-D mathematical model, which circumvents the need for carrying out detailed flow and heat transfer calculations, has also been proposed. Several important results have emerged from this investigation. Circumferential electrical heating of absorbers can mimic frontal heating by concentrated solar radiation reasonably well in testing and characterizing the performance of an OVAR. Circumferential heating, therefore, obviates the need for expensive high solar concentration simulators. Predictions suggest that the ratio of power on aperture (POA) and mass flow rate of air (MFR) is a normalizing parameter for characterizing the thermal performance of an OVAR. Increasing POA/MFR increases the maximum temperature of air, but decreases the thermal efficiency of an OVAR. Predictions of the 1-D mathematical are within 5% of ANSYS predictions and computation time is reduced from ~ 5 hours to a few seconds.Keywords: absorbers, mixer assembly, open volumetric air receiver, return air flow chamber, solar thermal energy
Procedia PDF Downloads 19716526 Personal Characteristics and Personality Traits as Predictors of Compassion Fatigue among Counselors from Dominican Schools in the Philippines
Authors: Neil Jordan M. Uy, Fe Pelilia V. Hernandez
Abstract:
A counselor is always regarded as a professional who embodies the willingness to help others through the process of counseling. He is knowledgeable and skillful of the different theories, tools, and techniques that are useful in aiding the client to cope with their dilemmas. The negative experiences of the clients that are shared during the counseling session can affect the professional counselor. Compassion fatigue, a professional impairment, is characterized by the decline of one’s productivity and the feeling of anxiety and stress brought about as the counselor empathizes, listens, and cares for others. This descriptive type of research aimed to explore variables that are predictors of compassion fatigue utilizing three research instruments; Demographic Profile Sheet, Professional Quality of Life Scale, and Neo-Pi-R. The 52 respondents of this study were counselors from the different Dominican schools in the Philippines. Generally, the counselors have low level of compassion fatigue across personal characteristics (age, gender, years of service, highest educational attainment, and professional status) and personality traits (extraversion, agreeableness, conscientiousness, openness, and neuroticism). ANOVA validated the findings of this that among the personal characteristics and personality traits, extraversion with f-value of 3.944 and p-value of 0.026, and conscientiousness, with f-value of 4.125 and p-value of 0.022 were found to have significant difference in the level of compassion fatigue. A very significant difference was observed with neuroticism with f-value of 6.878 and p-value 0.002. Among the personal characteristics and personal characteristics, only neuroticism was found to predict compassion fatigue. The computed r2 value of 0.204 using multiple regression analysis suggests that 20.4 percent of compassion fatigue can be predicted by neuroticism. The predicting power of neuroticism can be computed from the regression model Y=0.156x+26.464; where x is the number of neuroticism.Keywords: big five personality traits, compassion fatigue, counselors, professional quality of life scale
Procedia PDF Downloads 37816525 A Small Signal Model for Resonant Tunneling Diode
Authors: Rania M. Abdallah, Ahmed A. S. Dessouki, Moustafa H. Aly
Abstract:
This paper has presented a new simple small signal model for a resonant tunnelling diode device. The resonant tunnelling diode equivalent circuit elements were calculated and the results led to good agreement between the calculated equivalent circuit elements and the measurement results.Keywords: resonant tunnelling diode, small signal model, negative differential conductance, electronic engineering
Procedia PDF Downloads 44316524 Simulation of Red Blood Cells in Complex Micro-Tubes
Authors: Ting Ye, Nhan Phan-Thien, Chwee Teck Lim, Lina Peng, Huixin Shi
Abstract:
In biofluid flow systems, often the flow problems of fluids of complex structures, such as the flow of red blood cells (RBCs) through complex capillary vessels, need to be considered. In this paper, we aim to apply a particle-based method, Smoothed Dissipative Particle Dynamics (SDPD), to simulate the motion and deformation of RBCs in complex micro-tubes. We first present the theoretical models, including SDPD model, RBC-fluid interaction model, RBC deformation model, RBC aggregation model, and boundary treatment model. After that, we show the verification and validation of these models, by comparing our numerical results with the theoretical, experimental and previously-published numerical results. Finally, we provide some simulation cases, such as the motion and deformation of RBCs in rectangular, cylinder, curved, bifurcated, and constricted micro-tubes, respectively.Keywords: aggregation, deformation, red blood cell, smoothed dissipative particle dynamics
Procedia PDF Downloads 17416523 Students’ learning Effects in Physical Education between Sport Education Model with TPSR and Traditional Teaching Model with TPSR
Authors: Yi-Hsiang Pan, Chen-Hui Huang, Ching-Hsiang Chen, Wei-Ting Hsu
Abstract:
The purposes of the study were to explore the students' learning effect of physical education curriculum between merging Teaching Personal and Social Responsibility (TPSR) with sport education model and TPSR with traditional teaching model, which these learning effects included sport self-efficacy, sport enthusiastic, group cohesion, responsibility and game performance. The participants include 3 high school physical education teachers and 6 physical education classes, 133 participants with experience group 75 students and control group 58 students, and each teacher taught an experimental group and a control group for 16 weeks. The research methods used questionnaire investigation, interview, focus group meeting. The research instruments included personal and social responsibility questionnaire, sport enthusiastic scale, group cohesion scale, sport self-efficacy scale and game performance assessment instrument. Multivariate Analysis of covariance and Repeated measure ANOVA were used to test difference of students' learning effects between merging TPSR with sport education model and TPSR with traditional teaching model. The findings of research were: 1) The sport education model with TPSR could improve students' learning effects, including sport self-efficacy, game performance, sport enthusiastic, group cohesion and responsibility. 2) The traditional teaching model with TPSR could improve students' learning effect, including sport self-efficacy, responsibility and game performance. 3) the sport education model with TPSR could improve more learning effects than traditional teaching model with TPSR, including sport self-efficacy, sport enthusiastic,responsibility and game performance. 4) Based on qualitative data about learning experience of teachers and students, sport education model with TPSR significant improve learning motivation, group interaction and game sense. The conclusions indicated sport education model with TPSR could improve more learning effects in physical education curriculum. On other hand, the curricular projects of hybrid TPSR-Sport Education model and TPSR-Traditional Teaching model are both good curricular projects of moral character education, which may be applied in school physical education.Keywords: character education, sport season, game performance, sport competence
Procedia PDF Downloads 45216522 Simulation of the Large Hadrons Collisions Using Monte Carlo Tools
Authors: E. Al Daoud
Abstract:
In many cases, theoretical treatments are available for models for which there is no perfect physical realization. In this situation, the only possible test for an approximate theoretical solution is to compare with data generated from a computer simulation. In this paper, Monte Carlo tools are used to study and compare the elementary particles models. All the experiments are implemented using 10000 events, and the simulated energy is 13 TeV. The mean and the curves of several variables are calculated for each model using MadAnalysis 5. Anomalies in the results can be seen in the muons masses of the minimal supersymmetric standard model and the two Higgs doublet model.Keywords: Feynman rules, hadrons, Lagrangian, Monte Carlo, simulation
Procedia PDF Downloads 31916521 A Hybrid Particle Swarm Optimization-Nelder- Mead Algorithm (PSO-NM) for Nelson-Siegel- Svensson Calibration
Authors: Sofia Ayouche, Rachid Ellaia, Rajae Aboulaich
Abstract:
Today, insurers may use the yield curve as an indicator evaluation of the profit or the performance of their portfolios; therefore, they modeled it by one class of model that has the ability to fit and forecast the future term structure of interest rates. This class of model is the Nelson-Siegel-Svensson model. Unfortunately, many authors have reported a lot of difficulties when they want to calibrate the model because the optimization problem is not convex and has multiple local optima. In this context, we implement a hybrid Particle Swarm optimization and Nelder Mead algorithm in order to minimize by least squares method, the difference between the zero-coupon curve and the NSS curve.Keywords: optimization, zero-coupon curve, Nelson-Siegel-Svensson, particle swarm optimization, Nelder-Mead algorithm
Procedia PDF Downloads 43016520 Let-7 Mirnas Regulate Inflammatory Cytokine Production in Bovine Endometrial Cells after Lipopolysaccharide Challenge by Targeting TNFα
Authors: S. Ibrahim, D. Salilew-Wondim, M. Hoelker, C. Looft, E. Tholen, C. Grosse-Brinkhaus, K. Schellander, C. Neuhoff, D. Tesfaye
Abstract:
Bovine endometrial cells appear to have a key role in innate immune defense of the female genital tract. A better understanding of molecular changes in microRNAs (miRNAs) and their target genes expression may identify reliable prognostic indicators for cows that will resolve inflammation and resume cyclicity. In the current study, we hypothesized that let-7 miRNAs family has a primary role in the innate immune defence of the endometrium tissue against bacterial infection, which is partly achieved via regulating mRNA stability of pro-inflammatory cytokines at the post-transcriptional level. Therefore, we conducted two experiments. In the first experiment, primary bovine endometrial cells were challenged with clinical (3.0 μg/ml) and sub-clinical (0.5 μg/ml) doses of lipopolysaccharide (LPS) for 24h. In the 2nd experiment, we have investigated the potential role of let-7 miRNAs (let-7a and let-7f) using gain and loss of function approaches. Additionally, tumor necrosis factor alpha (TNFα), transforming growth factor beta 1 induced transcript 1 (TGFB1I1) and serum deprivation response (SDPR) genes were validated using reporter assay. Here we addressed for the first time that let-7 miRNAs have a precise role in bovine endometrium, where LPS dysregulated let-7 miRNAs family expression was associated with an increased pro-inflammatory cytokine level by directly/indirectly targeting the TNFα, interleukin 6 (IL6), nuclear factor kappa-light-chain enhancer of activated B cells (NFκB), TGFβ1I1 and SDPR genes. To our knowledge, this is the first study showing that TNFα, TGFβ1I1 and SDPR were identified and validated as novel let-7 miRNAs targets and could have a distinct role in inflammatory immune response of LPS challenged bovine endometrial cells. Our data represent a new finding by which uterine homeostasis is maintained through functional regulation of let-7a by down-regulation of pro-inflammatory cytokines expression (TNFα and IL6) at the mRNA and protein levels. These findings suggest that LPS serves as a negative regulator of let-7 miRNAs expression and provides a mechanism for the persistent pro-inflammatory phenotype, which is a hallmark of bovine subclinical endometritis.Keywords: bovine endometrial cells, let-7, lipopolysaccharide, pro-inflammatory cytokines
Procedia PDF Downloads 38016519 Uncertainty Quantification of Crack Widths and Crack Spacing in Reinforced Concrete
Authors: Marcel Meinhardt, Manfred Keuser, Thomas Braml
Abstract:
Cracking of reinforced concrete is a complex phenomenon induced by direct loads or restraints affecting reinforced concrete structures as soon as the tensile strength of the concrete is exceeded. Hence it is important to predict where cracks will be located and how they will propagate. The bond theory and the crack formulas in the actual design codes, for example, DIN EN 1992-1-1, are all based on the assumption that the reinforcement bars are embedded in homogeneous concrete without taking into account the influence of transverse reinforcement and the real stress situation. However, it can often be observed that real structures such as walls, slabs or beams show a crack spacing that is orientated to the transverse reinforcement bars or to the stirrups. In most Finite Element Analysis studies, the smeared crack approach is used for crack prediction. The disadvantage of this model is that the typical strain localization of a crack on element level can’t be seen. The crack propagation in concrete is a discontinuous process characterized by different factors such as the initial random distribution of defects or the scatter of material properties. Such behavior presupposes the elaboration of adequate models and methods of simulation because traditional mechanical approaches deal mainly with average material parameters. This paper concerned with the modelling of the initiation and the propagation of cracks in reinforced concrete structures considering the influence of transverse reinforcement and the real stress distribution in reinforced concrete (R/C) beams/plates in bending action. Therefore, a parameter study was carried out to investigate: (I) the influence of the transversal reinforcement to the stress distribution in concrete in bending mode and (II) the crack initiation in dependence of the diameter and distance of the transversal reinforcement to each other. The numerical investigations on the crack initiation and propagation were carried out with a 2D reinforced concrete structure subjected to quasi static loading and given boundary conditions. To model the uncertainty in the tensile strength of concrete in the Finite Element Analysis correlated normally and lognormally distributed random filed with different correlation lengths were generated. The paper also presents and discuss different methods to generate random fields, e.g. the Covariance Matrix Decomposition Method. For all computations, a plastic constitutive law with softening was used to model the crack initiation and the damage of the concrete in tension. It was found that the distributions of crack spacing and crack widths are highly dependent of the used random field. These distributions are validated to experimental studies on R/C panels which were carried out at the Laboratory for Structural Engineering at the University of the German Armed Forces in Munich. Also, a recommendation for parameters of the random field for realistic modelling the uncertainty of the tensile strength is given. The aim of this research was to show a method in which the localization of strains and cracks as well as the influence of transverse reinforcement on the crack initiation and propagation in Finite Element Analysis can be seen.Keywords: crack initiation, crack modelling, crack propagation, cracks, numerical simulation, random fields, reinforced concrete, stochastic
Procedia PDF Downloads 15716518 Comparison of Volume of Fluid Model: Experimental and Empirical Results for Flows over Stacked Drop Manholes
Authors: Ramin Mansouri
Abstract:
The manhole is one of the types of structures that are installed at the site of change direction or change in the pipe diameter or sewage pipes as well as in step slope areas to reduce the flow velocity. In this study, the flow characteristics of hydraulic structures in a manhole structure have been investigated with a numerical model. In this research, the types of computational grid coarse, medium, and fines have been used for simulation. In order to simulate flow, k-ε model (standard, RNG, Realizable) and k-w model (standard SST) are used. Also, in order to find the best wall conditions, two types of standard and non-equilibrium wall functions were investigated. The turbulent model k-ε has the highest correlation with experimental results or all models. In terms of boundary conditions, constant speed is set for the flow input boundary, the output pressure is set in the boundaries which are in contact with the air, and the standard wall function is used for the effect of the wall function. In the numerical model, the depth at the output of the second manhole is estimated to be less than that of the laboratory and the output jet from the span. In the second regime, the jet flow collides with the manhole wall and divides into two parts, so hydraulic characteristics are the same as large vertical shaft hydraulic characteristics. In this situation, the turbulence is in a high range since it can be seen more energy loss in it. According to the results, energy loss in numerical is estimated at 9.359%, which is more than experimental data.Keywords: manhole, energy, depreciation, turbulence model, wall function, flow
Procedia PDF Downloads 8216517 Static Properties of Ge and Sr Isotopes in the Cluster Model
Authors: Mohammad Reza Shojaei, Mahdeih Mirzaeinia
Abstract:
We have studied the cluster structure of even-even stable isotopes of Ge and Sr. The Schrodinger equation has been solved using the generalized parametric Nikiforov-Uvarov method with a phenomenological potential. This potential is the sum of the attractive Yukawa-like potential, a Manning-Rosen-type potential, and the repulsive Yukawa potential for interaction between the cluster and the core. We have shown that the available experimental data of the first rotational band energies can be well described by assuming a binary system of the α cluster and the core and using an analytical solution. Our results were consistent with experimental values. Hence, this model can be applied to study the other even-even isotopesKeywords: cluser model, NU method, ge and Sr, potential central
Procedia PDF Downloads 7616516 Conceptual Model for Logistics Information System
Authors: Ana María Rojas Chaparro, Cristian Camilo Sarmiento Chaves
Abstract:
Given the growing importance of logistics as a discipline for efficient management of materials flow and information, the adoption of tools that permit to create facilities in making decisions based on a global perspective of the system studied has been essential. The article shows how from a concepts-based model is possible to organize and represent in appropriate way the reality, showing accurate and timely information, features that make this kind of models an ideal component to support an information system, recognizing that information as relevant to establish particularities that allow get a better performance about the evaluated sector.Keywords: system, information, conceptual model, logistics
Procedia PDF Downloads 49716515 Automatic Flood Prediction Using Rainfall Runoff Model in Moravian-Silesian Region
Authors: B. Sir, M. Podhoranyi, S. Kuchar, T. Kocyan
Abstract:
Rainfall-runoff models play important role in hydrological predictions. However, the model is only one part of the process for creation of flood prediction. The aim of this paper is to show the process of successful prediction for flood event (May 15–May 18 2014). The prediction was performed by rainfall runoff model HEC–HMS, one of the models computed within Floreon+ system. The paper briefly evaluates the results of automatic hydrologic prediction on the river Olše catchment and its gages Český Těšín and Věřňovice.Keywords: flood, HEC-HMS, prediction, rainfall, runoff
Procedia PDF Downloads 39516514 A West Coast Estuarine Case Study: A Predictive Approach to Monitor Estuarine Eutrophication
Authors: Vedant Janapaty
Abstract:
Estuaries are wetlands where fresh water from streams mixes with salt water from the sea. Also known as “kidneys of our planet”- they are extremely productive environments that filter pollutants, absorb floods from sea level rise, and shelter a unique ecosystem. However, eutrophication and loss of native species are ailing our wetlands. There is a lack of uniform data collection and sparse research on correlations between satellite data and in situ measurements. Remote sensing (RS) has shown great promise in environmental monitoring. This project attempts to use satellite data and correlate metrics with in situ observations collected at five estuaries. Images for satellite data were processed to calculate 7 bands (SIs) using Python. Average SI values were calculated per month for 23 years. Publicly available data from 6 sites at ELK was used to obtain 10 parameters (OPs). Average OP values were calculated per month for 23 years. Linear correlations between the 7 SIs and 10 OPs were made and found to be inadequate (correlation = 1 to 64%). Fourier transform analysis on 7 SIs was performed. Dominant frequencies and amplitudes were extracted for 7 SIs, and a machine learning(ML) model was trained, validated, and tested for 10 OPs. Better correlations were observed between SIs and OPs, with certain time delays (0, 3, 4, 6 month delay), and ML was again performed. The OPs saw improved R² values in the range of 0.2 to 0.93. This approach can be used to get periodic analyses of overall wetland health with satellite indices. It proves that remote sensing can be used to develop correlations with critical parameters that measure eutrophication in situ data and can be used by practitioners to easily monitor wetland health.Keywords: estuary, remote sensing, machine learning, Fourier transform
Procedia PDF Downloads 10416513 Implementation of IWA-ASM1 Model for Simulating the Wastewater Treatment Plant of Beja by GPS-X 5.1
Authors: Fezzani Boubaker
Abstract:
The modified activated sludge model (ASM1 or Mantis) is a generic structured model and a common platform for dynamic simulation of varieties of aerobic processes for optimization and upgrading of existing plants and for new facilities design. In this study, the modified ASM1 included in the GPS-X software was used to simulate the wastewater treatment plant (WWTP) of Beja treating domestic sewage mixed with baker‘s yeast factory effluent. The results of daily measurements and operating records were used to calibrate the model. A sensitivity and an automatic optimization analysis were conducted to determine the most sensitive and optimal parameters. The results indicated that the ASM1 model could simulate with good accuracy: the COD concentration of effluents from the WWTP of Beja for all months of the year 2012. In addition, it prevents the disruption observed at the output of the plant by injecting the baker‘s yeast factory effluent at high concentrations varied between 20 and 80 g/l.Keywords: ASM1, activated sludge, baker’s yeast effluent, modelling, simulation, GPS-X 5.1 software
Procedia PDF Downloads 34316512 Multivariate Analysis on Water Quality Attributes Using Master-Slave Neural Network Model
Authors: A. Clementking, C. Jothi Venkateswaran
Abstract:
Mathematical and computational functionalities such as descriptive mining, optimization, and predictions are espoused to resolve natural resource planning. The water quality prediction and its attributes influence determinations are adopted optimization techniques. The water properties are tainted while merging water resource one with another. This work aimed to predict influencing water resource distribution connectivity in accordance to water quality and sediment using an innovative proposed master-slave neural network back-propagation model. The experiment results are arrived through collecting water quality attributes, computation of water quality index, design and development of neural network model to determine water quality and sediment, master–slave back propagation neural network back-propagation model to determine variations on water quality and sediment attributes between the water resources and the recommendation for connectivity. The homogeneous and parallel biochemical reactions are influences water quality and sediment while distributing water from one location to another. Therefore, an innovative master-slave neural network model [M (9:9:2)::S(9:9:2)] designed and developed to predict the attribute variations. The result of training dataset given as an input to master model and its maximum weights are assigned as an input to the slave model to predict the water quality. The developed master-slave model is predicted physicochemical attributes weight variations for 85 % to 90% of water quality as a target values.The sediment level variations also predicated from 0.01 to 0.05% of each water quality percentage. The model produced the significant variations on physiochemical attribute weights. According to the predicated experimental weight variation on training data set, effective recommendations are made to connect different resources.Keywords: master-slave back propagation neural network model(MSBPNNM), water quality analysis, multivariate analysis, environmental mining
Procedia PDF Downloads 47716511 The Future of Insurance: P2P Innovation versus Traditional Business Model
Authors: Ivan Sosa Gomez
Abstract:
Digitalization has impacted the entire insurance value chain, and the growing movement towards P2P platforms and the collaborative economy is also beginning to have a significant impact. P2P insurance is defined as innovation, enabling policyholders to pool their capital, self-organize, and self-manage their own insurance. In this context, new InsurTech start-ups are emerging as peer-to-peer (P2P) providers, based on a model that differs from traditional insurance. As a result, although P2P platforms do not change the fundamental basis of insurance, they do enable potentially more efficient business models to be established in terms of ensuring the coverage of risk. It is therefore relevant to determine whether p2p innovation can have substantial effects on the future of the insurance sector. For this purpose, it is considered necessary to develop P2P innovation from a business perspective, as well as to build a comparison between a traditional model and a P2P model from an actuarial perspective. Objectives: The objectives are (1) to represent P2P innovation in the business model compared to the traditional insurance model and (2) to establish a comparison between a traditional model and a P2P model from an actuarial perspective. Methodology: The research design is defined as action research in terms of understanding and solving the problems of a collectivity linked to an environment, applying theory and best practices according to the approach. For this purpose, the study is carried out through the participatory variant, which involves the collaboration of the participants, given that in this design, participants are considered experts. For this purpose, prolonged immersion in the field is carried out as the main instrument for data collection. Finally, an actuarial model is developed relating to the calculation of premiums that allows for the establishment of projections of future scenarios and the generation of conclusions between the two models. Main Contributions: From an actuarial and business perspective, we aim to contribute by developing a comparison of the two models in the coverage of risk in order to determine whether P2P innovation can have substantial effects on the future of the insurance sector.Keywords: Insurtech, innovation, business model, P2P, insurance
Procedia PDF Downloads 9216510 Modelling Enablers of Service Using ISM: Implications for Quality Improvements in Healthcare Sector of UAE
Authors: Flevy Lasrado
Abstract:
Purpose: The purpose of this paper is to show the relationship between the service quality dimensions and model them to propose quality improvements using interpretive structural modelling (ISM). Methodology: This paper used an interpretive structural modelling (ISM). The data was collected from the expert opinions that included a questionnaire. The detailed method of using ISM is discussed in the paper. Findings: The present research work provides an ISM based model to understand the relationships among the service quality dimensions. Practical implications or Original Value: An ISM based model has been developed for healthcare facility for improving customer satisfaction and increasing market share. Although there is lot of research on SERVQUAL model adapted to healthcare sector, no study has been done to understand the interactions among these dimensions. So the major contribution of this research work is the development of contextual relationships among identified variables through a systematic framework. The present research work provides an ISM based model to understand the relationships among the service quality dimensions.Keywords: SERQUAL, healthcare, quality, service quality
Procedia PDF Downloads 40516509 Predicting Financial Distress in South Africa
Authors: Nikki Berrange, Gizelle Willows
Abstract:
Business rescue has become increasingly popular since its inclusion in the Companies Act of South Africa in May 2011. The Alternate Exchange (AltX) of the Johannesburg Stock Exchange has experienced a marked increase in the number of companies entering business rescue. This study sampled twenty companies listed on the AltX to determine whether Altman’s Z-score model for emerging markets (ZEM) or Taffler’s Z-score model is a more accurate model in predicting financial distress for small to medium size companies in South Africa. The study was performed over three different time horizons; one, two and three years prior to the event of financial distress, in order to determine how many companies each model predicted would be unlikely to succeed as well as the predictive ability and accuracy of the respective models. The study found that Taffler’s Z-score model had a greater ability at predicting financial distress from all three-time horizons.Keywords: Altman’s ZEM-score, Altman’s Z-score, AltX, business rescue, Taffler’s Z-score
Procedia PDF Downloads 37316508 Normalizing Logarithms of Realized Volatility in an ARFIMA Model
Authors: G. L. C. Yap
Abstract:
Modelling realized volatility with high-frequency returns is popular as it is an unbiased and efficient estimator of return volatility. A computationally simple model is fitting the logarithms of the realized volatilities with a fractionally integrated long-memory Gaussian process. The Gaussianity assumption simplifies the parameter estimation using the Whittle approximation. Nonetheless, this assumption may not be met in the finite samples and there may be a need to normalize the financial series. Based on the empirical indices S&P500 and DAX, this paper examines the performance of the linear volatility model pre-treated with normalization compared to its existing counterpart. The empirical results show that by including normalization as a pre-treatment procedure, the forecast performance outperforms the existing model in terms of statistical and economic evaluations.Keywords: Gaussian process, long-memory, normalization, value-at-risk, volatility, Whittle estimator
Procedia PDF Downloads 354