Search results for: prediction modelling
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3907

Search results for: prediction modelling

2467 The Future of the Architect's Profession in France with the Emergence of Building Information Modelling

Authors: L. Mercier, D. Beladjine, K. Beddiar

Abstract:

The digital transition of building in France brings many changes which some have been able to face very quickly, while others are struggling to find their place and the interest that BIM can bring in their profession. BIM today is already adopted or initiated by construction professionals. However, this change, which can be drastic for some, prevents them from integrating it definitively. This is the case with architects. The profession is shared on the practice of BIM in its exercise. The risk of not adopting this new working method now and of not wanting to switch to its new digital tools leads us to question the future of the profession in view of the gap that is likely to be created within project management. In order to deal with the subject efficiently, our work was based on a documentary watch on BIM and then on the profession of architect, which allowed us to establish links on these two subjects. The observation of the economic model towards which the agencies tend and the trend of the sought after profiles made it possible to develop the opportunities and the brakes likely to impact the future of the profession of architect. The centralization of research directs work towards the conclusion that the model implemented by companies does not allow to integrate BIM within their structure. A solution hypothesis was then issued, focusing on the development of agencies through the diversity of profiles, skills to be integrated internally with the aim of diversifying their skills, and their business practices. In order to address this hypothesis of a multidisciplinary agency model, we conducted a survey of architectural firms. It is built on the model of Anglo-Saxon countries, which do not have the same functioning in comparison to the French model. The results obtained showed a risk of gradual disappearance on the market from small agencies in favor of those who will have and could take this BIM working method. This is why the architectural profession must, first of all, look at what is happening within its training before absolutely wanting to diversify the profiles to integrate into its structure. This directs the study on the training of architects. The schools of French architects are generally behind schedule if we allow the comparison to the schools of engineers. The latter is currently experiencing a slight improvement with the emergence of masters and BIM options during the university course. If the training of architects develops towards learning BIM and the agencies have the desire to integrate different but complementary profiles, then they will develop their skills internally and therefore open their profession to new functions. The place of BIM Management on projects will allow the architect to remain in control of the project because of their overall vision of the project. In addition, the integration of BIM and more generally of the life cycle analysis of the structure will make it possible to guarantee eco-design or eco-construction by approaching the constraints of sustainable development omnipresent on the planet.

Keywords: building information modelling, BIM, BIM management, BIM manager, BIM architect

Procedia PDF Downloads 113
2466 Your First Step to Understanding Research Ethics: Psychoneurolinguistic Approach

Authors: Sadeq Al Yaari, Ayman Al Yaari, Adham Al Yaari, Montaha Al Yaari, Aayah Al Yaari, Sajedah Al Yaari

Abstract:

Objective: This research aims at investigating the research ethics in the field of science. Method: It is an exploratory research wherein the researchers attempted to cover the phenomenon at hand from all specialists’ viewpoints. Results Discussion is based upon the findings resulted from the analysis the researcher undertook. Concerning the results’ prediction, the researcher needs first to seek highly qualified people in the field of research as well as in the field of statistics who share the philosophy of the research. Then s/he should make sure that s/he is adequately trained in the specific techniques, methods and statically programs that are used at the study. S/he should also believe in continually analysis for the data in the most current methods.

Keywords: research ethics, legal, rights, psychoneurolinguistics

Procedia PDF Downloads 46
2465 Numerical Investigation of the Effect of the Spark Plug Gap on Engine-Like Conditions

Authors: Fernanda Pinheiro Martins, Pedro Teixeira Lacava

Abstract:

The objective of this research is to analyze the effects of different spark plug conditions in engine-like conditions by applying computational fluid dynamics analysis. The 3D models applied consist of 3-Zones Extended Coherent Flame (ECFM-3Z) and Imposed Stretch Spark Ignition Model (ISSIM), respectively, for the combustion and the spark plug modelling. For this study, it was applied direct injection fuel system in a single cylinder engine operating with E0. The application of realistic operating conditions (load and speed) to the different cases studied will provide a deeper understanding of the effects of the spark plug gap, a result of parts outwearing in most of the cases, to the development of the combustion in engine-like conditions.

Keywords: engine, CFD, direct injection, combustion, spark plug

Procedia PDF Downloads 131
2464 Brain Age Prediction Based on Brain Magnetic Resonance Imaging by 3D Convolutional Neural Network

Authors: Leila Keshavarz Afshar, Hedieh Sajedi

Abstract:

Estimation of biological brain age from MR images is a topic that has been much addressed in recent years due to the importance it attaches to early diagnosis of diseases such as Alzheimer's. In this paper, we use a 3D Convolutional Neural Network (CNN) to provide a method for estimating the biological age of the brain. The 3D-CNN model is trained by MRI data that has been normalized. In addition, to reduce computation while saving overall performance, some effectual slices are selected for age estimation. By this method, the biological age of individuals using selected normalized data was estimated with Mean Absolute Error (MAE) of 4.82 years.

Keywords: brain age estimation, biological age, 3D-CNN, deep learning, T1-weighted image, SPM, preprocessing, MRI, canny, gray matter

Procedia PDF Downloads 148
2463 Assessment of Pedestrian Comfort in a Portuguese City Using Computational Fluid Dynamics Modelling and Wind Tunnel

Authors: Bruno Vicente, Sandra Rafael, Vera Rodrigues, Sandra Sorte, Sara Silva, Ana Isabel Miranda, Carlos Borrego

Abstract:

Wind comfort for pedestrians is an important condition in urban areas. In Portugal, a country with 900 km of coastline, the wind direction are predominantly from Nor-Northwest with an average speed of 2.3 m·s -1 (at 2 m height). As a result, a set of city authorities have been requesting studies of pedestrian wind comfort for new urban areas/buildings, as well as to mitigate wind discomfort issues related to existing structures. This work covers the efficiency evaluation of a set of measures to reduce the wind speed in an outdoor auditorium (open space) located in a coastal Portuguese urban area. These measures include the construction of barriers, placed at upstream and downstream of the auditorium, and the planting of trees, placed upstream of the auditorium. The auditorium is constructed in the form of a porch, aligned with North direction, driving the wind flow within the auditorium, promoting channelling effects and increasing its speed, causing discomfort in the users of this structure. To perform the wind comfort assessment, two approaches were used: i) a set of experiments using the wind tunnel (physical approach), with a representative mock-up of the study area; ii) application of the CFD (Computational Fluid Dynamics) model VADIS (numerical approach). Both approaches were used to simulate the baseline scenario and the scenarios considering a set of measures. The physical approach was conducted through a quantitative method, using hot-wire anemometer, and through a qualitative analysis (visualizations), using the laser technology and a fog machine. Both numerical and physical approaches were performed for three different velocities (2, 4 and 6 m·s-1 ) and two different directions (NorNorthwest and South), corresponding to the prevailing wind speed and direction of the study area. The numerical results show an effective reduction (with a maximum value of 80%) of the wind speed inside the auditorium, through the application of the proposed measures. A wind speed reduction in a range of 20% to 40% was obtained around the audience area, for a wind direction from Nor-Northwest. For southern winds, in the audience zone, the wind speed was reduced from 60% to 80%. Despite of that, for southern winds, the design of the barriers generated additional hot spots (high wind speed), namely, in the entrance to the auditorium. Thus, a changing in the location of the entrance would minimize these effects. The results obtained in the wind tunnel compared well with the numerical data, also revealing the high efficiency of the purposed measures (for both wind directions).

Keywords: urban microclimate, pedestrian comfort, numerical modelling, wind tunnel experiments

Procedia PDF Downloads 232
2462 Solving Momentum and Energy Equation by Using Differential Transform Techniques

Authors: Mustafa Ekici

Abstract:

Natural convection is a basic process which is important in a wide variety of practical applications. In essence, a heated fluid expands and rises from buoyancy due to decreased density. Numerous papers have been written on natural or mixed convection in vertical ducts heated on the side. These equations have been proved to be valuable tools for the modelling of many phenomena such as fluid dynamics. Finding solutions to such equations or system of equations are in general not an easy task. We propose a method, which is called differential transform method, of solving a non-linear equations and compare the results with some of the other techniques. Illustrative examples shows that the results are in good agreement.

Keywords: differential transform method, momentum, energy equation, boundry value problem

Procedia PDF Downloads 463
2461 Strengthening by Assessment: A Case Study of Rail Bridges

Authors: Evangelos G. Ilias, Panagiotis G. Ilias, Vasileios T. Popotas

Abstract:

The United Kingdom has one of the oldest railway networks in the world dating back to 1825 when the world’s first passenger railway was opened. The network has some 40,000 bridges of various construction types using a wide range of materials including masonry, steel, cast iron, wrought iron, concrete and timber. It is commonly accepted that the successful operation of the network is vital for the economy of the United Kingdom, consequently the cost effective maintenance of the existing infrastructure is a high priority to maintain the operability of the network, prevent deterioration and to extend the life of the assets. Every bridge on the railway network is required to be assessed every eighteen years and a structured approach to assessments is adopted with three main types of progressively more detailed assessments used. These assessment types include Level 0 (standardized spreadsheet assessment tools), Level 1 (analytical hand calculations) and Level 2 (generally finite element analyses). There is a degree of conservatism in the first two types of assessment dictated to some extent by the relevant standards which can lead to some structures not achieving the required load rating. In these situations, a Level 2 Assessment is often carried out using finite element analysis to uncover ‘latent strength’ and improve the load rating. If successful, the more sophisticated analysis can save on costly strengthening or replacement works and avoid disruption to the operational railway. This paper presents the ‘strengthening by assessment’ achieved by Level 2 analyses. The use of more accurate analysis assumptions and the implementation of non-linear modelling and functions (material, geometric and support) to better understand buckling modes and the structural behaviour of historic construction details that are not specifically covered by assessment codes are outlined. Metallic bridges which are susceptible to loss of section size through corrosion have largest scope for improvement by the Level 2 Assessment methodology. Three case studies are presented, demonstrating the effectiveness of the sophisticated Level 2 Assessment methodology using finite element analysis against the conservative approaches employed for Level 0 and Level 1 Assessments. One rail overbridge and two rail underbridges that did not achieve the required load rating by means of a Level 1 Assessment due to the inadequate restraint provided by U-Frame action are examined and the increase in assessed capacity given by the Level 2 Assessment is outlined.

Keywords: assessment, bridges, buckling, finite element analysis, non-linear modelling, strengthening

Procedia PDF Downloads 311
2460 Numerical Flow Simulation around HSP Propeller in Open Water and behind a Vessel Wake Using RANS CFD Code

Authors: Kadda Boumediene, Mohamed Bouzit

Abstract:

The prediction of the flow around marine propellers and vessel hulls propeller interaction is one of the challenges of Computational fluid dynamics (CFD). The CFD has emerged as a potential tool in recent years and has promising applications. The objective of the current study is to predict the hydrodynamic performances of HSP marine propeller in open water and behind a vessel. The unsteady 3-D flow was modeled numerically along with respectively the K-ω standard and K-ω SST turbulence models for steady and unsteady cases. The hydrodynamic performances such us a torque and thrust coefficients and efficiency show good agreement with the experiment results.

Keywords: seiun maru propeller, steady, unstead, CFD, HSP

Procedia PDF Downloads 308
2459 An Alternative Credit Scoring System in China’s Consumer Lendingmarket: A System Based on Digital Footprint Data

Authors: Minjuan Sun

Abstract:

Ever since the late 1990s, China has experienced explosive growth in consumer lending, especially in short-term consumer loans, among which, the growth rate of non-bank lending has surpassed bank lending due to the development in financial technology. On the other hand, China does not have a universal credit scoring and registration system that can guide lenders during the processes of credit evaluation and risk control, for example, an individual’s bank credit records are not available for online lenders to see and vice versa. Given this context, the purpose of this paper is three-fold. First, we explore if and how alternative digital footprint data can be utilized to assess borrower’s creditworthiness. Then, we perform a comparative analysis of machine learning methods for the canonical problem of credit default prediction. Finally, we analyze, from an institutional point of view, the necessity of establishing a viable and nationally universal credit registration and scoring system utilizing online digital footprints, so that more people in China can have better access to the consumption loan market. Two different types of digital footprint data are utilized to match with bank’s loan default records. Each separately captures distinct dimensions of a person’s characteristics, such as his shopping patterns and certain aspects of his personality or inferred demographics revealed by social media features like profile image and nickname. We find both datasets can generate either acceptable or excellent prediction results, and different types of data tend to complement each other to get better performances. Typically, the traditional types of data banks normally use like income, occupation, and credit history, update over longer cycles, hence they can’t reflect more immediate changes, like the financial status changes caused by the business crisis; whereas digital footprints can update daily, weekly, or monthly, thus capable of providing a more comprehensive profile of the borrower’s credit capabilities and risks. From the empirical and quantitative examination, we believe digital footprints can become an alternative information source for creditworthiness assessment, because of their near-universal data coverage, and because they can by and large resolve the "thin-file" issue, due to the fact that digital footprints come in much larger volume and higher frequency.

Keywords: credit score, digital footprint, Fintech, machine learning

Procedia PDF Downloads 166
2458 A Study on Numerical Modelling of Rigid Pavement: Temperature and Thickness Effect

Authors: Amin Chegenizadeh, Mahdi Keramatikerman, Hamid Nikraz

Abstract:

Pavement engineering plays a significant role to develop cost effective and efficient highway and road networks. In general, pavement regarding structure is categorized in two core group namely flexible and rigid pavements. There are various benefits in application of rigid pavement. For instance, they have a longer life and lower maintenance costs in compare with the flexible pavement. In rigid pavement designs, temperature and thickness are two effective parameters that could widely affect the total cost of the project. In this study, a numerical modeling using Kenpave-Kenslab was performed to investigate the effect of these two important parameters in the rigid pavement.   

Keywords: rigid pavement, Kenpave, Kenslab, thickness, temperature

Procedia PDF Downloads 374
2457 Nondestructive Prediction and Classification of Gel Strength in Ethanol-Treated Kudzu Starch Gels Using Near-Infrared Spectroscopy

Authors: John-Nelson Ekumah, Selorm Yao-Say Solomon Adade, Mingming Zhong, Yufan Sun, Qiufang Liang, Muhammad Safiullah Virk, Xorlali Nunekpeku, Nana Adwoa Nkuma Johnson, Bridget Ama Kwadzokpui, Xiaofeng Ren

Abstract:

Enhancing starch gel strength and stability is crucial. However, traditional gel property assessment methods are destructive, time-consuming, and resource-intensive. Thus, understanding ethanol treatment effects on kudzu starch gel strength and developing a rapid, nondestructive gel strength assessment method is essential for optimizing the treatment process and ensuring product quality consistency. This study investigated the effects of different ethanol concentrations on the microstructure of kudzu starch gels using a comprehensive microstructural analysis. We also developed a nondestructive method for predicting gel strength and classifying treatment levels using near-infrared (NIR) spectroscopy, and advanced data analytics. Scanning electron microscopy revealed progressive network densification and pore collapse with increasing ethanol concentration, correlating with enhanced mechanical properties. NIR spectroscopy, combined with various variable selection methods (CARS, GA, and UVE) and modeling algorithms (PLS, SVM, and ELM), was employed to develop predictive models for gel strength. The UVE-SVM model demonstrated exceptional performance, with the highest R² values (Rc = 0.9786, Rp = 0.9688) and lowest error rates (RMSEC = 6.1340, RMSEP = 6.0283). Pattern recognition algorithms (PCA, LDA, and KNN) successfully classified gels based on ethanol treatment levels, achieving near-perfect accuracy. This integrated approach provided a multiscale perspective on ethanol-induced starch gel modification, from molecular interactions to macroscopic properties. Our findings demonstrate the potential of NIR spectroscopy, coupled with advanced data analysis, as a powerful tool for rapid, nondestructive quality assessment in starch gel production. This study contributes significantly to the understanding of starch modification processes and opens new avenues for research and industrial applications in food science, pharmaceuticals, and biomaterials.

Keywords: kudzu starch gel, near-infrared spectroscopy, gel strength prediction, support vector machine, pattern recognition algorithms, ethanol treatment

Procedia PDF Downloads 41
2456 A Comparative Study of Optimization Techniques and Models to Forecasting Dengue Fever

Authors: Sudha T., Naveen C.

Abstract:

Dengue is a serious public health issue that causes significant annual economic and welfare burdens on nations. However, enhanced optimization techniques and quantitative modeling approaches can predict the incidence of dengue. By advocating for a data-driven approach, public health officials can make informed decisions, thereby improving the overall effectiveness of sudden disease outbreak control efforts. The National Oceanic and Atmospheric Administration and the Centers for Disease Control and Prevention are two of the U.S. Federal Government agencies from which this study uses environmental data. Based on environmental data that describe changes in temperature, precipitation, vegetation, and other factors known to affect dengue incidence, many predictive models are constructed that use different machine learning methods to estimate weekly dengue cases. The first step involves preparing the data, which includes handling outliers and missing values to make sure the data is prepared for subsequent processing and the creation of an accurate forecasting model. In the second phase, multiple feature selection procedures are applied using various machine learning models and optimization techniques. During the third phase of the research, machine learning models like the Huber Regressor, Support Vector Machine, Gradient Boosting Regressor (GBR), and Support Vector Regressor (SVR) are compared with several optimization techniques for feature selection, such as Harmony Search and Genetic Algorithm. In the fourth stage, the model's performance is evaluated using Mean Square Error (MSE), Mean Absolute Error (MAE), and Root Mean Square Error (RMSE) as assistance. Selecting an optimization strategy with the least number of errors, lowest price, biggest productivity, or maximum potential results is the goal. In a variety of industries, including engineering, science, management, mathematics, finance, and medicine, optimization is widely employed. An effective optimization method based on harmony search and an integrated genetic algorithm is introduced for input feature selection, and it shows an important improvement in the model's predictive accuracy. The predictive models with Huber Regressor as the foundation perform the best for optimization and also prediction.

Keywords: deep learning model, dengue fever, prediction, optimization

Procedia PDF Downloads 67
2455 Pres Syndrome in Pregnancy: A Case Series of Five Cases

Authors: Vaibhavi Birle

Abstract:

Posterior reversible encephalopathy syndrome is a rare clinic-radiological syndrome associated with acute changes in blood pressure during pregnancy. It is characterized symptomatically by headache, seizures, altered mental status, and visual blurring with radiological changes of white matter (vasogenic oedema) affecting the posterior occipital and parietal lobes of the brain. It is being increasingly recognized due to increased institutional deliveries and advances in imaging particularly magnetic resonance imaging (MRI). In spite of the increasing diagnosis the prediction of PRES and patient factors affecting susceptibility is still not clear. Hence, we conducted the retrospective study to analyse the factors associated with PRES at our tertiary centre.

Keywords: pres syndrome, eclampsia, maternal outcome, fetal outcome

Procedia PDF Downloads 153
2454 De-convolution Based IVIVC Correlation for Tacrolimus ER Tablet (Narrow Therapeutic Index Drug) With Widening of Dissolution Prediction for Virtual Bioequivalence

Authors: Sajad Khaliq Dar, Dipanjan Goswami, Arshad H. Khuroo, Mohd. Akhtar, Pulak Kumar Metia, Sudershan Kumar

Abstract:

Background: Development of modified-release oral dosage formulations (OSD) like tacrolimus in narrow therapeutic categories, together with high levels of intra-individual variability, impose greater challenges. The risk assessment for bioequivalence studies requires developing a suitable design through pilot studies involving the comparison of multiple formulations of the same product with a marketed product to understand the in-vivo behaviour. These formulations could have varying coating levels and other minor quantitative differences to achieve the desired release rate for the final product. Although small-scale studies are critical before the conduct of full-scale Pharmacokinetic (PK) studies, regulatory agencies evaluate critical bioavailability attributes (CBA) before approving the submitted dossiers. Since Tacrolimus is a BCS Class II drug, therefore developing the extended-release formulation, in addition to associated challenges, provides an opportunity to present the In vitro-in vivo correlations (IVIVC) to regulatory agencies, not only to exhibit product quality but also to reduce the burden of additional human trials and cost involved to them for bringing the product to market. Objective: The objective of this study was to develop a Level-A In vitro - In vivo Correlation (IVIVC) model for Sun Pharma’s test formulation Tacrolimus ER tablet 4mg and extend its application to a widened dissolution window of 25% at 2.5 hours (critical release time) sampling time point. Experimental Procedure: Post the conduct of two in-vivo studies, a pilot study evaluating two test prototypes on 24 subjects (under fasting) and a pivotal study having 50 subjects (under fasting), the observed pharmacokinetic profile was used for IVIVC model development. The dissolution media used was 0.005% HPC + 0.25% SLS in Water 900 mL at pH 4.50 using USP II (Paddle) apparatus with alternative sinkers operated at 100 RPM. The sampling time points were chosen to mimic the drug absorption in vivo. The dissolution best fit to data was obtained using Makoid Banakar kinetics. Then deconvolution, anchoring to concepts of the single compartment by Wagner Nelson method was applied for tacrolimus slow-release formulation batch with film coating weight build-up of 5.4% (used in pilot bio study), medium release with Hypromellose (retard-release exhibit batch used in the pivotal study) and fast release formulation batch with film coating weight build-up of 5.05% (used in pilot bio study). Results and Conclusion: The results were deemed acceptable as prediction errors for internal and external validation were < 3% depicting in-vitro drug release mimics in-vivo absorption. Moreover, the prediction result for the Test/Reference ratio was <15% for all test formulations and widening dissolution (i.e., 39%-64% drug release at 2.5hrs) predictions were well within 80-125% when compared against Envarsus XR (reference drug). This IVIVC-validated model can be used in the futuristic exploration of dose titration with 1mg tacrolimus ER OSD as a surrogate for In-vivo bioequivalence trials.

Keywords: pharmacokinetics, BCS, oral dosage form, Bioavailability, intra-individual variability

Procedia PDF Downloads 6
2453 Prediction of Anticancer Potential of Curcumin Nanoparticles by Means of Quasi-Qsar Analysis Using Monte Carlo Method

Authors: Ruchika Goyal, Ashwani Kumar, Sandeep Jain

Abstract:

The experimental data for anticancer potential of curcumin nanoparticles was calculated by means of eclectic data. The optimal descriptors were examined using Monte Carlo method based CORAL SEA software. The statistical quality of the model is following: n = 14, R² = 0.6809, Q² = 0.5943, s = 0.175, MAE = 0.114, F = 26 (sub-training set), n =5, R²= 0.9529, Q² = 0.7982, s = 0.086, MAE = 0.068, F = 61, Av Rm² = 0.7601, ∆R²m = 0.0840, k = 0.9856 and kk = 1.0146 (test set) and n = 5, R² = 0.6075 (validation set). This data can be used to build predictive QSAR models for anticancer activity.

Keywords: anticancer potential, curcumin, model, nanoparticles, optimal descriptors, QSAR

Procedia PDF Downloads 320
2452 Defects Estimation of Embedded Systems Components by a Bond Graph Approach

Authors: I. Gahlouz, A. Chellil

Abstract:

The paper concerns the estimation of system components faults by using an unknown inputs observer. To reach this goal, we used the Bond Graph approach to physical modelling. We showed that this graphical tool is allowing the representation of system components faults as unknown inputs within the state representation of the considered physical system. The study of the causal and structural features of the system (controllability, observability, finite structure, and infinite structure) based on the Bond Graph approach was hence fulfilled in order to design an unknown inputs observer which is used for the system component fault estimation.

Keywords: estimation, bond graph, controllability, observability

Procedia PDF Downloads 414
2451 Prediction of Fire Growth of the Office by Real-Scale Fire Experiment

Authors: Kweon Oh-Sang, Kim Heung-Youl

Abstract:

Estimating the engineering properties of fires is important to be prepared for the complex and various fire risks of large-scale structures such as super-tall buildings, large stadiums, and multi-purpose structures. In this study, a mock-up of a compartment which was 2.4(L) x 3.6 (W) x 2.4 (H) meter in dimensions was fabricated at the 10MW LSC (Large Scale Calorimeter) and combustible office supplies were placed in the compartment for a real-scale fire test. Maximum heat release rate was 4.1 MW and total energy release obtained through the application of t2 fire growth rate was 6705.9 MJ.

Keywords: fire growth, fire experiment, t2 curve, large scale calorimeter

Procedia PDF Downloads 338
2450 Ray Tracing Modified 3D Image Method Simulation of Picocellular Propagation Channel Environment

Authors: Fathi Alwafie

Abstract:

In this paper we present the simulation of the propagation characteristics of the picocellular propagation channel environment. The first aim has been to find a correct description of the environment for received wave. The result of the first investigations is that the environment of the indoor wave significantly changes as we change the electric parameters of material constructions. A modified 3D ray tracing image method tool has been utilized for the coverage prediction. A detailed analysis of the dependence of the indoor wave on the wide-band characteristics of the channel: Root Mean Square (RMS) delay spread characteristics and mean excess delay, is also investigated.

Keywords: propagation, ray tracing, network, mobile computing

Procedia PDF Downloads 401
2449 Thermomechanical Damage Modeling of F114 Carbon Steel

Authors: A. El Amri, M. El Yakhloufi Haddou, A. Khamlichi

Abstract:

The numerical simulation based on the Finite Element Method (FEM) is widely used in academic institutes and in the industry. It is a useful tool to predict many phenomena present in the classical manufacturing forming processes such as fracture. But, the results of such numerical model depend strongly on the parameters of the constitutive behavior model. The influences of thermal and mechanical loads cause damage. The temperature and strain rate dependent materials’ properties and their modelling are discussed. A Johnson-Cook Model of damage has been selected for the numerical simulations. Virtual software called the ABAQUS 6.11 is used for finite element analysis. This model was introduced in order to give information concerning crack initiation during thermal and mechanical loads.

Keywords: thermo-mechanical fatigue, failure, numerical simulation, fracture, damage

Procedia PDF Downloads 394
2448 An Incremental Refinement Approach to a Development of Dynamic Host Configuration Protocol (DHCP) Using Event-B

Authors: Rajaa Filali, Mohamed Bouhdadi

Abstract:

This paper presents an incremental development of the Dynamic Host Configuration Protocol (DHCP) in Event-B. DHCP is widely used communication protocol, which provides a standard mechanism to obtain configuration parameters. The specification is performed in a stepwise manner and verified through a series of refinements. The Event-B formal method uses the Rodin platform to modeling and verifying some properties of the protocol such as safety, liveness and deadlock freedom. To model and verify the protocol, we use the formal technique Event-B which provides an accessible and rigorous development method. This interaction between modelling and proving reduces the complexity and helps to eliminate misunderstandings, inconsistencies, and specification gaps.

Keywords: DHCP protocol, Event-B, refinement, proof obligation, Rodin

Procedia PDF Downloads 229
2447 EGF Serum Level in Diagnosis and Prediction of Mood Disorder in Adolescents and Young Adults

Authors: Monika Dmitrzak-Weglarz, Aleksandra Rajewska-Rager, Maria Skibinska, Natalia Lepczynska, Piotr Sibilski, Joanna Pawlak, Pawel Kapelski, Joanna Hauser

Abstract:

Epidermal growth factor (EGF) is a well-known neurotrophic factor that involves in neuronal growth and synaptic plasticity. The proteomic research provided in order to identify novel candidate biological markers for mood disorders focused on elevated EGF serum level in patients during depression episode. However, the EGF association with mood disorder spectrum among adolescents and young adults has not been studied extensively. In this study, we aim to investigate the serum levels of EGF in adolescents and young adults during hypo/manic, depressive episodes and in remission compared to healthy control group. In our study, we involved 80 patients aged 12-24 years in 2-year follow-up study with a primary diagnosis of mood disorder spectrum, and 35 healthy volunteers matched by age and gender. Diagnoses were established according to DSM-IV-TR criteria using structured clinical interviews: K-SADS for child and adolescents, and SCID for young adults. Clinical and biological evaluations were made at baseline and euthymic mood (at 3th or 6th month of treatment and after 1 and 2 years). The Young Mania Rating Scale and Hamilton Rating Scale for Depression were used for assessment. The study protocols were approved by the relevant ethics committee. Serum protein concentration was determined by Enzyme-Linked Immunosorbent Assays (ELISA) method. Human EGF (cat. no DY 236) DuoSet ELISA kit was used (R&D Systems). Serum EGF levels were analysed with following variables: age, age under 18 and above 18 years old, sex, family history of affective disorders, drug-free vs. medicated. Shapiro-Wilk test was used to test the normality of the data. The homogeneity of variance was calculated with Levene’s test. EGF levels showed non-normal distribution and the homogeneity of variance was violated. Non-parametric tests: Mann-Whitney U test, Kruskall-Wallis ANOVA, Friedman’s ANOVA, Wilcoxon signed rank test, Spearman correlation coefficient was applied in the analyses The statistical significance level was set at p<0.05. Elevated EGF level at baseline (p=0.001) and at month 24 (p=0.02) was detected in study subjects compared with controls. Increased EGF level in women at month 12 (p=0.02) compared to men in study group have been observed. Using Wilcoxon signed rank test differences in EGF levels were detected: decrease from baseline to month 3 (p=0.014) and increase comparing: month 3 vs. 24 (p=0.013); month 6 vs. 12 (p=0.021) and vs. 24 (p=0.008). EGF level at baseline was negatively correlated with depression and mania occurrence at 24 months. EGF level at 24 months was positively correlated with depression and mania occurrence at 12 months. No other correlations of EGF levels with clinical and demographical variables have been detected. The findings of the present study indicate that EGF serum level is significantly elevated in the study group of patients compared to the controls. We also observed fluctuations in EGF levels during two years of disease observation. EGF seems to be useful as an early marker for prediction of diagnosis, course of illness and treatment response in young patients during first episode od mood disorders, which requires further investigation. Grant was founded by National Science Center in Poland no 2011/03/D/NZ5/06146.

Keywords: biological marker, epidermal growth factor, mood disorders, prediction

Procedia PDF Downloads 190
2446 Effect of Downstream Pressure in Tuning the Flow Control Orifices of Pressure Fed Reaction Control System Thrusters

Authors: Prakash M.N, Mahesh G, Muhammed Rafi K.M, Shiju P. Nair

Abstract:

Introduction: In launch vehicle missions, Reaction Control thrusters are being used for the three-axis stabilization of the vehicle during the coasting phases. A pressure-fed propulsion system is used for the operation of these thrusters due to its less complexity. In liquid stages, these thrusters are designed to draw propellant from the same tank used for the main propulsion system. So in order to regulate the propellant flow rates of these thrusters, flow control orifices are used in feed lines. These orifices are calibrated separately as per the flow rate requirement of individual thrusters for the nominal operating conditions. In some missions, it was observed that the thrusters were operated at higher thrust than nominal. This point was addressed through a series of cold flow and hot tests carried out in-ground and this paper elaborates the details of the same. Discussion: In order to find out the exact reason for this phenomenon, two flight configuration thrusters were identified and hot tested in the ground with calibrated orifices and feed lines. During these tests, the chamber pressure, which is directly proportional to the thrust, is measured. In both cases, chamber pressures higher than the nominal by 0.32bar to 0.7bar were recorded. The increase in chamber pressure is due to an increase in the oxidizer flow rate of both the thrusters. Upon further investigation, it is observed that the calibration of the feed line is done with ambient pressure downstream. But in actual flight conditions, the orifices will be subjected to operate with 10 to 11bar pressure downstream. Due to this higher downstream pressure, the flow through the orifices increases and thereby, the thrusters operate with higher chamber pressure values. Conclusion: As part of further investigatory tests, two numbers of fresh thrusters were realized. Orifice tuning of these thrusters was carried out in three different ways. In the first trial, the orifice tuning was done by simulating 1bar pressure downstream. The second trial was done with the injector assembled downstream. In the third trial, the downstream pressure equal to the flight injection pressure was simulated downstream. Using these calibrated orifices, hot tests were carried out in simulated vacuum conditions. Chamber pressure and flow rate values were exactly matching with the prediction for the second and third trials. But for the first trial, the chamber pressure values obtained in the hot test were more than the prediction. This clearly shows that the flow is detached in the 1st trial and attached for the 2nd & 3rd trials. Hence, the error in tuning the flow control orifices is pinpointed as the reason for this higher chamber pressure observed in flight.

Keywords: reaction control thruster, propellent, orifice, chamber pressure

Procedia PDF Downloads 202
2445 Predictive Modelling of Curcuminoid Bioaccessibility as a Function of Food Formulation and Associated Properties

Authors: Kevin De Castro Cogle, Mirian Kubo, Maria Anastasiadi, Fady Mohareb, Claire Rossi

Abstract:

Background: The bioaccessibility of bioactive compounds is a critical determinant of the nutritional quality of various food products. Despite its importance, there is a limited number of comprehensive studies aimed at assessing how the composition of a food matrix influences the bioaccessibility of a compound of interest. This knowledge gap has prompted a growing need to investigate the intricate relationship between food matrix formulations and the bioaccessibility of bioactive compounds. One such class of bioactive compounds that has attracted considerable attention is curcuminoids. These naturally occurring phytochemicals, extracted from the roots of Curcuma longa, have gained popularity owing to their purported health benefits and also well known for their poor bioaccessibility Project aim: The primary objective of this research project is to systematically assess the influence of matrix composition on the bioaccessibility of curcuminoids. Additionally, this study aimed to develop a series of predictive models for bioaccessibility, providing valuable insights for optimising the formula for functional foods and provide more descriptive nutritional information to potential consumers. Methods: Food formulations enriched with curcuminoids were subjected to in vitro digestion simulation, and their bioaccessibility was characterized with chromatographic and spectrophotometric techniques. The resulting data served as the foundation for the development of predictive models capable of estimating bioaccessibility based on specific physicochemical properties of the food matrices. Results: One striking finding of this study was the strong correlation observed between the concentration of macronutrients within the food formulations and the bioaccessibility of curcuminoids. In fact, macronutrient content emerged as a very informative explanatory variable of bioaccessibility and was used, alongside other variables, as predictors in a Bayesian hierarchical model that predicted curcuminoid bioaccessibility accurately (optimisation performance of 0.97 R2) for the majority of cross-validated test formulations (LOOCV of 0.92 R2). These preliminary results open the door to further exploration, enabling researchers to investigate a broader spectrum of food matrix types and additional properties that may influence bioaccessibility. Conclusions: This research sheds light on the intricate interplay between food matrix composition and the bioaccessibility of curcuminoids. This study lays a foundation for future investigations, offering a promising avenue for advancing our understanding of bioactive compound bioaccessibility and its implications for the food industry and informed consumer choices.

Keywords: bioactive bioaccessibility, food formulation, food matrix, machine learning, probabilistic modelling

Procedia PDF Downloads 70
2444 Monitoring and Prediction of Intra-Crosstalk in All-Optical Network

Authors: Ahmed Jedidi, Mesfer Mohammed Alshamrani, Alwi Mohammad A. Bamhdi

Abstract:

Optical performance monitoring and optical network management are essential in building a reliable, high-capacity, and service-differentiation enabled all-optical network. One of the serious problems in this network is the fact that optical crosstalk is additive, and thus the aggregate effect of crosstalk over a whole AON may be more nefarious than a single point of crosstalk. As results, we note a huge degradation of the Quality of Service (QoS) in our network. For that, it is necessary to identify and monitor the impairments in whole network. In this way, this paper presents new system to identify and monitor crosstalk in AONs in real-time fashion. particular, it proposes a new technique to manage intra-crosstalk in objective to relax QoS of the network.

Keywords: all-optical networks, optical crosstalk, optical cross-connect, crosstalk, monitoring crosstalk

Procedia PDF Downloads 465
2443 Numerical Prediction of Width Crack of Concrete Dapped-End Beams

Authors: Jatziri Y. Moreno-Martinez, Arturo Galvan, Xavier Chavez Cardenas, Hiram Arroyo

Abstract:

Several methods have been utilized to study the prediction of cracking of concrete structural under loading. The finite element analysis is an alternative that shows good results. The aim of this work was the numerical study of the width crack in reinforced concrete beams with dapped ends, these are frequently found in bridge girders and precast concrete construction. Properly restricting cracking is an important aspect of the design in dapped ends, it has been observed that the cracks that exceed the allowable widths are unacceptable in an aggressive environment for reinforcing steel. For simulating the crack width, the discrete crack approach was considered by means of a Cohesive Zone (CZM) Model using a function to represent the crack opening. Two cases of dapped-end were constructed and tested in the laboratory of Structures and Materials of Engineering Institute of UNAM. The first case considers a reinforcement based on hangers as well as on vertical and horizontal ring, the second case considers 50% of the vertical stirrups in the dapped end to the main part of the beam were replaced by an equivalent area (vertically projected) of diagonal bars under. The loading protocol consisted on applying symmetrical loading to reach the service load. The models were performed using the software package ANSYS v. 16.2. The concrete structure was modeled using three-dimensional solid elements SOLID65 capable of cracking in tension and crushing in compression. Drucker-Prager yield surface was used to include the plastic deformations. The reinforcement was introduced with smeared approach. Interface delamination was modeled by traditional fracture mechanics methods such as the nodal release technique adopting softening relationships between tractions and the separations, which in turn introduce a critical fracture energy that is also the energy required to break apart the interface surfaces. This technique is called CZM. The interface surfaces of the materials are represented by a contact elements Surface-to-Surface (CONTA173) with bonded (initial contact). The Mode I dominated bilinear CZM model assumes that the separation of the material interface is dominated by the displacement jump normal to the interface. Furthermore, the opening crack was taken into consideration according to the maximum normal contact stress, the contact gap at the completion of debonding, and the maximum equivalent tangential contact stress. The contact elements were placed in the crack re-entrant corner. To validate the proposed approach, the results obtained with the previous procedure are compared with experimental test. A good correlation between the experimental and numerical Load-Displacement curves was presented, the numerical models also allowed to obtain the load-crack width curves. In these two cases, the proposed model confirms the capability of predicting the maximum crack width, with an error of ± 30 %. Finally, the orientation of the crack is a fundamental for the prediction of crack width. The results regarding the crack width can be considered as good from the practical point view. Load-Displacement curve of the test and the location of the crack were able to obtain favorable results.

Keywords: cohesive zone model, dapped-end beams, discrete crack approach, finite element analysis

Procedia PDF Downloads 169
2442 Modelling of Pervaporation Separation of Butanol from Aqueous Solutions Using Polydimethylsiloxane Mixed Matrix Membranes

Authors: Arian Ebneyamini, Hoda Azimi, Jules Thibaults, F. Handan Tezel

Abstract:

In this study, a modification of Hennepe model for pervaporation separation of butanol from aqueous solutions using Polydimethylsiloxane (PDMS) mixed matrix membranes has been introduced and validated by experimental data. The model was compared to the original Hennepe model and few other models which are applicable for membrane gas separation processes such as Maxwell, Lewis Nielson and Pal. Theoretical modifications for non-ideal interface morphology have been offered to predict the permeability in case of interface void, interface rigidification and pore-blockage. The model was in a good agreement with experimental data.

Keywords: butanol, PDMS, modeling, pervaporation, mixed matrix membranes

Procedia PDF Downloads 222
2441 Currency Exchange Rate Forecasts Using Quantile Regression

Authors: Yuzhi Cai

Abstract:

In this paper, we discuss a Bayesian approach to quantile autoregressive (QAR) time series model estimation and forecasting. Together with a combining forecasts technique, we then predict USD to GBP currency exchange rates. Combined forecasts contain all the information captured by the fitted QAR models at different quantile levels and are therefore better than those obtained from individual models. Our results show that an unequally weighted combining method performs better than other forecasting methodology. We found that a median AR model can perform well in point forecasting when the predictive density functions are symmetric. However, in practice, using the median AR model alone may involve the loss of information about the data captured by other QAR models. We recommend that combined forecasts should be used whenever possible.

Keywords: combining forecasts, MCMC, predictive density functions, quantile forecasting, quantile modelling

Procedia PDF Downloads 257
2440 Role of von Willebrand Factor and ADAMTS13 In The Prediction of Thrombotic Complications In Patients With COVID-19

Authors: Nataliya V. Dolgushina, Elena A. Gorodnova, Olga S. Beznoshenco, Andrey Yu Romanov, Irina V. Menzhinskaya, Lyubov V. Krechetova, Gennady T. Suchich

Abstract:

In patients with COVID-19, generalized hypercoagulability can lead to the development of severe coagulopathy. This event is accompanied by the development of a pronounced inflammatory reaction. The observational prospective study included 39 patients with mild COVID-19 and 102 patients with moderate and severe COVID-19. Patients were then stratified into groups depending on the risk of venous thromboembolism. vWF to ADAMTS-13 concentrations and activity ratios were significantly higher in patients with a high venous thromboembolism risks in patients with moderate and severe forms COVID-19.

Keywords: ADAMTS-13, COVID-19, hypercoagulation, thrombosis, von Willebrand factor

Procedia PDF Downloads 89
2439 An Empirical Dynamic Fuel Cell Model Used for Power System Verification in Aerospace

Authors: Giuliano Raimondo, Jörg Wangemann, Peer Drechsel

Abstract:

In systems development involving Fuel Cells generators, it is important to have from an early stage of the project a dynamic model for the electrical behavior of the stack to be shared between involved development parties. It allows independent and early design and tests of fuel cell related power electronic. This paper presents an empirical Fuel Cell system model derived from characterization tests on a real system. Moreover, it is illustrated how the obtained model is used to build and validate a real-time Fuel Cell system emulator which is used for aerospace electrical integration testing activities.

Keywords: fuel cell, modelling, real time emulation, testing

Procedia PDF Downloads 338
2438 Implementation of Iterative Algorithm for Earthquake Location

Authors: Hussain K. Chaiel

Abstract:

The development in the field of the digital signal processing (DSP) and the microelectronics technology reduces the complexity of the iterative algorithms that need large number of arithmetic operations. Virtex-Field Programmable Gate Arrays (FPGAs) are programmable silicon foundations which offer an important solution for addressing the needs of high performance DSP designer. In this work, Virtex-7 FPGA technology is used to implement an iterative algorithm to estimate the earthquake location. Simulation results show that an implementation based on block RAMB36E1 and DSP48E1 slices of Virtex-7 type reduces the number of cycles of the clock frequency. This enables the algorithm to be used for earthquake prediction.

Keywords: DSP, earthquake, FPGA, iterative algorithm

Procedia PDF Downloads 389