Search results for: predict
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2369

Search results for: predict

539 Application of Stochastic Models on the Portuguese Population and Distortion to Workers Compensation Pensioners Experience

Authors: Nkwenti Mbelli Njah

Abstract:

This research was motivated by a project requested by AXA on the topic of pensions payable under the workers compensation (WC) line of business. There are two types of pensions: the compulsorily recoverable and the not compulsorily recoverable. A pension is compulsorily recoverable for a victim when there is less than 30% of disability and the pension amount per year is less than six times the minimal national salary. The law defines that the mathematical provisions for compulsory recoverable pensions must be calculated by applying the following bases: mortality table TD88/90 and rate of interest 5.25% (maybe with rate of management). To manage pensions which are not compulsorily recoverable is a more complex task because technical bases are not defined by law and much more complex computations are required. In particular, companies have to predict the amount of payments discounted reflecting the mortality effect for all pensioners (this task is monitored monthly in AXA). The purpose of this research was thus to develop a stochastic model for the future mortality of the worker’s compensation pensioners of both the Portuguese market workers and AXA portfolio. Not only is past mortality modeled, also projections about future mortality are made for the general population of Portugal as well as for the two portfolios mentioned earlier. The global model was split in two parts: a stochastic model for population mortality which allows for forecasts, combined with a point estimate from a portfolio mortality model obtained through three different relational models (Cox Proportional, Brass Linear and Workgroup PLT). The one-year death probabilities for ages 0-110 for the period 2013-2113 are obtained for the general population and the portfolios. These probabilities are used to compute different life table functions as well as the not compulsorily recoverable reserves for each of the models required for the pensioners, their spouses and children under 21. The results obtained are compared with the not compulsory recoverable reserves computed using the static mortality table (TD 73/77) that is currently being used by AXA, to see the impact on this reserve if AXA adopted the dynamic tables.

Keywords: compulsorily recoverable, life table functions, relational models, worker’s compensation pensioners

Procedia PDF Downloads 153
538 Indian Premier League (IPL) Score Prediction: Comparative Analysis of Machine Learning Models

Authors: Rohini Hariharan, Yazhini R, Bhamidipati Naga Shrikarti

Abstract:

In the realm of cricket, particularly within the context of the Indian Premier League (IPL), the ability to predict team scores accurately holds significant importance for both cricket enthusiasts and stakeholders alike. This paper presents a comprehensive study on IPL score prediction utilizing various machine learning algorithms, including Support Vector Machines (SVM), XGBoost, Multiple Regression, Linear Regression, K-nearest neighbors (KNN), and Random Forest. Through meticulous data preprocessing, feature engineering, and model selection, we aimed to develop a robust predictive framework capable of forecasting team scores with high precision. Our experimentation involved the analysis of historical IPL match data encompassing diverse match and player statistics. Leveraging this data, we employed state-of-the-art machine learning techniques to train and evaluate the performance of each model. Notably, Multiple Regression emerged as the top-performing algorithm, achieving an impressive accuracy of 77.19% and a precision of 54.05% (within a threshold of +/- 10 runs). This research contributes to the advancement of sports analytics by demonstrating the efficacy of machine learning in predicting IPL team scores. The findings underscore the potential of advanced predictive modeling techniques to provide valuable insights for cricket enthusiasts, team management, and betting agencies. Additionally, this study serves as a benchmark for future research endeavors aimed at enhancing the accuracy and interpretability of IPL score prediction models.

Keywords: indian premier league (IPL), cricket, score prediction, machine learning, support vector machines (SVM), xgboost, multiple regression, linear regression, k-nearest neighbors (KNN), random forest, sports analytics

Procedia PDF Downloads 29
537 Resale Housing Development Board Price Prediction Considering Covid-19 through Sentiment Analysis

Authors: Srinaath Anbu Durai, Wang Zhaoxia

Abstract:

Twitter sentiment has been used as a predictor to predict price values or trends in both the stock market and housing market. The pioneering works in this stream of research drew upon works in behavioural economics to show that sentiment or emotions impact economic decisions. Latest works in this stream focus on the algorithm used as opposed to the data used. A literature review of works in this stream through the lens of data used shows that there is a paucity of work that considers the impact of sentiments caused due to an external factor on either the stock or the housing market. This is despite an abundance of works in behavioural economics that show that sentiment or emotions caused due to an external factor impact economic decisions. To address this gap, this research studies the impact of Twitter sentiment pertaining to the Covid-19 pandemic on resale Housing Development Board (HDB) apartment prices in Singapore. It leverages SNSCRAPE to collect tweets pertaining to Covid-19 for sentiment analysis, lexicon based tools VADER and TextBlob are used for sentiment analysis, Granger Causality is used to examine the relationship between Covid-19 cases and the sentiment score, and neural networks are leveraged as prediction models. Twitter sentiment pertaining to Covid-19 as a predictor of HDB price in Singapore is studied in comparison with the traditional predictors of housing prices i.e., the structural and neighbourhood characteristics. The results indicate that using Twitter sentiment pertaining to Covid19 leads to better prediction than using only the traditional predictors and performs better as a predictor compared to two of the traditional predictors. Hence, Twitter sentiment pertaining to an external factor should be considered as important as traditional predictors. This paper demonstrates the real world economic applications of sentiment analysis of Twitter data.

Keywords: sentiment analysis, Covid-19, housing price prediction, tweets, social media, Singapore HDB, behavioral economics, neural networks

Procedia PDF Downloads 93
536 Association of 1565C/T Polymorphism of Integrin Beta-3 (ITGB3) Gene and Increased Risk for Myocardial Infarction in Patients with Premature Coronary Artery Disease among Iranian Population

Authors: Mehrdad Sheikhvatan, Mohammad Ali Boroumand, Mehrdad Behmanesh, Shayan Ziaee

Abstract:

Contradictory results have been obtained regarding the role of integrin, beta 3 (ITGB3) gene polymorphisms in occurrence of acute myocardial infarction (MI) in patients with coronary artery disease (CAD). Hence, we aimed to assess the association between 1565C/T polymorphism of ITGB3 gene and increased risk for acute MI in patients who suffered premature CAD in Iranian population. Our prospective study included 1000 patients (492 men and 508 women aged 21 to 55 years) referred to Tehran Heart center during a period of four years from 2008 to 2011 with the final diagnosis of premature CAD and classified into two groups with history of MI (n = 461) and without of MI (n = 539). The polymorphism variants were determined by PCR-RFLP technique by entering 10% of randomized samples and then genotyping of the polymorphism was also conducted by High Resolution Melting (HRM) method. Among study samples, 640 were followed with a median follow-up time 45.74 months for determining association of long-term major adverse cardiac events (MACE) and genotypes of polymorphisms. There was no significant difference in the frequency of 1565C/T polymorphism between the MI and non-MI groups. The frequency of wild genotype was 69.2% and 72.2%, the frequency of homozygous genotype was 21.3% and 18.4%, and the frequency of mutant genotype was 9.5% and 9.5%, respectively (p=0.505). Results were also similar when adjusted for covariates in a multivariate logistic regression model. No significant difference was also found in total-MACE free survival rate between the patients with different genotypes of 1565C/T polymorphism in both MI and non-MI group. The carriage of the 1565C/T polymorphism of ITGB3 gene seems unlikely to be a significant risk factor for the development of MI in Iranian patients with premature CAD. The presence of this ITGB3 gene polymorphism may not also predict long-term cardiac events.

Keywords: coronary artery disease, myocardial infarction, gene, integrin, beta 3, polymorphism

Procedia PDF Downloads 383
535 Numerical Modelling of Shear Zone and Its Implications on Slope Instability at Letšeng Diamond Open Pit Mine, Lesotho

Authors: M. Ntšolo, D. Kalumba, N. Lefu, G. Letlatsa

Abstract:

Rock mass damage due to shear tectonic activity has been investigated largely in geoscience where fluid transport is of major interest. However, little has been studied on the effect of shear zones on rock mass behavior and its impact on stability of rock slopes. At Letšeng Diamonds open pit mine in Lesotho, the shear zone composed of sheared kimberlite material, calcite and altered basalt is forming part of the haul ramp into the main pit cut 3. The alarming rate at which the shear zone is deteriorating has triggered concerns about both local and global stability of pit the walls. This study presents the numerical modelling of the open pit slope affected by shear zone at Letšeng Diamond Mine (LDM). Analysis of the slope involved development of the slope model by using a two-dimensional finite element code RS2. Interfaces between shear zone and host rock were represented by special joint elements incorporated in the finite element code. The analysis of structural geological mapping data provided a good platform to understand the joint network. Major joints including shear zone were incorporated into the model for simulation. This approach proved successful by demonstrating that continuum modelling can be used to evaluate evolution of stresses, strain, plastic yielding and failure mechanisms that are consistent with field observations. Structural control due to geological shear zone structure proved to be important in its location, size and orientation. Furthermore, the model analyzed slope deformation and sliding possibility along shear zone interfaces. This type of approach can predict shear zone deformation and failure mechanism, hence mitigation strategies can be deployed for safety of human lives and property within mine pits.

Keywords: numerical modeling, open pit mine, shear zone, slope stability

Procedia PDF Downloads 285
534 Uncertainty and Multifunctionality as Bridging Concepts from Socio-Ecological Resilience to Infrastructure Finance in Water Resource Decision Making

Authors: Anita Lazurko, Laszlo Pinter, Jeremy Richardson

Abstract:

Uncertain climate projections, multiple possible development futures, and a financing gap create challenges for water infrastructure decision making. In contrast to conventional predict-plan-act methods, an emerging decision paradigm that enables social-ecological resilience supports decisions that are appropriate for uncertainty and leverage social, ecological, and economic multifunctionality. Concurrently, water infrastructure project finance plays a powerful role in sustainable infrastructure development but remains disconnected from discourse in socio-ecological resilience. At the time of research, a project to transfer water from Lesotho to Botswana through South Africa in the Orange-Senqu River Basin was at the pre-feasibility stage. This case was analysed through documents and interviews to investigate how uncertainty and multifunctionality are conceptualised and considered in decisions for the resilience of water infrastructure and to explore bridging concepts that might allow project finance to better enable socio-ecological resilience. Interviewees conceptualised uncertainty as risk, ambiguity and ignorance, and multifunctionality as politically-motivated shared benefits. Numerous efforts to adopt emerging decision methods that consider these terms were in use but required compromises to accommodate the persistent, conventional decision paradigm, though a range of future opportunities was identified. Bridging these findings to finance revealed opportunities to consider a more comprehensive scope of risk, to leverage risk mitigation measures, to diffuse risks and benefits over space, time and to diverse actor groups, and to clarify roles to achieve multiple objectives for resilience. In addition to insights into how multiple decision paradigms interact in real-world decision contexts, the research highlights untapped potential at the juncture between socio-ecological resilience and project finance.

Keywords: socio-ecological resilience, finance, multifunctionality, uncertainty

Procedia PDF Downloads 110
533 Role of Erythrocyte Fatty Acids in Predicting Cardiometabolic Risk among the Elderly: A Secondary Analysis of the Walnut and Healthy Aging Study

Authors: Tony Jehi, Sujatha Rajaram, Nader majzoub, Joan Sabate

Abstract:

Aging significantly increases the incidence of various cardiometabolic diseases, including cardiovascular disease (CVD). To combat CVD and its associated risk factors, it is imperative to adopt a healthy dietary pattern that is rife with beneficial nutrient and non-nutrient compounds. Unsaturated fats, specifically n-3 polyunsaturated fatty acids (n-3 PUFA), have cardio-protective effects; the opposite is true for saturated fatty acids. What role, if any, does the biomarker of fatty acid intake (specific fatty acids in the erythrocyte) play in predicting cardiometabolic risk among the elderly, a population highly susceptible to increased mortality and morbidity from CVD risk factors, remains unclear. This was a secondary analysis of the Walnuts and Healthy Aging Study. Briefly, elderly (n=192, mean age 69 y) participants followed their usual diet and were randomized into two groups to either eat walnuts daily or abstain from eating walnuts for a period of 2 years. The purpose was to identify potential associations between erythrocyte membrane fatty acids and cardiometabolic risk factors (body weight, blood pressure, blood lipids, and fasting glucose). Erythrocyte n-3 PUFA were inversely associated with total cholesterol (ß = -3.83; p= 0.02), triglycerides (ß = -7.66; p= <0.01), and fasting glucose (ß = -0.19; p=0.03). Specifically, erythrocyte ALA (ß= -1.59; P = 0.04) and DPA (ß= -0.62; P=0.04) were inversely associated with diastolic blood pressure and fasting glucose, respectively. N-6 PUFAs were positively associated with systolic blood pressure (ß=1.10; P=0.02). Mono-unsaturated fatty acids were positively associated with TAG (ß = 4.16; P=0.03). Total saturated fatty acids were not associated with any cardiometabolic risk factors. No association was found between any erythrocyte fatty acid and body weight. In conclusion, erythrocyte n-3 PUFA may be used as a biomarker to predict the cardiometabolic risk among healthy elders, providing support for the American Heart Association guidelines for including n-3 PUFA for preventing CVD.

Keywords: cardiometabolic diseases, erythrocyte fatty acids, elderly, n-3 PUFA

Procedia PDF Downloads 61
532 An Artificial Intelligence Framework to Forecast Air Quality

Authors: Richard Ren

Abstract:

Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.

Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms

Procedia PDF Downloads 103
531 Aero-Hydrodynamic Model for a Floating Offshore Wind Turbine

Authors: Beatrice Fenu, Francesco Niosi, Giovanni Bracco, Giuliana Mattiazzo

Abstract:

In recent years, Europe has seen a great development of renewable energy, in a perspective of reducing polluting emissions and transitioning to cleaner forms of energy, as established by the European Green New Deal. Wind energy has come to cover almost 15% of European electricity needs andis constantly growing. In particular, far-offshore wind turbines are attractive from the point of view of exploiting high-speed winds and high wind availability. Considering offshore wind turbine siting that combines the resources analysis, the bathymetry, environmental regulations, and maritime traffic and considering the waves influence in the stability of the platform, the hydrodynamic characteristics of the platform become fundamental for the evaluation of the performances of the turbine, especially for the pitch motion. Many platform's geometries have been studied and used in the last few years. Their concept is based upon different considerations as hydrostatic stability, material, cost and mooring system. A new method to reach a high-performances substructure for different kinds of wind turbines is proposed. The system that considers substructure, mooring, and wind turbine is implemented in Orcaflex, and the simulations are performed considering several sea states and wind speeds. An external dynamic library is implemented for the turbine control system. The study shows the comparison among different substructures and the new concepts developed. In order to validate the model, CFD simulations will be performed by mean of STAR CCM+, and a comparison between rigid and elastic body for what concerns blades and tower will be carried out. A global model will be built to predict the productivity of the floating turbine according to siting, resources, substructure, and mooring. The Levelized Cost of Electricity (LCOE) of the system is estimated, giving a complete overview about the advantages of floating offshore wind turbine plants. Different case studies will be presented.

Keywords: aero-hydrodynamic model, computational fluid dynamics, floating offshore wind, siting, verification, and validation

Procedia PDF Downloads 195
530 Development of Interaction Diagram for Eccentrically Loaded Reinforced Concrete Sandwich Walls with Different Design Parameters

Authors: May Haggag, Ezzat Fahmy, Mohamed Abdel-Mooty, Sherif Safar

Abstract:

Sandwich sections have a very complex nature due to variability of behavior of different materials within the section. Cracking, crushing and yielding capacity of constituent materials enforces high complexity of the section. Furthermore, slippage between the different layers adds to the section complex behavior. Conventional methods implemented in current industrial guidelines do not account for the above complexities. Thus, a throughout study is needed to understand the true behavior of the sandwich panels thus, increase the ability to use them effectively and efficiently. The purpose of this paper is to conduct numerical investigation using ANSYS software for the structural behavior of sandwich wall section under eccentric loading. Sandwich walls studied herein are composed of two RC faces, a foam core and linking shear connectors. Faces are modeled using solid elements and reinforcement together with connectors are modeled using link elements. The analysis conducted herein is nonlinear static analysis incorporating material nonlinearity, crashing and crushing of concrete and yielding of steel. The model is validated by comparing it to test results in literature. After validation, the model is used to establish extensive parametric analysis to investigate the effect of three key parameters on the axial force bending moment interaction diagram of the walls. These parameters are the concrete compressive strength, face thickness and number of shear connectors. Furthermore, the results of the parametric study are used to predict a coefficient that links the interaction diagram of a solid wall to that of a sandwich wall. The equation is predicted using the parametric study data and regression analysis. The predicted α was used to construct the interaction diagram of the investigated wall and the results were compared with ANSYS results and showed good agreement.

Keywords: sandwich walls, interaction diagrams, numerical modeling, eccentricity, reinforced concrete

Procedia PDF Downloads 395
529 Comparison of Multivariate Adaptive Regression Splines and Random Forest Regression in Predicting Forced Expiratory Volume in One Second

Authors: P. V. Pramila , V. Mahesh

Abstract:

Pulmonary Function Tests are important non-invasive diagnostic tests to assess respiratory impairments and provides quantifiable measures of lung function. Spirometry is the most frequently used measure of lung function and plays an essential role in the diagnosis and management of pulmonary diseases. However, the test requires considerable patient effort and cooperation, markedly related to the age of patients esulting in incomplete data sets. This paper presents, a nonlinear model built using Multivariate adaptive regression splines and Random forest regression model to predict the missing spirometric features. Random forest based feature selection is used to enhance both the generalization capability and the model interpretability. In the present study, flow-volume data are recorded for N= 198 subjects. The ranked order of feature importance index calculated by the random forests model shows that the spirometric features FVC, FEF 25, PEF,FEF 25-75, FEF50, and the demographic parameter height are the important descriptors. A comparison of performance assessment of both models prove that, the prediction ability of MARS with the `top two ranked features namely the FVC and FEF 25 is higher, yielding a model fit of R2= 0.96 and R2= 0.99 for normal and abnormal subjects. The Root Mean Square Error analysis of the RF model and the MARS model also shows that the latter is capable of predicting the missing values of FEV1 with a notably lower error value of 0.0191 (normal subjects) and 0.0106 (abnormal subjects). It is concluded that combining feature selection with a prediction model provides a minimum subset of predominant features to train the model, yielding better prediction performance. This analysis can assist clinicians with a intelligence support system in the medical diagnosis and improvement of clinical care.

Keywords: FEV, multivariate adaptive regression splines pulmonary function test, random forest

Procedia PDF Downloads 293
528 A Study on the Relationships among Teacher Empowerment, Professional Commitment and School Effectiveness

Authors: S. C. Lin, W. F. Hung, W. W. Cheng

Abstract:

Teacher empowerment was regarded as investing teachers with the right to participate in the determination of school goals and policies and to exercise professional judgment about what and how to teach. Professional commitment was considered as a person’s belief in and acceptance of the values of his or her chosen occupation or line of work, and a willingness to maintain membership in that occupation. An effective school has been defined as one in which students’ progress further than might be expected from consideration of its intake. An effective school thus adds extra value to its students' outcomes, in comparison with other schools serving similar intakes. A number of literature from various countries explored that teacher empowerment and professional commitment significantly influenced school effectiveness. However, there lacked more empirical studies to examine the relationships among them. Hence, this study was to explore the relationships among teacher empowerment, professional commitment and school effectiveness in junior high schools in Taiwan. Samples were seven hundred and five junior high school teachers selected from Taichung City, Changhua County and Nantou County. Questionnaire was applied to collect data. Data were analyzed by using descriptive statistics, t-test, one-way ANOVA, Pearson’s product-moment correlation, and multiple regression analysis. The findings of this study were as follows: First, the overall performances of teachers’ perceptions of teacher empowerment, teacher professional commitment and school effectiveness were above average. Second, the teachers’ perceptions of teacher empowerment were significant different in gender, designated duty, and school size. Third, the teachers’ perceptions of teacher professional commitment were significant different in gender, designated duty, and school size. Fourth, the teachers’ perceptions of school effectiveness were significant different in designated duty. Fifth, teacher empowerment was mid-positively correlation by teacher professional commitment. Sixth, there was mid-positively correlation between teacher empowerment and school effectiveness. Seventh, there was mid-positively correlation between teacher professional commitment and school effectiveness. Eighth, Teacher empowerment and professional commitment could significantly predict school effectiveness. Based on the findings of this study, the study proposed some suggestions for educational authorities, schools, teachers, and future studies as well.

Keywords: junior high school teacher, teacher empowerment, teacher professional commitment, school effectiveness

Procedia PDF Downloads 445
527 Experimental Studies on Stress Strain Behavior of Expanded Polystyrene Beads-Sand Mixture

Authors: K. N. Ashna

Abstract:

Lightweight fills are a viable alternative where weak soils such as soft clay, peat, and loose silt are encountered. Materials such as Expanded Polystyrene (EPS) geo-foam, plastics, tire wastes, rubber wastes have been used along with soil in order to obtain a lightweight fill. Out of these, Expanded Polystyrene (EPS) geo-foam has gained wide popularity in civil engineering over the past years due to its wide variety of applications. It is extremely lightweight, durable and is available in various densities to meet the strength requirements. It can be used as backfill behind retaining walls to reduce lateral load, as a fill over soft clay or weak soils to prevent the excessive settlements and to reduce seismic forces. Geo-foam is available in block form as well as beads form. In this project Expanded Polystyrene (EPS) beads of various diameters and varying densities were mixed along with sand to study their lightweight as well as strength properties. Four types of EPS beads were used 1mm, 2mm, 3-7 mm and a mix of 1-7 mm. In this project, EPS beads were varied at .25%, .5%, .75% and 1% by weight of sand. A water content of 10% by weight of sand was added to prevent segregation of the mixture. Unconsolidated Unconfined (UU) tri-axial test was conducted at 100kPa, 200 kPa and 300 kPa and angle of internal friction, and cohesion was obtained. Unit weight of the mix was obtained for a relative density of 65%. The results showed that by increasing the EPS content by weight, maximum deviator stress, unit weight, angle of internal friction and initial elastic modulus decreased. An optimum EPS bead content was arrived at by considering the strength as well as the unit weight. The stress-strain behaviour of the mix was found to be dependent on type of bead, bead content and density of the beads. Finally, regression equations were developed to predict the initial elastic modulus of the mix.

Keywords: expanded polystyrene beads, geofoam, lightweight fills, stress-strain behavior, triaxial test

Procedia PDF Downloads 250
526 Applying Semi-Automatic Digital Aerial Survey Technology and Canopy Characters Classification for Surface Vegetation Interpretation of Archaeological Sites

Authors: Yung-Chung Chuang

Abstract:

The cultural layers of archaeological sites are mainly affected by surface land use, land cover, and root system of surface vegetation. For this reason, continuous monitoring of land use and land cover change is important for archaeological sites protection and management. However, in actual operation, on-site investigation and orthogonal photograph interpretation require a lot of time and manpower. For this reason, it is necessary to perform a good alternative for surface vegetation survey in an automated or semi-automated manner. In this study, we applied semi-automatic digital aerial survey technology and canopy characters classification with very high-resolution aerial photographs for surface vegetation interpretation of archaeological sites. The main idea is based on different landscape or forest type can easily be distinguished with canopy characters (e.g., specific texture distribution, shadow effects and gap characters) extracted by semi-automatic image classification. A novel methodology to classify the shape of canopy characters using landscape indices and multivariate statistics was also proposed. Non-hierarchical cluster analysis was used to assess the optimal number of canopy character clusters and canonical discriminant analysis was used to generate the discriminant functions for canopy character classification (seven categories). Therefore, people could easily predict the forest type and vegetation land cover by corresponding to the specific canopy character category. The results showed that the semi-automatic classification could effectively extract the canopy characters of forest and vegetation land cover. As for forest type and vegetation type prediction, the average prediction accuracy reached 80.3%~91.7% with different sizes of test frame. It represented this technology is useful for archaeological site survey, and can improve the classification efficiency and data update rate.

Keywords: digital aerial survey, canopy characters classification, archaeological sites, multivariate statistics

Procedia PDF Downloads 125
525 Predicting the Turbulence Intensity, Excess Energy Available and Potential Power Generated by Building Mounted Wind Turbines over Four Major UK City

Authors: Emejeamara Francis

Abstract:

The future of potentials wind energy applications within suburban/urban areas are currently faced with various problems. These include insufficient assessment of urban wind resource, and the effectiveness of commercial gust control solutions as well as unavailability of effective and cheaper valuable tools for scoping the potentials of urban wind applications within built-up environments. In order to achieve effective assessment of the potentials of urban wind installations, an estimation of the total energy that would be available to them were effective control systems to be used, and evaluating the potential power to be generated by the wind system is required. This paper presents a methodology of predicting the power generated by a wind system operating within an urban wind resource. This method was developed by using high temporal resolution wind measurements from eight potential sites within the urban and suburban environment as inputs to a vertical axis wind turbine multiple stream tube model. A relationship between the unsteady performance coefficient obtained from the stream tube model results and turbulence intensity was demonstrated. Hence, an analytical methodology for estimating the unsteady power coefficient at a potential turbine site is proposed. This is combined with analytical models that were developed to predict the wind speed and the excess energy (EEC) available in estimating the potential power generated by wind systems at different heights within a built environment. Estimates of turbulence intensities, wind speed, EEC and turbine performance based on the current methodology allow a more complete assessment of available wind resource and potential urban wind projects. This methodology is applied to four major UK cities namely Leeds, Manchester, London and Edinburgh and the potential to map the turbine performance at different heights within a typical urban city is demonstrated.

Keywords: small-scale wind, turbine power, urban wind energy, turbulence intensity, excess energy content

Procedia PDF Downloads 260
524 Transcriptome Analysis for Insights into Disease Progression in Dengue Patients

Authors: Abhaydeep Pandey, Shweta Shukla, Saptamita Goswami, Bhaswati Bandyopadhyay, Vishnampettai Ramachandran, Sudhanshu Vrati, Arup Banerjee

Abstract:

Dengue virus infection is now considered as one of the most important mosquito-borne infection in human. The virus is known to promote vascular permeability, cerebral edema leading to Dengue hemorrhagic fever (DHF) or Dengue shock syndrome (DSS). Dengue infection has known to be endemic in India for over two centuries as a benign and self-limited disease. In the last couple of years, the disease symptoms have changed, manifesting severe secondary complication. So far, Delhi has experienced 12 outbreaks of dengue virus infection since 1997 with the last reported in 2014-15. Without specific antivirals, the case management of high-risk dengue patients entirely relies on supportive care, involving constant monitoring and timely fluid support to prevent hypovolemic shock. Nonetheless, the diverse clinical spectrum of dengue disease, as well as its initial similarity to other viral febrile illnesses, presents a challenge in the early identification of this high-risk group. WHO recommends the use of warning signs to identify high-risk patients, but warning signs generally appear during, or just one day before the development of severe illness, thus, providing only a narrow window for clinical intervention. The ability to predict which patient may develop DHF and DSS may improve the triage and treatment. With the recent discovery of high throughput RNA sequencing allows us to understand the disease progression at the genomic level. Here, we will collate the results of RNA-Sequencing data obtained recently from PBMC of different categories of dengue patients from India and will discuss the possible role of deregulated genes and long non-coding RNAs NEAT1 for development of disease progression.

Keywords: long non-coding RNA (lncRNA), dengue, peripheral blood mononuclear cell (PBMC), nuclear enriched abundant transcript 1 (NEAT1), dengue hemorrhagic fever (DHF), dengue shock syndrome (DSS)

Procedia PDF Downloads 296
523 Smoking and Alcohol Consumption Predicts Multiple Head and Neck Cancers

Authors: Kim Kennedy, Daren Gibson, Stephanie Flukes, Chandra Diwakarla, Lisa Spalding, Leanne Pilkington, Andrew Redfern

Abstract:

Introduction: It is well known that patients with Head and Neck Cancer (HNC) are at increased risk of subsequent head and neck cancers due to various aetiologies. Aim: We sought to determine the factors contributing to an increased risk of subsequent HNC primaries, and also to evaluate whether Aboriginal patients are at increased risk. Methods: We performed a retrospective cohort analysis of 320 HNC patients from a single centre in Western Australia, identifying 80 Aboriginal patients and 240 non-Aboriginal patients matched on a 1:3 ratio by site, histology, rurality, and age. We collected patient data including smoking and alcohol consumption, tumour and treatment data, and data on subsequent HNC primaries. Results: A subsequent HNC primary was seen in 37 patients (11.6%) overall. There was no significant difference in the rate of second primary HNCs between Aboriginal patients (12.5%) and nonAboriginal patients (11.2%) (p=0.408). Subsequent HNCs, were strongly associated with smoking and alcohol consumption however, with 95% of patients with a second primary being ever-smokers, and 54% of patients with a second primary having a history of excessive alcohol consumption. In the 37 patients with multiple HNC primaries, there were a total of 57 HNCs, with 29 patients having two primaries, six patients having 3 HNC primaries, one patient with four, and one with six. 54 out of the 57 cancers were in ever smokers (94.7%). There were only two multiple HNC primaries in a never smoker, non-drinker, and these cases were of unknown etiology with HPV/p16 status unknown in both cases. In the whole study population, there were 32 HPV-positive HNCs, and 67 p16-positive HNCs, with only two 2 nd HNCs in a p16-positive case, giving a rate of 3% in the p16+ population, which is actually much lower than the rate of second primaries seen in the overall population (11.6%), and was highest in the p16-negative population (15.7%). This suggests that p16-positivity is not a strong risk factor for subsequent primaries, and in fact p16-negativity appeared to be associated with increased risk, however this data is limited by the large number of patients without documented p16 status (45.3% overall, 12% for oropharyngeal, and 59.6% for oral cavity primaries had unknown p16 status). Summary: Subsequent HNC primaries were strongly associated with smoking and alcohol excess. Second and later HNC primaries did not appear to occur at increased rates in Aboriginal patients compared with non-Aboriginal patients, and p16-positivity did not predict increased risk, however p16-negativity was associated with an increased risk of subsequent HNCs.

Keywords: head and neck cancer, multiple primaries, aboriginal, p16 status, smoking, alcohol

Procedia PDF Downloads 55
522 Performance of AquaCrop Model for Simulating Maize Growth and Yield Under Varying Sowing Dates in Shire Area, North Ethiopia

Authors: Teklay Tesfay, Gebreyesus Brhane Tesfahunegn, Abadi Berhane, Selemawit Girmay

Abstract:

Adjusting the proper sowing date of a crop at a particular location with a changing climate is an essential management option to maximize crop yield. However, determining the optimum sowing date for rainfed maize production through field experimentation requires repeated trials for many years in different weather conditions and crop management. To avoid such long-term experimentation to determine the optimum sowing date, crop models such as AquaCrop are useful. Therefore, the overall objective of this study was to evaluate the performance of AquaCrop model in simulating maize productivity under varying sowing dates. A field experiment was conducted for two consecutive cropping seasons by deploying four maize seed sowing dates in a randomized complete block design with three replications. Input data required to run this model are stored as climate, crop, soil, and management files in the AquaCrop database and adjusted through the user interface. Observed data from separate field experiments was used to calibrate and validate the model. AquaCrop model was validated for its performance in simulating the green canopy and aboveground biomass of maize for the varying sowing dates based on the calibrated parameters. Results of the present study showed that there was a good agreement (an overall R2 =, Ef= d= RMSE =) between measured and simulated values of the canopy cover and biomass yields. Considering the overall values of the statistical test indicators, the performance of the model to predict maize growth and biomass yield was successful, and so this is a valuable tool help for decision-making. Hence, this calibrated and validated model is suggested to use for determining optimum maize crop sowing date for similar climate and soil conditions to the study area, instead of conducting long-term experimentation.

Keywords: AquaCrop model, calibration, validation, simulation

Procedia PDF Downloads 43
521 Experimental Investigation on the Effect of Cross Flow on Discharge Coefficient of an Orifice

Authors: Mathew Saxon A, Aneeh Rajan, Sajeev P

Abstract:

Many fluid flow applications employ different types of orifices to control the flow rate or to reduce the pressure. Discharge coefficients generally vary from 0.6 to 0.95 depending on the type of the orifice. The tabulated value of discharge coefficients of various types of orifices available can be used in most common applications. The upstream and downstream flow condition of an orifice is hardly considered while choosing the discharge coefficient of an orifice. But literature shows that the discharge coefficient can be affected by the presence of cross flow. Cross flow is defined as the condition wherein; a fluid is injected nearly perpendicular to a flowing fluid. Most researchers have worked on water being injected into a cross-flow of water. The present work deals with water to gas systems in which water is injected in a normal direction into a flowing stream of gas. The test article used in the current work is called thermal regulator, which is used in a liquid rocket engine to reduce the temperature of hot gas tapped from the gas generator by injecting water into the hot gas so that a cooler gas can be supplied to the turbine. In a thermal regulator, water is injected through an orifice in a normal direction into the hot gas stream. But the injection orifice had been calibrated under backpressure by maintaining a stagnant gas medium at the downstream. The motivation of the present study aroused due to the observation of a lower Cd of the orifice in flight compared to the calibrated Cd. A systematic experimental investigation is carried out in this paper to study the effect of cross-flow on the discharge coefficient of an orifice in water to a gas system. The study reveals that there is an appreciable reduction in the discharge coefficient with cross flow compared to that without cross flow. It is found that the discharge coefficient greatly depends on the ratio of momentum of water injected to the momentum of the gas cross flow. The effective discharge coefficient of different orifices was normalized using the discharge coefficient without cross-flow and it is observed that normalized curves of effective discharge coefficient of different orifices with momentum ratio collapsing into a single curve. Further, an equation is formulated using the test data to predict the effective discharge coefficient with cross flow using the calibrated Cd value without cross flow.

Keywords: cross flow, discharge coefficient, orifice, momentum ratio

Procedia PDF Downloads 126
520 Early Prediction of Diseases in a Cow for Cattle Industry

Authors: Ghufran Ahmed, Muhammad Osama Siddiqui, Shahbaz Siddiqui, Rauf Ahmad Shams Malick, Faisal Khan, Mubashir Khan

Abstract:

In this paper, a machine learning-based approach for early prediction of diseases in cows is proposed. Different ML algos are applied to extract useful patterns from the available dataset. Technology has changed today’s world in every aspect of life. Similarly, advanced technologies have been developed in livestock and dairy farming to monitor dairy cows in various aspects. Dairy cattle monitoring is crucial as it plays a significant role in milk production around the globe. Moreover, it has become necessary for farmers to adopt the latest early prediction technologies as the food demand is increasing with population growth. This highlight the importance of state-ofthe-art technologies in analyzing how important technology is in analyzing dairy cows’ activities. It is not easy to predict the activities of a large number of cows on the farm, so, the system has made it very convenient for the farmers., as it provides all the solutions under one roof. The cattle industry’s productivity is boosted as the early diagnosis of any disease on a cattle farm is detected and hence it is treated early. It is done on behalf of the machine learning output received. The learning models are already set which interpret the data collected in a centralized system. Basically, we will run different algorithms on behalf of the data set received to analyze milk quality, and track cows’ health, location, and safety. This deep learning algorithm draws patterns from the data, which makes it easier for farmers to study any animal’s behavioral changes. With the emergence of machine learning algorithms and the Internet of Things, accurate tracking of animals is possible as the rate of error is minimized. As a result, milk productivity is increased. IoT with ML capability has given a new phase to the cattle farming industry by increasing the yield in the most cost-effective and time-saving manner.

Keywords: IoT, machine learning, health care, dairy cows

Procedia PDF Downloads 45
519 Evaluation of Flexural Cracking Width of Steel Fibre Reinforced Concrete Beams

Authors: Touhami Tahenni

Abstract:

Excessively wide cracks are harmful to the serviceability of reinforced concrete (RC) beams and may lead to durability problems in the longer term. They also reduce the rigidity of RC sections, rendering the tensile concrete ineffective structurally. To reduce the negative effects of cracks, steel fibers are added to concrete mixes in the same manner as aggregates. In the present work, steel fibers reinforced concrete (SFRC) beams, made of normal strength and high strength concretes, were tested in a four-point bending test using a digital image correlation technique. The beams had different volume fractions of fibres and different aspect ratios (fiber length/fiber diameter). The evaluation of flexural cracking widths was determined using Gom-Aramis software. The experimental crack widths were compared with theoretical values predicted by the technical document of Rilem TC 162-TDF. The model proposed in this document seems to be the only one that considers the efficiency of steel fibres in restraining the crack widths. However, the model of Rilem takes into account only the aspect ratio of steel fibres to predict the crack width of SFRC beams. It has been reported in several pieces of research that the contribution of steel fibres to the limitation of flexural cracking widths is based on three essential parameters namely, the volume fraction, the orientation and the aspect ratio of fibres. Referring to the literature on the flexural cracking behavior of SFRC beams and the experimental observations of the present work, a correction of the Rilem model by the introduction of these parameters in the formula is proposed. The crack widths predicted by the new empirical model were compared with the experimental results and assessed against other test data on SFRC beams taken from the literature. The modified Rilem model gives better results and is found more satisfactory in predicting the crack widths of fibres concrete.

Keywords: stee fibres, reinforced concrete, flexural cracking, tensile strength, crack width

Procedia PDF Downloads 75
518 Analysis of Biomarkers Intractable Epileptogenic Brain Networks with Independent Component Analysis and Deep Learning Algorithms: A Comprehensive Framework for Scalable Seizure Prediction with Unimodal Neuroimaging Data in Pediatric Patients

Authors: Bliss Singhal

Abstract:

Epilepsy is a prevalent neurological disorder affecting approximately 50 million individuals worldwide and 1.2 million Americans. There exist millions of pediatric patients with intractable epilepsy, a condition in which seizures fail to come under control. The occurrence of seizures can result in physical injury, disorientation, unconsciousness, and additional symptoms that could impede children's ability to participate in everyday tasks. Predicting seizures can help parents and healthcare providers take precautions, prevent risky situations, and mentally prepare children to minimize anxiety and nervousness associated with the uncertainty of a seizure. This research proposes a comprehensive framework to predict seizures in pediatric patients by evaluating machine learning algorithms on unimodal neuroimaging data consisting of electroencephalogram signals. The bandpass filtering and independent component analysis proved to be effective in reducing the noise and artifacts from the dataset. Various machine learning algorithms’ performance is evaluated on important metrics such as accuracy, precision, specificity, sensitivity, F1 score and MCC. The results show that the deep learning algorithms are more successful in predicting seizures than logistic Regression, and k nearest neighbors. The recurrent neural network (RNN) gave the highest precision and F1 Score, long short-term memory (LSTM) outperformed RNN in accuracy and convolutional neural network (CNN) resulted in the highest Specificity. This research has significant implications for healthcare providers in proactively managing seizure occurrence in pediatric patients, potentially transforming clinical practices, and improving pediatric care.

Keywords: intractable epilepsy, seizure, deep learning, prediction, electroencephalogram channels

Procedia PDF Downloads 69
517 Comparison of Receiver Operating Characteristic Curve Smoothing Methods

Authors: D. Sigirli

Abstract:

The Receiver Operating Characteristic (ROC) curve is a commonly used statistical tool for evaluating the diagnostic performance of screening and diagnostic test with continuous or ordinal scale results which aims to predict the presence or absence probability of a condition, usually a disease. When the test results were measured as numeric values, sensitivity and specificity can be computed across all possible threshold values which discriminate the subjects as diseased and non-diseased. There are infinite numbers of possible decision thresholds along the continuum of the test results. The ROC curve presents the trade-off between sensitivity and the 1-specificity as the threshold changes. The empirical ROC curve which is a non-parametric estimator of the ROC curve is robust and it represents data accurately. However, especially for small sample sizes, it has a problem of variability and as it is a step function there can be different false positive rates for a true positive rate value and vice versa. Besides, the estimated ROC curve being in a jagged form, since the true ROC curve is a smooth curve, it underestimates the true ROC curve. Since the true ROC curve is assumed to be smooth, several smoothing methods have been explored to smooth a ROC curve. These include using kernel estimates, using log-concave densities, to fit parameters for the specified density function to the data with the maximum-likelihood fitting of univariate distributions or to create a probability distribution by fitting the specified distribution to the data nd using smooth versions of the empirical distribution functions. In the present paper, we aimed to propose a smooth ROC curve estimation based on the boundary corrected kernel function and to compare the performances of ROC curve smoothing methods for the diagnostic test results coming from different distributions in different sample sizes. We performed simulation study to compare the performances of different methods for different scenarios with 1000 repetitions. It is seen that the performance of the proposed method was typically better than that of the empirical ROC curve and only slightly worse compared to the binormal model when in fact the underlying samples were generated from the normal distribution.

Keywords: empirical estimator, kernel function, smoothing, receiver operating characteristic curve

Procedia PDF Downloads 138
516 Evaluation of Stress Relief using Ultrasonic Peening in GTAW Welding and Stress Corrosion Cracking (SCC) in Stainless Steel, and Comparison with the Thermal Method

Authors: Hamidreza Mansouri

Abstract:

In the construction industry, the lifespan of a metal structure is directly related to the quality of welding. In most metal structures, the welded area is considered critical and is one of the most important factors in design. To date, many fracture incidents caused by these types of cracks have occurred. Various methods exist to increase the lifespan of welds to prevent failure in the welded area. Among these methods, the application of ultrasonic peening, in addition to the stress relief process, can manually and more precisely adjust the geometry of the weld toe and prevent stress concentration in this part. This research examined Gas Tungsten Arc Welding (GTAW) on common structural steels and 316 stainless steel, which require precise welding, to predict the optimal condition. The GTAW method was used to create residual stress; two samples underwent ultrasonic stress relief, and for comparison, two samples underwent thermal stress relief. Also, no treatment was considered for two samples. The residual stress of all six pieces was measured by X-Ray Diffraction (XRD) method. Then, the two ultrasonically stress-relieved samples and two untreated samples were exposed to a corrosive environment to initiate cracking and determine the effectiveness of the ultrasonic stress relief method. Thus, the residual stress caused by GTAW in the samples decreased by 3.42% with thermal treatment and by 7.69% with ultrasonic peening. Furthermore, the results show that the untreated sample developed cracks after 740 hours, while the ultrasonically stress-relieved piece showed no cracks. Given the high costs of welding and post-welding zone modification processes, finding an economical, effective, and comprehensive method that has the least limitations alongside a broad spectrum of usage is of great importance. Therefore, the impact of various ultrasonic peening stress relief parameters and the selection of the best stress relief parameter to achieve the longest lifespan for the weld area is highly significant.

Keywords: GTAW welding, stress corrosion cracking(SCC), thermal method, ultrasonic peening.

Procedia PDF Downloads 30
515 Performance of Reinforced Concrete Beams under Different Fire Durations

Authors: Arifuzzaman Nayeem, Tafannum Torsha, Tanvir Manzur, Shaurav Alam

Abstract:

Performance evaluation of reinforced concrete (RC) beams subjected to accidental fire is significant for post-fire capacity measurement. Mechanical properties of any RC beam degrade due to heating since the strength and modulus of concrete and reinforcement suffer considerable reduction under elevated temperatures. Moreover, fire-induced thermal dilation and shrinkage cause internal stresses within the concrete and eventually result in cracking, spalling, and loss of stiffness, which ultimately leads to lower service life. However, conducting full-scale comprehensive experimental investigation for RC beams exposed to fire is difficult and cost-intensive, where the finite element (FE) based numerical study can provide an economical alternative for evaluating the post-fire capacity of RC beams. In this study, an attempt has been made to study the fire behavior of RC beams using FE software package ABAQUS under different durations of fire. The damaged plasticity model of concrete in ABAQUS was used to simulate behavior RC beams. The effect of temperature on strength and modulus of concrete and steel was simulated following relevant Eurocodes. Initially, the result of FE models was validated using several experimental results from available scholarly articles. It was found that the response of the developed FE models matched quite well with the experimental outcome for beams without heat. The FE analysis of beams subjected to fire showed some deviation from the experimental results, particularly in terms of stiffness degradation. However, the ultimate strength and deflection of FE models were similar to that of experimental values. The developed FE models, thus, exhibited the good potential to predict the fire behavior of RC beams. Once validated, FE models were then used to analyze several RC beams having different strengths (ranged between 20 MPa and 50 MPa) exposed to the standard fire curve (ASTM E119) for different durations. The post-fire performance of RC beams was investigated in terms of load-deflection behavior, flexural strength, and deflection characteristics.

Keywords: fire durations, flexural strength, post fire capacity, reinforced concrete beam, standard fire

Procedia PDF Downloads 124
514 Comparative Study of Water Quality Parameters in the Proximity of Various Landfills Sites in India

Authors: Abhishek N. Srivastava, Rahul Singh, Sumedha Chakma

Abstract:

The rapid urbanization in the developing countries is generating an enormous amount of waste leading to the creation of unregulated landfill sites at various places at its disposal. The liquid waste, known as leachate, produced from these landfills sites is severely affecting the surrounding water quality. The water quality in the proximity areas of the landfill is found affected by various physico-chemical parameters of leachate such as pH, alkalinity, total hardness, conductivity, chloride, total dissolved solids (TDS), total suspended solids (TSS), sulphate, nitrate, phosphate, fluoride, sodium and potassium, biological parameters such as biochemical oxygen demand (BOD), chemical oxygen demand (COD), Faecal coliform, and heavy metals such as cadmium (Cd), lead (Pb), iron (Fe), mercury (Hg), arsenic (As), cobalt (Co), manganese (Mn), zinc (Zn), copper (Cu), chromium (Cr), nickel (Ni). However, all these parameters are distributive in leachate that produced according to the nature of waste being dumped at various landfill sites, therefore, it becomes very difficult to predict the main responsible parameter of leachate for water quality contamination. The present study is endeavour the comparative analysis of the physical, chemical and biological parameters of various landfills in India viz. Okhla landfill, Ghazipur landfill, Bhalswa ladfill in NCR Delhi, Deonar landfill in Mumbai, Dhapa landfill in Kolkata and Kodungayaiyur landfill, Perungudi landfill in Chennai. The statistical analysis of the parameters was carried out using the Statistical Packages for the Social Sciences (SPSS) and LandSim 2.5 model to simulate the long term effect of various parameters on different time scale. Further, the uncertainties characterization of various input parameters has also been analysed using fuzzy alpha cut (FAC) technique to check the sensitivity of various water quality parameters at the proximity of numerous landfill sites. Finally, the study would help to suggest the best method for the prevention of pollution migration from the landfill sites on priority basis.

Keywords: landfill leachate, water quality, LandSim, fuzzy alpha cut

Procedia PDF Downloads 114
513 Numerical Methodology to Support the Development of a Double Chamber Syringe

Authors: Lourenço Bastos, Filipa Carneiro, Bruno Vale, Rita Marques Joana Silva, Ricardo Freitas, Ângelo Marques, Sara Cortez, Alberta Coelho, Pedro Parreira, Liliana Sousa, Anabela Salgueiro, Bruno Silva

Abstract:

The process of flushing is considered to be an adequate technique to reduce the risk of infection during the clinical practice of venous catheterization. Nonetheless, there is still a lack of adhesion to this method, in part due to the complexity of this procedure. The project SeringaDuo aimed to develop an innovative double-chamber syringe for intravenous sequential administration of drugs and serums. This device served the purpose of improving the adherence to the practice, through the reduction of manipulations needed, which also improves patient safety, and though the promotion of flushing practice by health professionals, by simplifying this task. To assist on the development of this innovative syringe, a numerical methodology was developed and validated in order to predict the syringe’s mechanical and flow behavior during the fluids’ loading and administration phases, as well as to allow the material behavior evaluation during its production. For this, three commercial numerical simulation software was used, namely ABAQUS, ANSYS/FLUENT, and MOLDFLOW. This methodology aimed to evaluate the concepts feasibility and to optimize the geometries of the syringe’s components, creating this way an iterative process for product development based on numerical simulations, validated by the production of prototypes. Through this methodology, it was possible to achieve a final design that fulfils all the characteristics and specifications defined. This iterative process based on numerical simulations is a powerful tool for product development that allows obtaining fast and accurate results without the strict need for prototypes. An iterative process can be implemented, consisting of consecutive constructions and evaluations of new concepts, to obtain an optimized solution, which fulfils all the predefined specifications and requirements.

Keywords: Venous catheterization, flushing, syringe, numerical simulation

Procedia PDF Downloads 147
512 Comparison of Water Equivalent Ratio of Several Dosimetric Materials in Proton Therapy Using Monte Carlo Simulations and Experimental Data

Authors: M. R. Akbari , H. Yousefnia, E. Mirrezaei

Abstract:

Range uncertainties of protons are currently a topic of interest in proton therapy. Two of the parameters that are often used to specify proton range are water equivalent thickness (WET) and water equivalent ratio (WER). Since WER values for a specific material is nearly constant at different proton energies, it is a more useful parameter to compare. In this study, WER values were calculated for different proton energies in polymethyl methacrylate (PMMA), polystyrene (PS) and aluminum (Al) using FLUKA and TRIM codes. The results were compared with analytical, experimental and simulated SEICS code data obtained from the literature. In FLUKA simulation, a cylindrical phantom, 1000 mm in height and 300 mm in diameter, filled with the studied materials was simulated. A typical mono-energetic proton pencil beam in a wide range of incident energies usually applied in proton therapy (50 MeV to 225 MeV) impinges normally on the phantom. In order to obtain the WER values for the considered materials, cylindrical detectors, 1 mm in height and 20 mm in diameter, were also simulated along the beam trajectory in the phantom. In TRIM calculations, type of projectile, energy and angle of incidence, type of target material and thickness should be defined. The mode of 'detailed calculation with full damage cascades' was selected for proton transport in the target material. The biggest difference in WER values between the codes was 3.19%, 1.9% and 0.67% for Al, PMMA and PS, respectively. In Al and PMMA, the biggest difference between each code and experimental data was 1.08%, 1.26%, 2.55%, 0.94%, 0.77% and 0.95% for SEICS, FLUKA and SRIM, respectively. FLUKA and SEICS had the greatest agreement (≤0.77% difference in PMMA and ≤1.08% difference in Al, respectively) with the available experimental data in this study. It is concluded that, FLUKA and TRIM codes have capability for Bragg curves simulation and WER values calculation in the studied materials. They can also predict Bragg peak location and range of proton beams with acceptable accuracy.

Keywords: water equivalent ratio, dosimetric materials, proton therapy, Monte Carlo simulations

Procedia PDF Downloads 304
511 Experimental Verification of Similarity Criteria for Sound Absorption of Perforated Panels

Authors: Aleksandra Majchrzak, Katarzyna Baruch, Monika Sobolewska, Bartlomiej Chojnacki, Adam Pilch

Abstract:

Scaled modeling is very common in the areas of science such as aerodynamics or fluid mechanics, since defining characteristic numbers enables to determine relations between objects under test and their models. In acoustics, scaled modeling is aimed mainly at investigation of room acoustics, sound insulation and sound absorption phenomena. Despite such a range of application, there is no method developed that would enable scaling acoustical perforated panels freely, maintaining their sound absorption coefficient in a desired frequency range. However, conducted theoretical and numerical analyses have proven that it is not physically possible to obtain given sound absorption coefficient in a desired frequency range by directly scaling only all of the physical dimensions of a perforated panel, according to a defined characteristic number. This paper is a continuation of the research mentioned above and presents practical evaluation of theoretical and numerical analyses. The measurements of sound absorption coefficient of perforated panels were performed in order to verify previous analyses and as a result find the relations between full-scale perforated panels and their models which will enable to scale them properly. The measurements were conducted in a one-to-eight model of a reverberation chamber of Technical Acoustics Laboratory, AGH. Obtained results verify theses proposed after theoretical and numerical analyses. Finding the relations between full-scale and modeled perforated panels will allow to produce measurement samples equivalent to the original ones. As a consequence, it will make the process of designing acoustical perforated panels easier and will also lower the costs of prototypes production. Having this knowledge, it will be possible to emulate in a constructed model panels used, or to be used, in a full-scale room more precisely and as a result imitate or predict the acoustics of a modeled space more accurately.

Keywords: characteristic numbers, dimensional analysis, model study, scaled modeling, sound absorption coefficient

Procedia PDF Downloads 180
510 AI for Efficient Geothermal Exploration and Utilization

Authors: Velimir "monty" Vesselinov, Trais Kliplhuis, Hope Jasperson

Abstract:

Artificial intelligence (AI) is a powerful tool in the geothermal energy sector, aiding in both exploration and utilization. Identifying promising geothermal sites can be challenging due to limited surface indicators and the need for expensive drilling to confirm subsurface resources. Geothermal reservoirs can be located deep underground and exhibit complex geological structures, making traditional exploration methods time-consuming and imprecise. AI algorithms can analyze vast datasets of geological, geophysical, and remote sensing data, including satellite imagery, seismic surveys, geochemistry, geology, etc. Machine learning algorithms can identify subtle patterns and relationships within this data, potentially revealing hidden geothermal potential in areas previously overlooked. To address these challenges, a SIML (Science-Informed Machine Learning) technology has been developed. SIML methods are different from traditional ML techniques. In both cases, the ML models are trained to predict the spatial distribution of an output (e.g., pressure, temperature, heat flux) based on a series of inputs (e.g., permeability, porosity, etc.). The traditional ML (a) relies on deep and wide neural networks (NNs) based on simple algebraic mappings to represent complex processes. In contrast, the SIML neurons incorporate complex mappings (including constitutive relationships and physics/chemistry models). This results in ML models that have a physical meaning and satisfy physics laws and constraints. The prototype of the developed software, called GeoTGO, is accessible through the cloud. Our software prototype demonstrates how different data sources can be made available for processing, executed demonstrative SIML analyses, and presents the results in a table and graphic form.

Keywords: science-informed machine learning, artificial inteligence, exploration, utilization, hidden geothermal

Procedia PDF Downloads 27