Search results for: interactive models
6146 Artificial Intelligence Applications in Kahoot!
Authors: Jana, Walah, Salma, Dareen
Abstract:
This study looks at how the game-based learning platform Kahoot! has changed education, with a particular emphasis on how it incorporates artificial intelligence (AI). From humanly made questions to AI-driven features that improve the learning process, Kahoot! has changed since its 2013 introduction. The software successfully engages educators and students by delivering adaptive learning paths, regulating content, and offering individualized tests. This study also highlights the AI features of Kahoot! by contrasting it with comparable platforms like Quizizz, Socrative, Gimkit, and Nearpod. User satisfaction with Kahoot!'s "PDF to Story" and "Story Text Enhancer" functions ranges from moderate to high, according to a review of user input; yet, there are still issues with consistent accuracy and usability. The results demonstrate how AI can improve learning's effectiveness, adaptability, and interactivity while offering useful insights for educators and developers seeking to optimize educational tools.Keywords: PDF to story feature, story text enhancer, AI-driven learning, interactive content creation
Procedia PDF Downloads 126145 Analyzing Bridge Response to Wind Loads and Optimizing Design for Wind Resistance and Stability
Authors: Abdul Haq
Abstract:
The goal of this research is to better understand how wind loads affect bridges and develop strategies for designing bridges that are more stable and resistant to wind. The effect of wind on bridges is essential to their safety and functionality, especially in areas that are prone to high wind speeds or violent wind conditions. The study looks at the aerodynamic forces and vibrations caused by wind and how they affect bridge construction. Part of the research method involves first understanding the underlying ideas influencing wind flow near bridges. Computational fluid dynamics (CFD) simulations are used to model and forecast the aerodynamic behaviour of bridges under different wind conditions. These models incorporate several factors, such as wind directionality, wind speed, turbulence intensity, and the influence of nearby structures or topography. The results provide significant new insights into the loads and pressures that wind places on different bridge elements, such as decks, pylons, and connections. Following the determination of the wind loads, the structural response of bridges is assessed. By simulating their dynamic behavior under wind-induced forces, Finite Element Analysis (FEA) is used to model the bridge's component parts. This work contributes to the understanding of which areas are at risk of experiencing excessive stresses, vibrations, or oscillations due to wind excitations. Because the bridge has inherent modes and frequencies, the study considers both static and dynamic responses. Various strategies are examined to maximize the design of bridges to withstand wind. It is possible to alter the bridge's geometry, add aerodynamic components, add dampers or tuned mass dampers to lessen vibrations, and boost structural rigidity. Through an analysis of several design modifications and their effectiveness, the study aims to offer guidelines and recommendations for wind-resistant bridge design. In addition to the numerical simulations and analyses, there are experimental studies. In order to assess the computational models and validate the practicality of proposed design strategies, scaled bridge models are tested in a wind tunnel. These investigations help to improve numerical models and prediction precision by providing valuable information on wind-induced forces, pressures, and flow patterns. Using a combination of numerical models, actual testing, and long-term performance evaluation, the project aims to offer practical insights and recommendations for building wind-resistant bridges that are secure, long-lasting, and comfortable for users.Keywords: wind effects, aerodynamic forces, computational fluid dynamics, finite element analysis
Procedia PDF Downloads 696144 Comparative Review of Models for Forecasting Permanent Deformation in Unbound Granular Materials
Authors: Shamsulhaq Amin
Abstract:
Unbound granular materials (UGMs) are pivotal in ensuring long-term quality, especially in the layers under the surface of flexible pavements and other constructions. This study seeks to better understand the behavior of the UGMs by looking at popular models for predicting lasting deformation under various levels of stresses and load cycles. These models focus on variables such as the number of load cycles, stress levels, and features specific to materials and were evaluated on the basis of their ability to accurately predict outcomes. The study showed that these factors play a crucial role in how well the models work. Therefore, the research highlights the need to look at a wide range of stress situations to more accurately predict how much the UGMs bend or shift. The research looked at important factors, like how permanent deformation relates to the number of times a load is applied, how quickly this phenomenon happens, and the shakedown effect, in two different types of UGMs: granite and limestone. A detailed study was done over 100,000 load cycles, which provided deep insights into how these materials behave. In this study, a number of factors, such as the level of stress applied, the number of load cycles, the density of the material, and the moisture present were seen as the main factors affecting permanent deformation. It is vital to fully understand these elements for better designing pavements that last long and handle wear and tear. A series of laboratory tests were performed to evaluate the mechanical properties of materials and acquire model parameters. The testing included gradation tests, CBR tests, and Repeated load triaxial tests. The repeated load triaxial tests were crucial for studying the significant components that affect deformation. This test involved applying various stress levels to estimate model parameters. In addition, certain model parameters were established by regression analysis, and optimization was conducted to improve outcomes. Afterward, the material parameters that were acquired were used to construct graphs for each model. The graphs were subsequently compared to the outcomes obtained from the repeated load triaxial testing. Additionally, the models were evaluated to determine if they demonstrated the two inherent deformation behaviors of materials when subjected to repetitive load: the initial phase, post-compaction, and the second phase volumetric changes. In this study, using log-log graphs was key to making the complex data easier to understand. This method made the analysis clearer and helped make the findings easier to interpret, adding both precision and depth to the research. This research provides important insight into picking the right models for predicting how these materials will act under expected stress and load conditions. Moreover, it offers crucial information regarding the effect of load cycle and permanent deformation as well as the shakedown effect on granite and limestone UGMs.Keywords: permanent deformation, unbound granular materials, load cycles, stress level
Procedia PDF Downloads 446143 The Combination of the Mel Frequency Cepstral Coefficients (MFCC), Perceptual Linear Prediction (PLP), JITTER and SHIMMER Coefficients for the Improvement of Automatic Recognition System for Dysarthric Speech
Authors: Brahim-Fares Zaidi, Malika Boudraa, Sid-Ahmed Selouani
Abstract:
Our work aims to improve our Automatic Recognition System for Dysarthria Speech (ARSDS) based on the Hidden Models of Markov (HMM) and the Hidden Markov Model Toolkit (HTK) to help people who are sick. With pronunciation problems, we applied two techniques of speech parameterization based on Mel Frequency Cepstral Coefficients (MFCC's) and Perceptual Linear Prediction (PLP's) and concatenated them with JITTER and SHIMMER coefficients in order to increase the recognition rate of a dysarthria speech. For our tests, we used the NEMOURS database that represents speakers with dysarthria and normal speakers.Keywords: hidden Markov model toolkit (HTK), hidden models of Markov (HMM), Mel-frequency cepstral coefficients (MFCC), perceptual linear prediction (PLP’s)
Procedia PDF Downloads 1656142 A Review of Research on Pre-training Technology for Natural Language Processing
Authors: Moquan Gong
Abstract:
In recent years, with the rapid development of deep learning, pre-training technology for natural language processing has made great progress. The early field of natural language processing has long used word vector methods such as Word2Vec to encode text. These word vector methods can also be regarded as static pre-training techniques. However, this context-free text representation brings very limited improvement to subsequent natural language processing tasks and cannot solve the problem of word polysemy. ELMo proposes a context-sensitive text representation method that can effectively handle polysemy problems. Since then, pre-training language models such as GPT and BERT have been proposed one after another. Among them, the BERT model has significantly improved its performance on many typical downstream tasks, greatly promoting the technological development in the field of natural language processing, and has since entered the field of natural language processing. The era of dynamic pre-training technology. Since then, a large number of pre-trained language models based on BERT and XLNet have continued to emerge, and pre-training technology has become an indispensable mainstream technology in the field of natural language processing. This article first gives an overview of pre-training technology and its development history, and introduces in detail the classic pre-training technology in the field of natural language processing, including early static pre-training technology and classic dynamic pre-training technology; and then briefly sorts out a series of enlightening technologies. Pre-training technology, including improved models based on BERT and XLNet; on this basis, analyze the problems faced by current pre-training technology research; finally, look forward to the future development trend of pre-training technology.Keywords: natural language processing, pre-training, language model, word vectors
Procedia PDF Downloads 636141 Virtual Reality Learning Environment in Embryology Education
Authors: Salsabeel F. M. Alfalah, Jannat F. Falah, Nadia Muhaidat, Amjad Hudaib, Diana Koshebye, Sawsan AlHourani
Abstract:
Educational technology is changing the way how students engage and interact with learning materials. This improved the learning process amongst various subjects. Virtual Reality (VR) applications are considered one of the evolving methods that have contributed to enhancing medical education. This paper utilizes VR to provide a solution to improve the delivery of the subject of Embryology to medical students, and facilitate the teaching process by providing a useful aid to lecturers, whilst proving the effectiveness of this new technology in this particular area. After evaluating the current teaching methods and identifying students ‘needs, a VR system was designed that demonstrates in an interactive fashion the development of the human embryo from fertilization to week ten of intrauterine development. This system aims to overcome some of the problems faced by the students’ in the current educational methods, and to increase the efficacy of the learning process.Keywords: virtual reality, student assessment, medical education, 3D, embryology
Procedia PDF Downloads 1936140 An Alternative Richards’ Growth Model Based on Hyperbolic Sine Function
Authors: Samuel Oluwafemi Oyamakin, Angela Unna Chukwu
Abstract:
Richrads growth equation being a generalized logistic growth equation was improved upon by introducing an allometric parameter using the hyperbolic sine function. The integral solution to this was called hyperbolic Richards growth model having transformed the solution from deterministic to a stochastic growth model. Its ability in model prediction was compared with the classical Richards growth model an approach which mimicked the natural variability of heights/diameter increment with respect to age and therefore provides a more realistic height/diameter predictions using the coefficient of determination (R2), Mean Absolute Error (MAE) and Mean Square Error (MSE) results. The Kolmogorov-Smirnov test and Shapiro-Wilk test was also used to test the behavior of the error term for possible violations. The mean function of top height/Dbh over age using the two models under study predicted closely the observed values of top height/Dbh in the hyperbolic Richards nonlinear growth models better than the classical Richards growth model.Keywords: height, diameter at breast height, DBH, hyperbolic sine function, Pinus caribaea, Richards' growth model
Procedia PDF Downloads 3976139 The Confluence between Autism Spectrum Disorder and the Schizoid Personality
Authors: Murray David Schane
Abstract:
Though years of clinical encounters with patients with autism spectrum disorders and those with a schizoid personality the many defining diagnostic features shared between these conditions have been explored and current neurobiological differences have been reviewed; and, critical and different treatment strategies for each have been devised. The paper compares and contrasts the apparent similarities between autism spectrum disorders and the schizoid personality are found in these DSM descriptive categories: restricted range of social-emotional reciprocity; poor non-verbal communicative behavior in social interactions; difficulty developing and maintaining relationships; detachment from social relationships; lack of the desire for or enjoyment of close relationships; and preference for solitary activities. In this paper autism, fundamentally a communicative disorder, is revealed to present clinically as a pervasive aversive response to efforts to engage with or be engaged by others. Autists with the Asperger presentation typically have language but have difficulty understanding humor, irony, sarcasm, metaphoric speech, and even narratives about social relationships. They also tend to seek sameness, possibly to avoid problems of social interpretation. Repetitive behaviors engage many autists as a screen against ambient noise, social activity, and challenging interactions. Also in this paper, the schizoid personality is revealed as a pattern of social avoidance, self-sufficiency and apparent indifference to others as a complex psychological defense against a deep, long-abiding fear of appropriation and perverse manipulation. Neither genetic nor MRI studies have yet located the explanatory data that identifies the cause or the neurobiology of autism. Similarly, studies of the schizoid have yet to group that condition with those found in schizophrenia. Through presentations of clinical examples, the treatment of autists of the Asperger type is revealed to address the autist’s extreme social aversion which also precludes the experience of empathy. Autists will be revealed as forming social attachments but without the capacity to interact with mutual concern. Empathy will be shown be teachable and, as social avoidance relents, understanding of the meaning and signs of empathic needs that autists can recognize and acknowledge. Treatment of schizoids will be shown to revolve around joining empathically with the schizoid’s apprehensions about interpersonal, interactive proximity. Models of both autism and schizoid personality traits have yet to be replicated in animals, thereby eliminating the role of translational research in providing the kind of clues to behavioral patterns that can be related to genetic, epigenetic and neurobiological measures. But as these clinical examples will attest, treatment strategies have significant impact.Keywords: autism spectrum, schizoid personality traits, neurobiological implications, critical diagnostic distinctions
Procedia PDF Downloads 1166138 Study on 3D FE Analysis on Normal and Osteoporosis Mouse Models Based on 3-Point Bending Tests
Authors: Tae-min Byun, Chang-soo Chon, Dong-hyun Seo, Han-sung Kim, Bum-mo Ahn, Hui-suk Yun, Cheolwoong Ko
Abstract:
In this study, a 3-point bending computational analysis of normal and osteoporosis mouse models was performed based on the Micro-CT image information of the femurs. The finite element analysis (FEA) found 1.68 N (normal group) and 1.39 N (osteoporosis group) in the average maximum force, and 4.32 N/mm (normal group) and 3.56 N/mm (osteoporosis group) in the average stiffness. In the comparison of the 3-point bending test results, the maximum force and the stiffness were different about 9.4 times in the normal group and about 11.2 times in the osteoporosis group. The difference between the analysis and the test was greatly significant and this result demonstrated improvement points of the material properties applied to the computational analysis of this study. For the next study, the material properties of the mouse femur will be supplemented through additional computational analysis and test.Keywords: 3-point bending test, mouse, osteoporosis, FEA
Procedia PDF Downloads 3556137 The Role of Creative Entrepreneurship in the Development of Croatian Economy
Authors: Marko Kolakovic
Abstract:
Creative industries are an important sector of growth and development of knowledge economies. They have a positive impact on employment, economic growth, export and the quality of life in the areas where they are developed. Creative sectors include architecture, design, advertising, publishing, music, film, television and radio, video games, visual and performing arts and heritage. Following the positive trends of development of creative industries on the global and European level, this paper analyzes creative industries in general and specific characteristics of creative entrepreneurship. Special focus in this paper is put on the influence of the information communication technology on the development of new creative business models and protection of the intellectual property rights. One part of the paper is oriented on the analysis of the status of creative industries and creative entrepreneurship in Croatia. The main objective of the paper is by using the statistical analysis of creative industries in Croatia and information gained during the interviews with entrepreneurs, to make conclusions about potentials and development of creative industries in Croatia. Creative industries in Croatia are at the beginning of their development and growth strategy still does not exist at the national level. Statistical analysis pointed out that in 2015 creative enterprises made 9% of all enterprises in Croatia, employed 5,5% of employed people and their share in GDP was 4,01%. Croatian creative entrepreneurs are building competitive advantage using their creative resources and creating specific business models. The main obstacles they meet are lack of business experience and impossibility of focusing on the creative activities only. In their business, they use digital technologies and are focused on export. The conclusion is that creative industries in Croatia have development potential, but it is necessary to take adequate measures to use this potential in a right way.Keywords: creative entrepreneurship, knowledge economy, business models, intellectual property
Procedia PDF Downloads 2116136 Reservoir Potential, Net Pay Zone and 3D Modeling of Cretaceous Clastic Reservoir in Eastern Sulieman Belt Pakistan
Authors: Hadayat Ullah, Pervez Khalid, Saad Ahmed Mashwani, Zaheer Abbasi, Mubashir Mehmood, Muhammad Jahangir, Ehsan ul Haq
Abstract:
The aim of the study is to explore subsurface structures through data that is acquired from the seismic survey to delineate the characteristics of the reservoir through petrophysical analysis. Ghazij Shale of Eocene age is regional seal rock in this field. In this research work, 3D property models of subsurface were prepared by applying Petrel software to identify various lithologies and reservoir fluids distribution throughout the field. The 3D static modeling shows a better distribution of the discrete and continuous properties in the field. This model helped to understand the reservoir properties and enhance production by selecting the best location for future drilling. A complete workflow is proposed for formation evaluation, electrofacies modeling, and structural interpretation of the subsurface geology. Based on the wireline logs, it is interpreted that the thickness of the Pab Sandstone varies from 250 m to 350 m in the entire study area. The sandstone is massive with high porosity and intercalated layers of shales. Faulted anticlinal structures are present in the study area, which are favorable for the accumulation of hydrocarbon. 3D structural models and various seismic attribute models were prepared to analyze the reservoir character of this clastic reservoir. Based on wireline logs and seismic data, clean sand, shaly sand, and shale are marked as dominant facies in the study area. However, clean sand facies are more favorable to act as a potential net pay zone.Keywords: cretaceous, pab sandstone, petrophysics, electrofacies, hydrocarbon
Procedia PDF Downloads 1466135 A Study on Game Theory Approaches for Wireless Sensor Networks
Authors: M. Shoukath Ali, Rajendra Prasad Singh
Abstract:
Game Theory approaches and their application in improving the performance of Wireless Sensor Networks (WSNs) are discussed in this paper. The mathematical modeling and analysis of WSNs may have low success rate due to the complexity of topology, modeling, link quality, etc. However, Game Theory is a field, which can efficiently use to analyze the WSNs. Game Theory is related to applied mathematics that describes and analyzes interactive decision situations. Game theory has the ability to model independent, individual decision makers whose actions affect the surrounding decision makers. The outcome of complex interactions among rational entities can be predicted by a set of analytical tools. However, the rationality demands a stringent observance to a strategy based on measured of perceived results. Researchers are adopting game theory approaches to model and analyze leading wireless communication networking issues, which includes QoS, power control, resource sharing, etc.Keywords: wireless sensor network, game theory, cooperative game theory, non-cooperative game theory
Procedia PDF Downloads 4366134 Recommendations for Data Quality Filtering of Opportunistic Species Occurrence Data
Authors: Camille Van Eupen, Dirk Maes, Marc Herremans, Kristijn R. R. Swinnen, Ben Somers, Stijn Luca
Abstract:
In ecology, species distribution models are commonly implemented to study species-environment relationships. These models increasingly rely on opportunistic citizen science data when high-quality species records collected through standardized recording protocols are unavailable. While these opportunistic data are abundant, uncertainty is usually high, e.g., due to observer effects or a lack of metadata. Data quality filtering is often used to reduce these types of uncertainty in an attempt to increase the value of studies relying on opportunistic data. However, filtering should not be performed blindly. In this study, recommendations are built for data quality filtering of opportunistic species occurrence data that are used as input for species distribution models. Using an extensive database of 5.7 million citizen science records from 255 species in Flanders, the impact on model performance was quantified by applying three data quality filters, and these results were linked to species traits. More specifically, presence records were filtered based on record attributes that provide information on the observation process or post-entry data validation, and changes in the area under the receiver operating characteristic (AUC), sensitivity, and specificity were analyzed using the Maxent algorithm with and without filtering. Controlling for sample size enabled us to study the combined impact of data quality filtering, i.e., the simultaneous impact of an increase in data quality and a decrease in sample size. Further, the variation among species in their response to data quality filtering was explored by clustering species based on four traits often related to data quality: commonness, popularity, difficulty, and body size. Findings show that model performance is affected by i) the quality of the filtered data, ii) the proportional reduction in sample size caused by filtering and the remaining absolute sample size, and iii) a species ‘quality profile’, resulting from a species classification based on the four traits related to data quality. The findings resulted in recommendations on when and how to filter volunteer generated and opportunistically collected data. This study confirms that correctly processed citizen science data can make a valuable contribution to ecological research and species conservation.Keywords: citizen science, data quality filtering, species distribution models, trait profiles
Procedia PDF Downloads 2086133 Estimation Atmospheric parameters for Weather Study and Forecast over Equatorial Regions Using Ground-Based Global Position System
Authors: Asmamaw Yehun, Tsegaye Kassa, Addisu Hunegnaw, Martin Vermeer
Abstract:
There are various models to estimate the neutral atmospheric parameter values, such as in-suite and reanalysis datasets from numerical models. Accurate estimated values of the atmospheric parameters are useful for weather forecasting and, climate modeling and monitoring of climate change. Recently, Global Navigation Satellite System (GNSS) measurements have been applied for atmospheric sounding due to its robust data quality and wide horizontal and vertical coverage. The Global Positioning System (GPS) solutions that includes tropospheric parameters constitute a reliable set of data to be assimilated into climate models. The objective of this paper is, to estimate the neutral atmospheric parameters such as Wet Zenith Delay (WZD), Precipitable Water Vapour (PWV) and Total Zenith Delay (TZD) using six selected GPS stations in the equatorial regions, more precisely, the Ethiopian GPS stations from 2012 to 2015 observational data. Based on historic estimated GPS-derived values of PWV, we forecasted the PWV from 2015 to 2030. During data processing and analysis, we applied GAMIT-GLOBK software packages to estimate the atmospheric parameters. In the result, we found that the annual averaged minimum values of PWV are 9.72 mm for IISC and maximum 50.37 mm for BJCO stations. The annual averaged minimum values of WZD are 6 cm for IISC and maximum 31 cm for BDMT stations. In the long series of observations (from 2012 to 2015), we also found that there is a trend and cyclic patterns of WZD, PWV and TZD for all stations.Keywords: atmosphere, GNSS, neutral atmosphere, precipitable water vapour
Procedia PDF Downloads 656132 The Relationship between Depression, HIV Stigma and Adherence to Antiretroviral Therapy among Adult Patients Living with HIV at a Tertiary Hospital in Durban, South Africa: The Mediating Roles of Self-Efficacy and Social Support
Authors: Muziwandile Luthuli
Abstract:
Although numerous factors predicting adherence to antiretroviral therapy (ART) among people living with HIV/AIDS (PLWHA) have been broadly studied on both regional and global level, up-to-date adherence of patients to ART remains an overarching, dynamic and multifaceted problem that needs to be investigated over time and across various contexts. There is a rarity of empirical data in the literature on interactive mechanisms by which psychosocial factors influence adherence to ART among PLWHA within the South African context. Therefore, this study was designed to investigate the relationship between depression, HIV stigma, and adherence to ART among adult patients living with HIV at a tertiary hospital in Durban, South Africa, and the mediating roles of self-efficacy and social support. The health locus of control theory and the social support theory were the underlying theoretical frameworks for this study. Using a cross-sectional research design, a total of 201 male and female adult patients aged between 18-75 years receiving ART at a tertiary hospital in Durban, KwaZulu-Natal were sampled, using time location sampling (TLS). A self-administered questionnaire was employed to collect the data in this study. Data were analysed through SPSS version 27. Several statistical analyses were conducted in this study, namely univariate statistical analysis, correlational analysis, Pearson’s chi-square analysis, cross-tabulation analysis, binary logistic regression analysis, and mediational analysis. Univariate analysis indicated that the sample mean age was 39.28 years (SD=12.115), while most participants were females 71.0% (n=142), never married 74.2% (n=147), and most were also secondary school educated 48.3% (n=97), as well as unemployed 65.7% (n=132). The prevalence rate of participants who had high adherence to ART was 53.7% (n=108), and 46.3% (n=93) of participants had low adherence to ART. Chi-square analysis revealed that employment status was the only statistically significant socio-demographic influence of adherence to ART in this study (χ2 (3) = 8.745; p < .033). Chi-square analysis showed that there was a statistically significant difference found between depression and adherence to ART (χ2 (4) = 16.140; p < .003), while between HIV stigma and adherence to ART, no statistically significant difference was found (χ2 (1) = .323; p >.570). Binary logistic regression indicated that depression was statistically associated with adherence to ART (OR= .853; 95% CI, .789–.922, P < 001), while the association between self-efficacy and adherence to ART was statistically significant (OR= 1.04; 95% CI, 1.001– 1.078, P < .045) after controlling for the effect of depression. However, the findings showed that the effect of depression on adherence to ART was not significantly mediated by self-efficacy (Sobel test for indirect effect, Z= 1.01, P > 0.31). Binary logistic regression showed that the effect of HIV stigma on adherence to ART was not statistically significant (OR= .980; 95% CI, .937– 1.025, P > .374), but the effect of social support on adherence to ART was statistically significant, only after the effect of HIV stigma was controlled for (OR= 1.017; 95% CI, 1.000– 1.035, P < .046). This study promotes behavioral and social change effected through evidence-based interventions by emphasizing the need for additional research that investigates the interactive mechanisms by which psychosocial factors influence adherence to ART. Depression is a significant predictor of adherence to ART. Thus, to alleviate the psychosocial impact of depression on adherence to ART, effective interventions must be devised, along with special consideration of self-efficacy and social support. Therefore, this study is helpful in informing and effecting change in health policy and healthcare services through its findingsKeywords: ART adherence, depression, HIV/AIDS, PLWHA
Procedia PDF Downloads 1816131 Mechanistic Understanding of the Difference in two Strains Cholerae Causing Pathogens and Predicting Therapeutic Strategies for Cholera Patients Affected with new Strain Vibrio Cholerae El.tor. Using Constrain-based Modelling
Authors: Faiz Khan Mohammad, Saumya Ray Chaudhari, Raghunathan Rengaswamy, Swagatika Sahoo
Abstract:
Cholera caused by pathogenic gut bacteria Vibrio Cholerae (VC), is a major health problem in developing countries. Different strains of VC exhibit variable responses subject to different extracellular medium (Nag et al, Infect Immun, 2018). In this study, we present a new approach to model the variable VC responses in mono- and co-cultures, subject to continuously changing growth medium, which is otherwise difficult via simple FBA model. Nine VC strain and seven E. coli (EC) models were assembled and considered. A continuously changing medium is modelled using a new iterative-based controlled medium technique (ITC). The medium is appropriately prefixed with the VC model secretome. As the flux through the bacteria biomass increases secretes certain by-products. These products shall add-on to the medium, either deviating the nutrient potential or block certain metabolic components of the model, effectively forming a controlled feed-back loop. Different VC models were setup as monoculture of VC in glucose enriched medium, and in co-culture with VC strains and EC. Constrained to glucose enriched medium, (i) VC_Classical model resulted in higher flux through acidic secretome suggesting a pH change of the medium, leading to lowering of its biomass. This is in consonance with the literature reports. (ii) When compared for neutral secretome, flux through acetoin exchange was higher in VC_El tor than the classical models, suggesting El tor requires an acidic partner to lower its biomass. (iii) Seven of nine VC models predicted 3-methyl-2-Oxovaleric acid, mysirtic acid, folic acid, and acetate significantly affect corresponding biomass reactions. (iv) V. parhemolyticus and vulnificus were found to be phenotypically similar to VC Classical strain, across the nine VC strains. The work addresses the advantage of the ITC over regular flux balance analysis for modelling varying growth medium. Future expansion to co-cultures, potentiates the identification of novel interacting partners as effective cholera therapeutics.Keywords: cholera, vibrio cholera El. tor, vibrio cholera classical, acetate
Procedia PDF Downloads 1666130 Biosorption of Fluoride from Aqueous Solutions by Tinospora Cordifolia Leaves
Authors: Srinivasulu Dasaiah, Kalyan Yakkala, Gangadhar Battala, Pavan Kumar Pindi, Ramakrishna Naidu Gurijala
Abstract:
Tinospora cordifolia leaves biomass used for the removal fluoride from aqueous solutions. Batch biosorption technique was applied, pH, contact time, biosorbent dose and initial fluoride concentration was studied. The Scanning Electron Microscopy (SEM) and Fourier Transform Infrared (FTIR) techniques used to study the surface characteristics and the presence of chemical functional groups on the biosorbent. Biosorption isotherm models and kinetic models were applied to understand the sorption mechanism. Results revealed that pH, contact time, biosorbent dose and initial fluoride concentration played a significant effect on fluoride removal from aqueous solutions. The developed biosorbent derived from Tinospora cordifolia leaves biomass found to be a low-cost biosorbent and could be used for the effective removal of fluoride in synthetic as well as real water samples.Keywords: biosorption, contact time, fluoride, isotherms
Procedia PDF Downloads 1806129 Pedagogical Technologies of Teaching Natural Geography
Authors: Mirzahmedov Ismoiljon Karimjon Ugli, Juraeva Shakhnoza Abdumalik Kizi
Abstract:
The article deals with the current scientific problems of natural geography related to the development of new pedagogical technologies and their implementation in the educational process. The use of recommended interactive methods in independent study is considered very effective and is a very useful method for students, especially for students who work more on themselves. Today's demand is to make young people talented, intelligent, innovative, as well as mature and well-rounded individuals, as a result of the work carried out in the field of education today. This is how creating tables of different contents and filling them out shows the student's talent and desire for innovation. Also, the techniques and methods necessary for today's student are shown, the role of the teacher in conducting lessons meaningfully, the suitability of the method used by the teacher for the lesson, factors affecting the quality of education, and natural issues of the use of methods based on the specific features of geography are highlighted.Keywords: teaching methods, educational process, educational technologies, education, problem, didactics, natural geography
Procedia PDF Downloads 686128 Financial Inclusion and Modernization: Secure Energy Performance in Shanghai Cooperation Organization
Authors: Shama Urooj
Abstract:
The present work investigates the relationship among financial inclusion, modernization, and energy performance in SCO member countries during the years 2011–2021. PCA is used to create composite indexes of financial inclusion, modernization, and energy performance. We used panel regression models that are both reliable and heteroscedasticity-consistent to look at the relationship among variables. The findings indicate that financial inclusion (FI) and modernization, along with the increased FDI, all appear to contribute to the energy performance in the SCO member countries. However, per capita GDP has a negative impact on energy performance. These results are unbiased and consistent with the robust results obtained by applying different econometric models. Feasible Generalized Least Square (FGLS) estimation is also used for checking the uniformity of the main model results. This research work concludes that there has been no policy coherence in SCO member countries regarding the coordination of growing financial inclusion and modernization for energy sustainability in recent years. In order to improve energy performance with modern development, policies regarding financial inclusion and modernization need be integrated both at national as well as international levels.Keywords: financial inclusion, energy performance, modernization, technological development, SCO.
Procedia PDF Downloads 786127 Predictive Models of Ruin Probability in Retirement Withdrawal Strategies
Authors: Yuanjin Liu
Abstract:
Retirement withdrawal strategies are very important to minimize the probability of ruin in retirement. The ruin probability is modeled as a function of initial withdrawal age, gender, asset allocation, inflation rate, and initial withdrawal rate. The ruin probability is obtained based on the 2019 period life table for the Social Security, IRS Required Minimum Distribution (RMD) Worksheets, US historical bond and equity returns, and inflation rates using simulation. Several popular machine learning algorithms of the generalized additive model, random forest, support vector machine, extreme gradient boosting, and artificial neural network are built. The model validation and selection are based on the test errors using hyperparameter tuning and train-test split. The optimal model is recommended for retirees to monitor the ruin probability. The optimal withdrawal strategy can be obtained based on the optimal predictive model.Keywords: ruin probability, retirement withdrawal strategies, predictive models, optimal model
Procedia PDF Downloads 776126 Low-Cost Parking Lot Mapping and Localization for Home Zone Parking Pilot
Authors: Hongbo Zhang, Xinlu Tang, Jiangwei Li, Chi Yan
Abstract:
Home zone parking pilot (HPP) is a fast-growing segment in low-speed autonomous driving applications. It requires the car automatically cruise around a parking lot and park itself in a range of up to 100 meters inside a recurrent home/office parking lot, which requires precise parking lot mapping and localization solution. Although Lidar is ideal for SLAM, the car OEMs favor a low-cost fish-eye camera based visual SLAM approach. Recent approaches have employed segmentation models to extract semantic features and improve mapping accuracy, but these AI models are memory unfriendly and computationally expensive, making deploying on embedded ADAS systems difficult. To address this issue, we proposed a new method that utilizes object detection models to extract robust and accurate parking lot features. The proposed method could reduce computational costs while maintaining high accuracy. Once combined with vehicles’ wheel-pulse information, the system could construct maps and locate the vehicle in real-time. This article will discuss in detail (1) the fish-eye based Around View Monitoring (AVM) with transparent chassis images as the inputs, (2) an Object Detection (OD) based feature point extraction algorithm to generate point cloud, (3) a low computational parking lot mapping algorithm and (4) the real-time localization algorithm. At last, we will demonstrate the experiment results with an embedded ADAS system installed on a real car in the underground parking lot.Keywords: ADAS, home zone parking pilot, object detection, visual SLAM
Procedia PDF Downloads 696125 Mathematical Modeling of Bi-Substrate Enzymatic Reactions in the Presence of Different Types of Inhibitors
Authors: Rafayel Azizyan, Valeri Arakelyan, Aram Gevorgyan, Varduhi Balayan, Emil Gevorgyan
Abstract:
Currently, mathematical and computer modeling are widely used in different biological studies to predict or assess behavior of such complex systems as biological ones. This study deals with mathematical and computer modeling of bi-substrate enzymatic reactions, which play an important role in different biochemical pathways. The main objective of this study is to represent the results from in silico investigation of bi-substrate enzymatic reactions in the presence of uncompetitive inhibitors, as well as to describe in details the inhibition effects. Four models of uncompetitive inhibition were designed using different software packages. Particularly, uncompetitive inhibitor to the first [ES1] and the second ([ES1S2]; [FS2]) enzyme-substrate complexes have been studied. The simulation, using the same kinetic parameters for all models allowed investigating the behavior of reactions as well as determined some interesting aspects concerning influence of different cases of uncompetitive inhibition. Besides that shown, that uncompetitive inhibitors exhibit specific selectivity depending on mechanism of bi-substrate enzymatic reaction.Keywords: mathematical modeling, bi-substrate enzymatic reactions, reversible inhibition
Procedia PDF Downloads 3496124 Thin-Layer Drying Characteristics and Modelling of Instant Coffee Solution
Authors: Apolinar Picado, Ronald Solís, Rafael Gamero
Abstract:
The thin-layer drying characteristics of instant coffee solution were investigated in a laboratory tunnel dryer. Drying experiments were carried out at three temperatures (80, 100 and 120 °C) and an air velocity of 1.2 m/s. Drying experimental data obtained are fitted to six (6) thin-layer drying models using the non-linear least squares regression analysis. The acceptability of the thin-layer drying model has been based on a value of the correlation coefficient that should be close to one, and low values for root mean square error (RMSE) and chi-square (x²). According to this evaluation, the most suitable model for describing drying process of thin-layer instant coffee solution is the Page model. Further, the effective moisture diffusivity and the activation energy were computed employing the drying experimental data. The effective moisture diffusivity values varied from 1.6133 × 10⁻⁹ to 1.6224 × 10⁻⁹ m²/s over the temperature range studied and the activation energy was estimated to be 162.62 J/mol.Keywords: activation energy, diffusivity, instant coffee, thin-layer models
Procedia PDF Downloads 2646123 Comparing Business Excellence Models Using Quantitative Methods: A First Step
Authors: Mohammed Alanazi, Dimitrios Tsagdis
Abstract:
Established Business Excellence Models (BEMs), like the Malcolm Baldrige National Quality Award (MBNQA) model and the European Foundation for Quality Management (EFQM) model, have been adopted by firms all over the world. They exist alongside more recent country-specific BEMs; e.g. the Australian, Canadian, China, New Zealand, Singapore, and Taiwan quality awards that although not as widespread as MBNQA and EFQM have nonetheless strong national followings. Regardless of any differences in their following or prestige, the emergence and development of all BEMs have been shaped both by their local context (e.g. underlying socio-economic dynamics) as well as by global best practices. Besides such similarities, that render them into objects (i.e. models) of the same class (i.e. BEMs), BEMs exhibit non-trivial differences in their criteria, relations, and emphasis. Given the evolution of BEMs (e.g. the MBNQA underwent seven evolutions since its inception in 1987 while the EFQM five since 1993), it is unsurprising that comparative studies of their validity are few and far in between. This poses challenges for practitioners and policy makers alike; as it is not always clear which BEM is to be preferred or better fitting to a particular context. Especially, in contexts that differ substantially from the original context of BEM development. This paper aims to fill this gap by presenting a research design and measurement model for comparing BEMs using quantitative methods (e.g. structural equations). Three BEMs will be focused upon in particular for illustration purposes; the MBNQA, the EFQM, and the King Abdul Aziz Quality Award (KAQA) model. They have been selected so to reflect the two established and widely spread traditions as well as a more recent context-specific arrival promising a better fit.Keywords: Baldrige, business excellence, European Foundation for Quality Management, Structural Equation Model, total quality management
Procedia PDF Downloads 2406122 Why and When to Teach Definitions: Necessary and Unnecessary Discontinuities Resulting from the Definition of Mathematical Concepts
Authors: Josephine Shamash, Stuart Smith
Abstract:
We examine reasons for introducing definitions in teaching mathematics in a number of different cases. We try to determine if, where, and when to provide a definition, and which definition to choose. We characterize different types of definitions and the different purposes we may have for formulating them, and detail examples of each type. Giving a definition at a certain stage can sometimes be detrimental to the development of the concept image. In such a case, it is advisable to delay the precise definition to a later stage. We describe two models, the 'successive approximation model', and the 'model of the extending definition' that fit such situations. Detailed examples that fit the different models are given based on material taken from a number of textbooks, and analysis of the way the concept is introduced, and where and how its definition is given. Our conclusions, based on this analysis, is that some of the definitions given may cause discontinuities in the learning sequence and constitute obstacles and unnecessary cognitive conflicts in the formation of the concept definition. However, in other cases, the discontinuity in passing from definition to definition actually serves a didactic purpose, is unavoidable for the mathematical evolution of the concept image, and is essential for students to deepen their understanding.Keywords: concept image, mathematical definitions, mathematics education, mathematics teaching
Procedia PDF Downloads 1336121 On-Line Data-Driven Multivariate Statistical Prediction Approach to Production Monitoring
Authors: Hyun-Woo Cho
Abstract:
Detection of incipient abnormal events in production processes is important to improve safety and reliability of manufacturing operations and reduce losses caused by failures. The construction of calibration models for predicting faulty conditions is quite essential in making decisions on when to perform preventive maintenance. This paper presents a multivariate calibration monitoring approach based on the statistical analysis of process measurement data. The calibration model is used to predict faulty conditions from historical reference data. This approach utilizes variable selection techniques, and the predictive performance of several prediction methods are evaluated using real data. The results shows that the calibration model based on supervised probabilistic model yielded best performance in this work. By adopting a proper variable selection scheme in calibration models, the prediction performance can be improved by excluding non-informative variables from their model building steps.Keywords: calibration model, monitoring, quality improvement, feature selection
Procedia PDF Downloads 3606120 Discrete State Prediction Algorithm Design with Self Performance Enhancement Capacity
Authors: Smail Tigani, Mohamed Ouzzif
Abstract:
This work presents a discrete quantitative state prediction algorithm with intelligent behavior making it able to self-improve some performance aspects. The specificity of this algorithm is the capacity of self-rectification of the prediction strategy before the final decision. The auto-rectification mechanism is based on two parallel mathematical models. In one hand, the algorithm predicts the next state based on event transition matrix updated after each observation. In the other hand, the algorithm extracts its residues trend with a linear regression representing historical residues data-points in order to rectify the first decision if needs. For a normal distribution, the interactivity between the two models allows the algorithm to self-optimize its performance and then make better prediction. Designed key performance indicator, computed during a Monte Carlo simulation, shows the advantages of the proposed approach compared with traditional one.Keywords: discrete state, Markov Chains, linear regression, auto-adaptive systems, decision making, Monte Carlo Simulation
Procedia PDF Downloads 4996119 Effect of Plasticizer Additives on the Mechanical Properties of Cement Composite: A Molecular Dynamics Analysis
Authors: R. Mohan, V. Jadhav, A. Ahmed, J. Rivas, A. Kelkar
Abstract:
Cementitious materials are an excellent example of a composite material with complex hierarchical features and random features that range from nanometer (nm) to millimeter (mm) scale. Multi-scale modeling of complex material systems requires starting from fundamental building blocks to capture the scale relevant features through associated computational models. In this paper, molecular dynamics (MD) modeling is employed to predict the effect of plasticizer additive on the mechanical properties of key hydrated cement constituent calcium-silicate-hydrate (CSH) at the molecular, nanometer scale level. Due to complexity, still unknown molecular configuration of CSH, a representative configuration widely accepted in the field of mineral Jennite is employed. The effectiveness of the Molecular Dynamics modeling to understand the predictive influence of material chemistry changes based on molecular/nanoscale models is demonstrated.Keywords: cement composite, mechanical properties, molecular dynamics, plasticizer additives
Procedia PDF Downloads 4576118 Kou Jump Diffusion Model: An Application to the SP 500; Nasdaq 100 and Russell 2000 Index Options
Authors: Wajih Abbassi, Zouhaier Ben Khelifa
Abstract:
The present research points towards the empirical validation of three options valuation models, the ad-hoc Black-Scholes model as proposed by Berkowitz (2001), the constant elasticity of variance model of Cox and Ross (1976) and the Kou jump-diffusion model (2002). Our empirical analysis has been conducted on a sample of 26,974 options written on three indexes, the S&P 500, Nasdaq 100 and the Russell 2000 that were negotiated during the year 2007 just before the sub-prime crisis. We start by presenting the theoretical foundations of the models of interest. Then we use the technique of trust-region-reflective algorithm to estimate the structural parameters of these models from cross-section of option prices. The empirical analysis shows the superiority of the Kou jump-diffusion model. This superiority arises from the ability of this model to portray the behavior of market participants and to be closest to the true distribution that characterizes the evolution of these indices. Indeed the double-exponential distribution covers three interesting properties that are: the leptokurtic feature, the memory less property and the psychological aspect of market participants. Numerous empirical studies have shown that markets tend to have both overreaction and under reaction over good and bad news respectively. Despite of these advantages there are not many empirical studies based on this model partly because probability distribution and option valuation formula are rather complicated. This paper is the first to have used the technique of nonlinear curve-fitting through the trust-region-reflective algorithm and cross-section options to estimate the structural parameters of the Kou jump-diffusion model.Keywords: jump-diffusion process, Kou model, Leptokurtic feature, trust-region-reflective algorithm, US index options
Procedia PDF Downloads 4316117 Metrology-Inspired Methods to Assess the Biases of Artificial Intelligence Systems
Authors: Belkacem Laimouche
Abstract:
With the field of artificial intelligence (AI) experiencing exponential growth, fueled by technological advancements that pave the way for increasingly innovative and promising applications, there is an escalating need to develop rigorous methods for assessing their performance in pursuit of transparency and equity. This article proposes a metrology-inspired statistical framework for evaluating bias and explainability in AI systems. Drawing from the principles of metrology, we propose a pioneering approach, using a concrete example, to evaluate the accuracy and precision of AI models, as well as to quantify the sources of measurement uncertainty that can lead to bias in their predictions. Furthermore, we explore a statistical approach for evaluating the explainability of AI systems based on their ability to provide interpretable and transparent explanations of their predictions.Keywords: artificial intelligence, metrology, measurement uncertainty, prediction error, bias, machine learning algorithms, probabilistic models, interlaboratory comparison, data analysis, data reliability, measurement of bias impact on predictions, improvement of model accuracy and reliability
Procedia PDF Downloads 107