Search results for: estimation algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3842

Search results for: estimation algorithms

722 Comparison of Direction of Arrival Estimation Method for Drone Based on Phased Microphone Array

Authors: Jiwon Lee, Yeong-Ju Go, Jong-Soo Choi

Abstract:

Drones were first developed for military use and were used in World War 1. But recently drones have been used in a variety of fields. Several companies actively utilize drone technology to strengthen their services, and in agriculture, drones are used for crop monitoring and sowing. Other people use drones for hobby activities such as photography. However, as the range of use of drones expands rapidly, problems caused by drones such as improperly flying, privacy and terrorism are also increasing. As the need for monitoring and tracking of drones increases, researches are progressing accordingly. The drone detection system estimates the position of the drone using the physical phenomena that occur when the drones fly. The drone detection system measures being developed utilize many approaches, such as radar, infrared camera, and acoustic detection systems. Among the various drone detection system, the acoustic detection system is advantageous in that the microphone array system is small, inexpensive, and easy to operate than other systems. In this paper, the acoustic signal is acquired by using minimum microphone when drone is flying, and direction of drone is estimated. When estimating the Direction of Arrival(DOA), there is a method of calculating the DOA based on the Time Difference of Arrival(TDOA) and a method of calculating the DOA based on the beamforming. The TDOA technique requires less number of microphones than the beamforming technique, but is weak in noisy environments and can only estimate the DOA of a single source. The beamforming technique requires more microphones than the TDOA technique. However, it is strong against the noisy environment and it is possible to simultaneously estimate the DOA of several drones. When estimating the DOA using acoustic signals emitted from the drone, it is impossible to measure the position of the drone, and only the direction can be estimated. To overcome this problem, in this work we show how to estimate the position of drones by arranging multiple microphone arrays. The microphone array used in the experiments was four tetrahedral microphones. We simulated the performance of each DOA algorithm and demonstrated the simulation results through experiments.

Keywords: acoustic sensing, direction of arrival, drone detection, microphone array

Procedia PDF Downloads 160
721 An Integrated Label Propagation Network for Structural Condition Assessment

Authors: Qingsong Xiong, Cheng Yuan, Qingzhao Kong, Haibei Xiong

Abstract:

Deep-learning-driven approaches based on vibration responses have attracted larger attention in rapid structural condition assessment while obtaining sufficient measured training data with corresponding labels is relevantly costly and even inaccessible in practical engineering. This study proposes an integrated label propagation network for structural condition assessment, which is able to diffuse the labels from continuously-generating measurements by intact structure to those of missing labels of damage scenarios. The integrated network is embedded with damage-sensitive features extraction by deep autoencoder and pseudo-labels propagation by optimized fuzzy clustering, the architecture and mechanism which are elaborated. With a sophisticated network design and specified strategies for improving performance, the present network achieves to extends the superiority of self-supervised representation learning, unsupervised fuzzy clustering and supervised classification algorithms into an integration aiming at assessing damage conditions. Both numerical simulations and full-scale laboratory shaking table tests of a two-story building structure were conducted to validate its capability of detecting post-earthquake damage. The identifying accuracy of a present network was 0.95 in numerical validations and an average 0.86 in laboratory case studies, respectively. It should be noted that the whole training procedure of all involved models in the network stringently doesn’t rely upon any labeled data of damage scenarios but only several samples of intact structure, which indicates a significant superiority in model adaptability and feasible applicability in practice.

Keywords: autoencoder, condition assessment, fuzzy clustering, label propagation

Procedia PDF Downloads 97
720 Downscaling Grace Gravity Models Using Spectral Combination Techniques for Terrestrial Water Storage and Groundwater Storage Estimation

Authors: Farzam Fatolazadeh, Kalifa Goita, Mehdi Eshagh, Shusen Wang

Abstract:

The Gravity Recovery and Climate Experiment (GRACE) is a satellite mission with twin satellites for the precise determination of spatial and temporal variations in the Earth’s gravity field. The products of this mission are monthly global gravity models containing the spherical harmonic coefficients and their errors. These GRACE models can be used for estimating terrestrial water storage (TWS) variations across the globe at large scales, thereby offering an opportunity for surface and groundwater storage (GWS) assessments. Yet, the ability of GRACE to monitor changes at smaller scales is too limited for local water management authorities. This is largely due to the low spatial and temporal resolutions of its models (~200,000 km2 and one month, respectively). High-resolution GRACE data products would substantially enrich the information that is needed by local-scale decision-makers while offering the data for the regions that lack adequate in situ monitoring networks, including northern parts of Canada. Such products could eventually be obtained through downscaling. In this study, we extended the spectral combination theory to simultaneously downscale spatiotemporally the 3o spatial coarse resolution of GRACE to 0.25o degrees resolution and monthly coarse resolution to daily resolution. This method combines the monthly gravity field solution of GRACE and daily hydrological model products in the form of both low and high-frequency signals to produce high spatiotemporal resolution TWSA and GWSA products. The main contribution and originality of this study are to comprehensively and simultaneously consider GRACE and hydrological variables and their uncertainties to form the estimator in the spectral domain. Therefore, it is predicted that we reach downscale products with an acceptable accuracy.

Keywords: GRACE satellite, groundwater storage, spectral combination, terrestrial water storage

Procedia PDF Downloads 83
719 Advancements in Autonomous Drones for Enhanced Healthcare Logistics

Authors: Bhaargav Gupta P., Vignesh N., Nithish Kumar R., Rahul J., Nivetha Ruvah D.

Abstract:

Delivering essential medical supplies to rural and underserved areas is challenging due to infrastructure limitations and logistical barriers, often resulting in inefficiencies and delays. Traditional delivery methods are hindered by poor road networks, long distances, and difficult terrains, compromising timely access to vital resources, especially in emergencies. This paper introduces an autonomous drone system engineered to optimize last-mile delivery. By utilizing advanced navigation and object-detection algorithms, such as region-based convolutional neural networks (R-CNN), our drones efficiently avoid obstacles, identify safe landing zones, and adapt dynamically to varying environments. Equipped with high-precision GPS and autonomous capabilities, the drones effectively navigate complex, remote areas with minimal dependence on established infrastructure. The system includes a dedicated mobile application for secure order placement and real-time tracking, and a secure payload box with OTP verification ensures tamper-resistant delivery to authorized recipients. This project demonstrates the potential of automated drone technology in healthcare logistics, offering a scalable and eco-friendly approach to enhance accessibility and service delivery in underserved regions. By addressing logistical gaps through advanced automation, this system represents a significant advancement toward sustainable, accessible healthcare in remote areas.

Keywords: region-based convolutional neural network, one time password, global positioning system, autonomous drones, healthcare logistics

Procedia PDF Downloads 12
718 Comparative Analysis of Change in Vegetation in Four Districts of Punjab through Satellite Imagery, Land Use Statistics and Machine Learning

Authors: Mirza Waseem Abbas, Syed Danish Raza

Abstract:

For many countries agriculture is still the major force driving the economy and a critically important socioeconomic sector, despite exceptional industrial development across the globe. In countries like Pakistan, this sector is considered the backbone of the economy, and most of the economic decision making revolves around agricultural outputs and data. Timely and accurate facts and figures about this vital sector hold immense significance and have serious implications for the long-term development of the economy. Therefore, any significant improvements in the statistics and other forms of data regarding agriculture sector are considered important by all policymakers. This is especially true for decision making for the betterment of crops and the agriculture sector in general. Provincial and federal agricultural departments collect data for all cash and non-cash crops and the sector, in general, every year. Traditional data collection for such a large sector i.e. agriculture, being time-consuming, prone to human error and labor-intensive, is slowly but gradually being replaced by remote sensing techniques. For this study, remotely sensed data were used for change detection (machine learning, supervised & unsupervised classification) to assess the increase or decrease in area under agriculture over the last fifteen years due to urbanization. Detailed Landsat Images for the selected agricultural districts were acquired for the year 2000 and compared to images of the same area acquired for the year 2016. Observed differences validated through detailed analysis of the areas show that there was a considerable decrease in vegetation during the last fifteen years in four major agricultural districts of the Punjab province due to urbanization (housing societies).

Keywords: change detection, area estimation, machine learning, urbanization, remote sensing

Procedia PDF Downloads 251
717 Quantification and Evaluation of Tumors Heterogeneity Utilizing Multimodality Imaging

Authors: Ramin Ghasemi Shayan, Morteza Janebifam

Abstract:

Tumors are regularly inhomogeneous. Provincial varieties in death, metabolic action, multiplication and body part are watched. There’s expanding proof that strong tumors may contain subpopulations of cells with various genotypes and phenotypes. These unmistakable populaces of malignancy cells can connect during a serious way and may contrast in affectability to medications. Most tumors show organic heterogeneity1–3 remembering heterogeneity for genomic subtypes, varieties inside the statement of development variables and genius, and hostile to angiogenic factors4–9 and varieties inside the tumoural microenvironment. These can present as contrasts between tumors in a few people. for instance, O6-methylguanine-DNA methyltransferase, a DNA fix compound, is hushed by methylation of the quality advertiser in half of glioblastoma (GBM), adding to chemosensitivity, and improved endurance. From the outset, there includes been specific enthusiasm inside the usage of dissemination weighted imaging (DWI) and dynamic complexity upgraded MRI (DCE-MRI). DWI sharpens MRI to water dispersion inside the extravascular extracellular space (EES) and is wiped out with the size and setup of the cell populace. Additionally, DCE-MRI utilizes dynamic obtaining of pictures during and after the infusion of intravenous complexity operator. Signal changes are additionally changed to outright grouping of differentiation permitting examination utilizing pharmacokinetic models. PET scan modality gives one of a kind natural particularity, permitting dynamic or static imaging of organic atoms marked with positron emanating isotopes (for example, 15O, 18F, 11C). The strategy is explained to a colossal radiation portion, which points of confinement rehashed estimations, particularly when utilized together with PC tomography (CT). At long last, it's of incredible enthusiasm to quantify territorial hemoglobin state, which could be joined with DCE-CT vascular physiology estimation to create significant experiences for understanding tumor hypoxia.

Keywords: heterogeneity, computerized tomography scan, magnetic resonance imaging, PET

Procedia PDF Downloads 147
716 Applications of Evolutionary Optimization Methods in Reinforcement Learning

Authors: Rahul Paul, Kedar Nath Das

Abstract:

The paradigm of Reinforcement Learning (RL) has become prominent in training intelligent agents to make decisions in environments that are both dynamic and uncertain. The primary objective of RL is to optimize the policy of an agent in order to maximize the cumulative reward it receives throughout a given period. Nevertheless, the process of optimization presents notable difficulties as a result of the inherent trade-off between exploration and exploitation, the presence of extensive state-action spaces, and the intricate nature of the dynamics involved. Evolutionary Optimization Methods (EOMs) have garnered considerable attention as a supplementary approach to tackle these challenges, providing distinct capabilities for optimizing RL policies and value functions. The ongoing advancement of research in both RL and EOMs presents an opportunity for significant advancements in autonomous decision-making systems. The convergence of these two fields has the potential to have a transformative impact on various domains of artificial intelligence (AI) applications. This article highlights the considerable influence of EOMs in enhancing the capabilities of RL. Taking advantage of evolutionary principles enables RL algorithms to effectively traverse extensive action spaces and discover optimal solutions within intricate environments. Moreover, this paper emphasizes the practical implementations of EOMs in the field of RL, specifically in areas such as robotic control, autonomous systems, inventory problems, and multi-agent scenarios. The article highlights the utilization of EOMs in facilitating RL agents to effectively adapt, evolve, and uncover proficient strategies for complex tasks that may pose challenges for conventional RL approaches.

Keywords: machine learning, reinforcement learning, loss function, optimization techniques, evolutionary optimization methods

Procedia PDF Downloads 81
715 Fexofenadine Hydrochloride Orodispersisble Tablets: Formulation and in vitro/in vivo Evaluation in Healthy Human Volunteers

Authors: Soad Ali Yehia, Mohamed Shafik El-Ridi, Mina Ibrahim Tadros, Nolwa Gamal El-Sherif

Abstract:

Fexofenadine hydrochloride (FXD) is a slightly soluble, bitter-tasting, drug having an oral bioavailability of 35%. The maximum plasma concentration is reached 2.6 hours (Tmax) post-dose. The current work aimed to develop taste-masked FXD orodispersible tablets (ODTs) to increase extent of drug absorption and reduce Tmax. Taste masking was achieved via solid dispersion (SD) with chitosan (CS) or sodium alginate (ALG). FT-IR, DSC and XRD were performed to identify physicochemical interactions and FXD crystallinity. Taste-masked FXD-ODTs were developed via addition of superdisintegrants (crosscarmelose sodium or sodium starch glycolate, 5% and 10%, w/w) or sublimable agents (camphor, menthol or thymol; 10% and 20%, w/w) to FXD-SDs. ODTs were evaluated for weight variation, drug-content, friability, wetting time, disintegration time and drug release. Camphor-based (20%, w/w) FXD-ODT (F12) was optimized (F23) by incorporation of a more hydrophilic lubricant, sodium stearyl fumarate (Pruv®). The topography of the latter formula was examined via scanning electron microscopy (SEM). The in vivo estimation of FXD pharmacokinetics, relative to Allegra® tablets, was evaluated in healthy human volunteers. Based on the gustatory sensation test in healthy volunteers, FXD:CS (1:1) and FXD:ALG (1:0.5) SDs were selected. Taste-masked FXD-ODTs had appropriate physicochemical properties and showed short wetting and disintegration times. Drug release profiles of F23 and phenylalanine-containing Allegra® ODT were similar (f2 = 96) showing a complete release in two minutes. SEM micrographs revealed pores following camphor sublimation. Compared to Allegra® tablets, pharmacokinetic studies in healthy volunteers proved F23 ability to increase extent of FXD absorption (14%) and reduce Tmax to 1.83 h.

Keywords: fexofenadine hydrochloride, taste masking, chitosan, orodispersible

Procedia PDF Downloads 344
714 Low-Cost Image Processing System for Evaluating Pavement Surface Distress

Authors: Keerti Kembhavi, M. R. Archana, V. Anjaneyappa

Abstract:

Most asphalt pavement condition evaluation use rating frameworks in which asphalt pavement distress is estimated by type, extent, and severity. Rating is carried out by the pavement condition rating (PCR), which is tedious and expensive. This paper presents the development of a low-cost technique for image pavement distress analysis that permits the identification of pothole and cracks. The paper explores the application of image processing tools for the detection of potholes and cracks. Longitudinal cracking and pothole are detected using Fuzzy-C- Means (FCM) and proceeded with the Spectral Theory algorithm. The framework comprises three phases, including image acquisition, processing, and extraction of features. A digital camera (Gopro) with the holder is used to capture pavement distress images on a moving vehicle. FCM classifier and Spectral Theory algorithms are used to compute features and classify the longitudinal cracking and pothole. The Matlab2016Ra Image preparing tool kit utilizes performance analysis to identify the viability of pavement distress on selected urban stretches of Bengaluru city, India. The outcomes of image evaluation with the utilization semi-computerized image handling framework represented the features of longitudinal crack and pothole with an accuracy of about 80%. Further, the detected images are validated with the actual dimensions, and it is seen that dimension variability is about 0.46. The linear regression model y=1.171x-0.155 is obtained using the existing and experimental / image processing area. The R2 correlation square obtained from the best fit line is 0.807, which is considered in the linear regression model to be ‘large positive linear association’.

Keywords: crack detection, pothole detection, spectral clustering, fuzzy-c-means

Procedia PDF Downloads 181
713 Predictive Modeling of Bridge Conditions Using Random Forest

Authors: Miral Selim, May Haggag, Ibrahim Abotaleb

Abstract:

The aging of transportation infrastructure presents significant challenges, particularly concerning the monitoring and maintenance of bridges. This study investigates the application of Random Forest algorithms for predictive modeling of bridge conditions, utilizing data from the US National Bridge Inventory (NBI). The research is significant as it aims to improve bridge management through data-driven insights that can enhance maintenance strategies and contribute to overall safety. Random Forest is chosen for its robustness, ability to handle complex, non-linear relationships among variables, and its effectiveness in feature importance evaluation. The study begins with comprehensive data collection and cleaning, followed by the identification of key variables influencing bridge condition ratings, including age, construction materials, environmental factors, and maintenance history. Random Forest is utilized to examine the relationships between these variables and the predicted bridge conditions. The dataset is divided into training and testing subsets to evaluate the model's performance. The findings demonstrate that the Random Forest model effectively enhances the understanding of factors affecting bridge conditions. By identifying bridges at greater risk of deterioration, the model facilitates proactive maintenance strategies, which can help avoid costly repairs and minimize service disruptions. Additionally, this research underscores the value of data-driven decision-making, enabling better resource allocation to prioritize maintenance efforts where they are most necessary. In summary, this study highlights the efficiency and applicability of Random Forest in predictive modeling for bridge management. Ultimately, these findings pave the way for more resilient and proactive management of bridge systems, ensuring their longevity and reliability for future use.

Keywords: data analysis, random forest, predictive modeling, bridge management

Procedia PDF Downloads 23
712 Analyzing the Influence of Hydrometeorlogical Extremes, Geological Setting, and Social Demographic on Public Health

Authors: Irfan Ahmad Afip

Abstract:

This main research objective is to accurately identify the possibility for a Leptospirosis outbreak severity of a certain area based on its input features into a multivariate regression model. The research question is the possibility of an outbreak in a specific area being influenced by this feature, such as social demographics and hydrometeorological extremes. If the occurrence of an outbreak is being subjected to these features, then the epidemic severity for an area will be different depending on its environmental setting because the features will influence the possibility and severity of an outbreak. Specifically, this research objective was three-fold, namely: (a) to identify the relevant multivariate features and visualize the patterns data, (b) to develop a multivariate regression model based from the selected features and determine the possibility for Leptospirosis outbreak in an area, and (c) to compare the predictive ability of multivariate regression model and machine learning algorithms. Several secondary data features were collected locations in the state of Negeri Sembilan, Malaysia, based on the possibility it would be relevant to determine the outbreak severity in the area. The relevant features then will become an input in a multivariate regression model; a linear regression model is a simple and quick solution for creating prognostic capabilities. A multivariate regression model has proven more precise prognostic capabilities than univariate models. The expected outcome from this research is to establish a correlation between the features of social demographic and hydrometeorological with Leptospirosis bacteria; it will also become a contributor for understanding the underlying relationship between the pathogen and the ecosystem. The relationship established can be beneficial for the health department or urban planner to inspect and prepare for future outcomes in event detection and system health monitoring.

Keywords: geographical information system, hydrometeorological, leptospirosis, multivariate regression

Procedia PDF Downloads 115
711 Assessment of Zinc Content in Nuts by Atomic Absorption Spectrometry Method

Authors: Katarzyna Socha, Konrad Mielcarek, Grzegorz Kangowski, Renata Markiewicz-Zukowska, Anna Puscion-Jakubik, Jolanta Soroczynska, Maria H. Borawska

Abstract:

Nuts have high nutritional value. They are a good source of polyunsaturated fatty acids, dietary fiber, vitamins (B₁, B₆, E, K) and minerals: magnesium, selenium, zinc (Zn). Zn is an essential element for proper functioning and development of human organism. Due to antioxidant and anti-inflammatory properties, Zn has an influence on immunological and central nervous system. It also affects proper functioning of reproductive organs and has beneficial impact on the condition of skin, hair, and nails. The objective of this study was estimation of Zn content in edible nuts. The research material consisted of 10 types of nuts, 12 samples of each type: almonds, brazil nuts, cashews, hazelnuts, macadamia nuts, peanuts, pecans, pine nuts, pistachios, and walnuts. The samples of nuts were digested in concentrated nitric acid using microwave mineralizer (Berghof, Germany). The concentration of Zn was determined by flame atomic absorption spectrometry method with Zeeman background correction (Hitachi, Japan). The accuracy of the method was verified on certified reference material: Simulated Diet D. The statistical analysis was performed using Statistica v. 13.0 software. For comparison between the groups, t-Student test was used. The highest content of Zn was shown in pine nuts and cashews: 78.57 ± 21.9, 70.02 ± 10,2 mg/kg, respectively, significantly higher than in other types of nuts. The lowest content of Zn was found in macadamia nuts: 16.25 ± 4.1 mg/kg. The consumption of a standard 42-gram portion of almonds, brazil nuts, cashews, peanuts, pecans, and pine nuts covers the daily requirement for Zn above 15% of recommended daily allowances (RDA) for women, while in the case of men consumption all of the above types of nuts, except peanuts. Selected types of nuts can be a good source of Zn in the diet.

Keywords: atomic absorption spectrometry, microelement, nuts, zinc

Procedia PDF Downloads 196
710 Climate Change Impact Due to Timber Product Imports in the UK

Authors: Juan A. Ferriz-Papi, Allan L. Nantel, Talib E. Butt

Abstract:

Buildings are thought to consume about 50% of the total energy in the UK. The use stage in a building life cycle has the largest energy consumption, although different assessments are showing that the construction can equal several years of maintenance and operations. The selection of materials with lower embodied energy is very important to reduce this consumption. For this reason, timber is one adequate material due to its low embodied energy and the capacity to be used as carbon storage. The use of timber in the construction industry is very significant. Sawn wood, for example, is one of the top 5 construction materials consumed in the UK according to National Statistics. Embodied energy for building products considers the energy consumed in extraction and production stages. However, it is not the same consideration if this product is produced locally as when considering the resource produced further afield. Transport is a very relevant matter that profoundly influences in the results of embodied energy. The case of timber use in the UK is important because the balance between imports and exports is far negative, industry consuming more imported timber than produced. Nearly 80% of sawn softwood used in construction is imported. The imports-exports deficit for sawn wood accounted for more than 180 million pounds during the first four-month period of 2016. More than 85% of these imports come from Europe (83% from the EU). The aim of this study is to analyze climate change impact due to transport for timber products consumed in the UK. An approximate estimation of energy consumed and carbon emissions are calculated considering the timber product’s import origin. The results are compared to the total consumption of each product, estimating the impact of transport on the final embodied energy and carbon emissions. The analysis of these results can help deduce that one big challenge for climate change is the reduction of external dependency, with the associated improvement of internal production of timber products. A study of different types of timber products produced in the UK and abroad is developed to understand the possibilities for this country to improve sustainability and self-management. Reuse and recycle possibilities are also considered.

Keywords: embodied energy, climate change, CO2 emissions, timber, transport

Procedia PDF Downloads 344
709 Benchmarking Machine Learning Approaches for Forecasting Hotel Revenue

Authors: Rachel Y. Zhang, Christopher K. Anderson

Abstract:

A critical aspect of revenue management is a firm’s ability to predict demand as a function of price. Historically hotels have used simple time series models (regression and/or pick-up based models) owing to the complexities of trying to build casual models of demands. Machine learning approaches are slowly attracting attention owing to their flexibility in modeling relationships. This study provides an overview of approaches to forecasting hospitality demand – focusing on the opportunities created by machine learning approaches, including K-Nearest-Neighbors, Support vector machine, Regression Tree, and Artificial Neural Network algorithms. The out-of-sample performances of above approaches to forecasting hotel demand are illustrated by using a proprietary sample of the market level (24 properties) transactional data for Las Vegas NV. Causal predictive models can be built and evaluated owing to the availability of market level (versus firm level) data. This research also compares and contrast model accuracy of firm-level models (i.e. predictive models for hotel A only using hotel A’s data) to models using market level data (prices, review scores, location, chain scale, etc… for all hotels within the market). The prospected models will be valuable for hotel revenue prediction given the basic characters of a hotel property or can be applied in performance evaluation for an existed hotel. The findings will unveil the features that play key roles in a hotel’s revenue performance, which would have considerable potential usefulness in both revenue prediction and evaluation.

Keywords: hotel revenue, k-nearest-neighbors, machine learning, neural network, prediction model, regression tree, support vector machine

Procedia PDF Downloads 133
708 Optimizing CNC Production Line Efficiency Using NSGA-II: Adaptive Layout and Operational Sequence for Enhanced Manufacturing Flexibility

Authors: Yi-Ling Chen, Dung-Ying Lin

Abstract:

In the manufacturing process, computer numerical control (CNC) machining plays a crucial role. CNC enables precise machinery control through computer programs, achieving automation in the production process and significantly enhancing production efficiency. However, traditional CNC production lines often require manual intervention for loading and unloading operations, which limits the production line's operational efficiency and production capacity. Additionally, existing CNC automation systems frequently lack sufficient intelligence and fail to achieve optimal configuration efficiency, resulting in the need for substantial time to reconfigure production lines when producing different products, thereby impacting overall production efficiency. Using the NSGA-II algorithm, we generate production line layout configurations that consider field constraints and select robotic arm specifications from an arm list. This allows us to calculate loading and unloading times for each job order, perform demand allocation, and assign processing sequences. The NSGA-II algorithm is further employed to determine the optimal processing sequence, with the aim of minimizing demand completion time and maximizing average machine utilization. These objectives are used to evaluate the performance of each layout, ultimately determining the optimal layout configuration. By employing this method, it enhance the configuration efficiency of CNC production lines and establish an adaptive capability that allows the production line to respond promptly to changes in demand. This will minimize production losses caused by the need to reconfigure the layout, ensuring that the CNC production line can maintain optimal efficiency even when adjustments are required due to fluctuating demands.

Keywords: evolutionary algorithms, multi-objective optimization, pareto optimality, layout optimization, operations sequence

Procedia PDF Downloads 24
707 Attenuation of Endotoxin Induced Hepatotoxicity by Dexamethasone, Melatonin and Pentoxifylline in White Albino Mice: A Comparative Study

Authors: Ammara Khan

Abstract:

Sepsis is characterized by an overwhelming surge of cytokines and oxidative stress to one of many factors, gram-negative bacteria commonly implicated. Despite major expansion and elaboration of sepsis pathophysiology and therapeutic approach; death rate remains very high in septic patients due to multiple organ damages including hepatotoxicity.The present study was aimed to ascertain the adequacy of three different drugs delivered separately and collectively- low dose steroid-dexamethasone (3mg/kg i.p) ,antioxidant-melatonin(10 mg/kg i.p) ,and phosphodiesterases inhibitor - pentoxifylline (75 mg/kg i.p)in endotoxin-induced hepatotoxicity in mice. Endotoxin/lipopolysaccharides induced hepatotoxicity was reproduced in mice by giving lipopolysaccharide of serotype E.Coli intraperitoneally. The preventive role was questioned by giving the experimental agent half an hour prior to LPS injection whereas the therapeutic potential of the experimental agent was searched out via post-LPS delivering. The extent of liver damage was adjudged via serum alanine aminotransferases (ALT) and aspartate aminotransferase (AST) estimation along with a histopathological examination of liver tissue. Dexamethasone is given before (Group 3) and after LPS (group 4) significantly attenuated LPS generated liver injury.Pentoxifylline generated similar results and serum ALT; AST histological alteration abated considerably (p≤ 0.05) both in animals subjected to pentoxifylline pre (Group 5) and post-treatment(Group 6). Melatonin was also prosperous in aversion (Group 7) and curation (Group 8) of LPS invoked hepatotoxicity as evident by lessening of augmented ALT (≤0.01) and AST (≤0.01) along with restoration of pathological changes in liver sections (p≤0.05). Combination therapies with dexamethasone in conjunction with melatonin (Group 9), dexamethasone together with pentoxifylline (Group 10), and pentoxifylline along with melatonin (Group 11) after LPS administration tapered LPS evoked hepatic dysfunction statistically considerably. In conclusion, both melatonin and pentoxifylline set up promising results in endotoxin-induced hepatotoxicity and can be used therapeutic adjuncts to conventional treatment strategies in sepsis-induced liver failure.

Keywords: endotoxin/lipopolysacchride, dexamethasone, hepatotoxicity, melatonin, pentoxifylline

Procedia PDF Downloads 280
706 Numerical Modeling and Prediction of Nanoscale Transport Phenomena in Vertically Aligned Carbon Nanotube Catalyst Layers by the Lattice Boltzmann Simulation

Authors: Seungho Shin, Keunwoo Choi, Ali Akbar, Sukkee Um

Abstract:

In this study, the nanoscale transport properties and catalyst utilization of vertically aligned carbon nanotube (VACNT) catalyst layers are computationally predicted by the three-dimensional lattice Boltzmann simulation based on the quasi-random nanostructural model in pursuance of fuel cell catalyst performance improvement. A series of catalyst layers are randomly generated with statistical significance at the 95% confidence level to reflect the heterogeneity of the catalyst layer nanostructures. The nanoscale gas transport phenomena inside the catalyst layers are simulated by the D3Q19 (i.e., three-dimensional, 19 velocities) lattice Boltzmann method, and the corresponding mass transport characteristics are mathematically modeled in terms of structural properties. Considering the nanoscale reactant transport phenomena, a transport-based effective catalyst utilization factor is defined and statistically analyzed to determine the structure-transport influence on catalyst utilization. The tortuosity of the reactant mass transport path of VACNT catalyst layers is directly calculated from the streaklines. Subsequently, the corresponding effective mass diffusion coefficient is statistically predicted by applying the pre-estimated tortuosity factors to the Knudsen diffusion coefficient in the VACNT catalyst layers. The statistical estimation results clearly indicate that the morphological structures of VACNT catalyst layers reduce the tortuosity of reactant mass transport path when compared to conventional catalyst layer and significantly improve consequential effective mass diffusion coefficient of VACNT catalyst layer. Furthermore, catalyst utilization of the VACNT catalyst layer is substantially improved by enhanced mass diffusion and electric current paths despite the relatively poor interconnections of the ion transport paths.

Keywords: Lattice Boltzmann method, nano transport phenomena, polymer electrolyte fuel cells, vertically aligned carbon nanotube

Procedia PDF Downloads 201
705 A Comprehensive Theory of Communication with Biological and Non-Biological Intelligence for a 21st Century Curriculum

Authors: Thomas Schalow

Abstract:

It is commonly recognized that our present curriculum is not preparing students to function in the 21st century. This is particularly true in regard to communication needs across cultures - both human and non-human. In this paper, a comprehensive theory of communication-based on communication with non-human cultures and intelligences is presented to meet the following three imminent contingencies: communicating with sentient biological intelligences, communicating with extraterrestrial intelligences, and communicating with artificial super-intelligences. The paper begins with the argument that we need to become much more serious about communicating with the non-human, intelligent life forms that already exists around us here on Earth. We need to broaden our definition of communication and reach out to other sentient life forms in order to provide humanity with a better perspective of its place within our ecosystem. The paper next examines the science and philosophy behind CETI (communication with extraterrestrial intelligences) and how it could prove useful even in the absence of contact with alien life. However, CETI’s assumptions and methodology need to be revised in accordance with the communication theory being proposed in this paper if we are truly serious about finding and communicating with life beyond Earth. The final theme explored in this paper is communication with non-biological super-intelligences. Humanity has never been truly compelled to converse with other species, and our failure to seriously consider such intercourse has left us largely unprepared to deal with communication in a future that will be mediated and controlled by computer algorithms. Fortunately, our experience dealing with other cultures can provide us with a framework for this communication. The basic concepts behind intercultural communication can be applied to the three types of communication envisioned in this paper if we are willing to recognize that we are in fact dealing with other cultures when we interact with other species, alien life, and artificial super-intelligence. The ideas considered in this paper will require a new mindset for humanity, but a new disposition will yield substantial gains. A curriculum that is truly ready for the 21st century needs to be aligned with this new theory of communication.

Keywords: artificial intelligence, CETI, communication, language

Procedia PDF Downloads 364
704 Implementation of Integrated Multi-Channel Analysis of Surface Waves and Waveform Inversion Techniques for Seismic Hazard Estimation with Emphasis on Associated Uncertainty: A Case Study at Zafarana Wind Turbine Towers Farm, Egypt

Authors: Abd El-Aziz Khairy Abd El-Aal, Yuji Yagi, Heba Kamal

Abstract:

In this study, an integrated multi-channel analysis of Surface Waves (MASW) technique is applied to explore the geotechnical parameters of subsurface layers at the Zafarana wind farm. Moreover, a seismic hazard procedure based on the extended deterministic technique is used to estimate the seismic hazard load for the investigated area. The study area includes many active fault systems along the Gulf of Suez that cause many moderate and large earthquakes. Overall, the seismic activity of the area has recently become better understood following the use of new waveform inversion methods and software to develop accurate focal mechanism solutions for recent recorded earthquakes around the studied area. These earthquakes resulted in major stress-drops in the Eastern desert and the Gulf of Suez area. These findings have helped to reshape the understanding of the seismotectonic environment of the Gulf of Suez area, which is a perplexing tectonic domain. Based on the collected new information and data, this study uses an extended deterministic approach to re-examine the seismic hazard for the Gulf of Suez region, particularly the wind turbine towers at Zafarana Wind Farm and its vicinity. Alternate seismic source and magnitude-frequency relationships were combined with various indigenous attenuation relationships, adapted within a logic tree formulation, to quantify and project the regional exposure on a set of hazard maps. We select two desired exceedance probabilities (10 and 20%) that any of the applied scenarios may exceed the largest median ground acceleration. The ground motion was calculated at 50th, 84th percentile levels.

Keywords: MASW, seismic hazard, wind turbine towers, Zafarana wind farm

Procedia PDF Downloads 403
703 Marginal Productivity of Small Scale Yam and Cassava Farmers in Kogi State, Nigeria: Data Envelopment Analysis as a Complement

Authors: M. A. Ojo, O. A. Ojo, A. I. Odine, A. Ogaji

Abstract:

The study examined marginal productivity analysis of small scale yam and cassava farmers in Kogi State, Nigeria. Data used for the study were obtained from primary source using a multi-stage sampling technique with structured questionnaires administered to 150 randomly selected yam and cassava farmers from three Local Government Areas of the State. Description statistics, data envelopment analysis and Cobb-Douglas production function were used to analyze the data. The DEA result on the overall technical efficiency of the farmers showed that 40% of the sampled yam and cassava farmers in the study area were operating at frontier and optimum level of production with mean technical efficiency of 1.00. This implies that 60% of the yam and cassava farmers in the study area can still improve their level of efficiency through better utilization of available resources, given the current state of technology. The results of the Cobb-Douglas analysis of factors affecting the output of yam and cassava farmers showed that labour, planting materials, fertilizer and capital inputs positively and significantly affected the output of the yam and cassava farmers in the study area. The study further revealed that yam and cassava farms in the study area operated under increasing returns to scale. This result of marginal productivity analysis further showed that relatively efficient farms were more marginally productive in resource utilization This study also shows that estimating production functions without separating the farms to efficient and inefficient farms bias the parameter values obtained from such production function. It is therefore recommended that yam and cassava farmers in the study area should form cooperative societies so as to enable them have access to productive inputs that will enable them expand. Also, since using a single equation model for production function produces a bias parameter estimates as confirmed above, farms should, therefore, be decomposed into efficient and inefficient ones before production function estimation is done.

Keywords: marginal productivity, DEA, production function, Kogi state

Procedia PDF Downloads 483
702 Multi-Impairment Compensation Based Deep Neural Networks for 16-QAM Coherent Optical Orthogonal Frequency Division Multiplexing System

Authors: Ying Han, Yuanxiang Chen, Yongtao Huang, Jia Fu, Kaile Li, Shangjing Lin, Jianguo Yu

Abstract:

In long-haul and high-speed optical transmission system, the orthogonal frequency division multiplexing (OFDM) signal suffers various linear and non-linear impairments. In recent years, researchers have proposed compensation schemes for specific impairment, and the effects are remarkable. However, different impairment compensation algorithms have caused an increase in transmission delay. With the widespread application of deep neural networks (DNN) in communication, multi-impairment compensation based on DNN will be a promising scheme. In this paper, we propose and apply DNN to compensate multi-impairment of 16-QAM coherent optical OFDM signal, thereby improving the performance of the transmission system. The trained DNN models are applied in the offline digital signal processing (DSP) module of the transmission system. The models can optimize the constellation mapping signals at the transmitter and compensate multi-impairment of the OFDM decoded signal at the receiver. Furthermore, the models reduce the peak to average power ratio (PAPR) of the transmitted OFDM signal and the bit error rate (BER) of the received signal. We verify the effectiveness of the proposed scheme for 16-QAM Coherent Optical OFDM signal and demonstrate and analyze transmission performance in different transmission scenarios. The experimental results show that the PAPR and BER of the transmission system are significantly reduced after using the trained DNN. It shows that the DNN with specific loss function and network structure can optimize the transmitted signal and learn the channel feature and compensate for multi-impairment in fiber transmission effectively.

Keywords: coherent optical OFDM, deep neural network, multi-impairment compensation, optical transmission

Procedia PDF Downloads 143
701 Quoting Jobshops Due Dates Subject to Exogenous Factors in Developing Nations

Authors: Idris M. Olatunde, Kareem B.

Abstract:

In manufacturing systems, especially job shops, service performance is a key factor that determines customer satisfaction. Service performance depends not only on the quality of the output but on the delivery lead times as well. Besides product quality enhancement, delivery lead time must be minimized for optimal patronage. Quoting accurate due dates is sine quo non for job shop operational survival in a global competitive environment. Quoting accurate due dates in job shops has been a herculean task that nearly defiled solutions from many methods employed due to complex jobs routing nature of the system. This class of NP-hard problems possessed no rigid algorithms that can give an optimal solution. Jobshop operational problem is more complex in developing nations due to some peculiar factors. Operational complexity in job shops emanated from political instability, poor economy, technological know-how, and the non-promising socio-political environment. The mentioned exogenous factors were hardly considered in the previous studies on scheduling problem related to due date determination in job shops. This study has filled the gap created in the past studies by developing a dynamic model that incorporated the exogenous factors for accurate determination of due dates for varying jobs complexity. Real data from six job shops selected from the different part of Nigeria, were used to test the efficacy of the model, and the outcomes were analyzed statistically. The results of the analyzes showed that the model is more promising in determining accurate due dates than the traditional models deployed by many job shops in terms of patronage and lead times minimization.

Keywords: due dates prediction, improved performance, customer satisfaction, dynamic model, exogenous factors, job shops

Procedia PDF Downloads 412
700 Transmission Line Congestion Management Using Hybrid Fish-Bee Algorithm with Unified Power Flow Controller

Authors: P. Valsalal, S. Thangalakshmi

Abstract:

There is a widespread changeover in the electrical power industry universally from old-style monopolistic outline towards a horizontally distributed competitive structure to come across the demand of rising consumption. When the transmission lines of derestricted system are incapable to oblige the entire service needs, the lines are overloaded or congested. The governor between customer and power producer is nominated as Independent System Operator (ISO) to lessen the congestion without obstructing transmission line restrictions. Among the existing approaches for congestion management, the frequently used approaches are reorganizing the generation and load curbing. There is a boundary for reorganizing the generators, and further loads may not be supplemented with the prevailing resources unless more private power producers are added in the system by considerably raising the cost. Hence, congestion is relaxed by appropriate Flexible AC Transmission Systems (FACTS) devices which boost the existing transfer capacity of transmission lines. The FACTs device, namely, Unified Power Flow Controller (UPFC) is preferred, and the correct placement of UPFC is more vital and should be positioned in the highly congested line. Hence, the weak line is identified by using power flow performance index with the new objective function with proposed hybrid Fish – Bee algorithm. Further, the location of UPFC at appropriate line reduces the branch loading and minimizes the voltage deviation. The power transfer capacity of lines is determined with and without UPFC in the identified congested line of IEEE 30 bus structure and the simulated results are compared with prevailing algorithms. It is observed that the transfer capacity of existing line is increased with the presented algorithm and thus alleviating the congestion.

Keywords: available line transfer capability, congestion management, FACTS device, Hybrid Fish-Bee Algorithm, ISO, UPFC

Procedia PDF Downloads 383
699 Advantages of Neural Network Based Air Data Estimation for Unmanned Aerial Vehicles

Authors: Angelo Lerro, Manuela Battipede, Piero Gili, Alberto Brandl

Abstract:

Redundancy requirements for UAV (Unmanned Aerial Vehicle) are hardly faced due to the generally restricted amount of available space and allowable weight for the aircraft systems, limiting their exploitation. Essential equipment as the Air Data, Attitude and Heading Reference Systems (ADAHRS) require several external probes to measure significant data as the Angle of Attack or the Sideslip Angle. Previous research focused on the analysis of a patented technology named Smart-ADAHRS (Smart Air Data, Attitude and Heading Reference System) as an alternative method to obtain reliable and accurate estimates of the aerodynamic angles. This solution is based on an innovative sensor fusion algorithm implementing soft computing techniques and it allows to obtain a simplified inertial and air data system reducing external devices. In fact, only one external source of dynamic and static pressures is needed. This paper focuses on the benefits which would be gained by the implementation of this system in UAV applications. A simplification of the entire ADAHRS architecture will bring to reduce the overall cost together with improved safety performance. Smart-ADAHRS has currently reached Technology Readiness Level (TRL) 6. Real flight tests took place on ultralight aircraft equipped with a suitable Flight Test Instrumentation (FTI). The output of the algorithm using the flight test measurements demonstrates the capability for this fusion algorithm to embed in a single device multiple physical and virtual sensors. Any source of dynamic and static pressure can be integrated with this system gaining a significant improvement in terms of versatility.

Keywords: aerodynamic angles, air data system, flight test, neural network, unmanned aerial vehicle, virtual sensor

Procedia PDF Downloads 221
698 Emotion Detection in Twitter Messages Using Combination of Long Short-Term Memory and Convolutional Deep Neural Networks

Authors: Bahareh Golchin, Nooshin Riahi

Abstract:

One of the most significant issues as attended a lot in recent years is that of recognizing the sentiments and emotions in social media texts. The analysis of sentiments and emotions is intended to recognize the conceptual information such as the opinions, feelings, attitudes and emotions of people towards the products, services, organizations, people, topics, events and features in the written text. These indicate the greatness of the problem space. In the real world, businesses and organizations are always looking for tools to gather ideas, emotions, and directions of people about their products, services, or events related to their own. This article uses the Twitter social network, one of the most popular social networks with about 420 million active users, to extract data. Using this social network, users can share their information and opinions about personal issues, policies, products, events, etc. It can be used with appropriate classification of emotional states due to the availability of its data. In this study, supervised learning and deep neural network algorithms are used to classify the emotional states of Twitter users. The use of deep learning methods to increase the learning capacity of the model is an advantage due to the large amount of available data. Tweets collected on various topics are classified into four classes using a combination of two Bidirectional Long Short Term Memory network and a Convolutional network. The results obtained from this study with an average accuracy of 93%, show good results extracted from the proposed framework and improved accuracy compared to previous work.

Keywords: emotion classification, sentiment analysis, social networks, deep neural networks

Procedia PDF Downloads 138
697 Prediction of Seismic Damage Using Scalar Intensity Measures Based on Integration of Spectral Values

Authors: Konstantinos G. Kostinakis, Asimina M. Athanatopoulou

Abstract:

A key issue in seismic risk analysis within the context of Performance-Based Earthquake Engineering is the evaluation of the expected seismic damage of structures under a specific earthquake ground motion. The assessment of the seismic performance strongly depends on the choice of the seismic Intensity Measure (IM), which quantifies the characteristics of a ground motion that are important to the nonlinear structural response. Several conventional IMs of ground motion have been used to estimate their damage potential to structures. Yet, none of them has been proved to be able to predict adequately the seismic damage. Therefore, alternative, scalar intensity measures, which take into account not only ground motion characteristics but also structural information have been proposed. Some of these IMs are based on integration of spectral values over a range of periods, in an attempt to account for the information that the shape of the acceleration, velocity or displacement spectrum provides. The adequacy of a number of these IMs in predicting the structural damage of 3D R/C buildings is investigated in the present paper. The investigated IMs, some of which are structure specific and some are nonstructure-specific, are defined via integration of spectral values. To achieve this purpose three symmetric in plan R/C buildings are studied. The buildings are subjected to 59 bidirectional earthquake ground motions. The two horizontal accelerograms of each ground motion are applied along the structural axes. The response is determined by nonlinear time history analysis. The structural damage is expressed in terms of the maximum interstory drift as well as the overall structural damage index. The values of the aforementioned seismic damage measures are correlated with seven scalar ground motion IMs. The comparative assessment of the results revealed that the structure-specific IMs present higher correlation with the seismic damage of the three buildings. However, the adequacy of the IMs for estimation of the structural damage depends on the response parameter adopted. Furthermore, it was confirmed that the widely used spectral acceleration at the fundamental period of the structure is a good indicator of the expected earthquake damage level.

Keywords: damage measures, bidirectional excitation, spectral based IMs, R/C buildings

Procedia PDF Downloads 328
696 A Simple Approach to Establish Urban Energy Consumption Map Using the Combination of LiDAR and Thermal Image

Authors: Yu-Cheng Chen, Tzu-Ping Lin, Feng-Yi Lin, Chih-Yu Chen

Abstract:

Due to the urban heat island effect caused by highly development of city, the heat stress increased in recent year rapidly. Resulting in a sharp raise of the energy used in urban area. The heat stress during summer time exacerbated the usage of air conditioning and electric equipment, which caused more energy consumption and anthropogenic heat. Therefore, an accurate and simple method to measure energy used in urban area can be helpful for the architectures and urban planners to develop better energy efficiency goals. This research applies the combination of airborne LiDAR data and thermal imager to provide an innovate method to estimate energy consumption. Owing to the high resolution of remote sensing data, the accurate current volume and total floor area and the surface temperature of building derived from LiDAR and thermal imager can be herein obtained to predict energy used. In the estimate process, the LiDAR data will be divided into four type of land cover which including building, road, vegetation, and other obstacles. In this study, the points belong to building were selected to overlay with the land use information; therefore, the energy consumption can be estimated precisely with the real value of total floor area and energy use index for different use of building. After validating with the real energy used data from the government, the result shows the higher building in high development area like commercial district will present in higher energy consumption, caused by the large quantity of total floor area and more anthropogenic heat. Furthermore, because of the surface temperature can be warm up by electric equipment used, this study also applies the thermal image of building to find the hot spots of energy used and make the estimation method more complete.

Keywords: urban heat island, urban planning, LiDAR, thermal imager, energy consumption

Procedia PDF Downloads 239
695 Non-Destructive Static Damage Detection of Structures Using Genetic Algorithm

Authors: Amir Abbas Fatemi, Zahra Tabrizian, Kabir Sadeghi

Abstract:

To find the location and severity of damage that occurs in a structure, characteristics changes in dynamic and static can be used. The non-destructive techniques are more common, economic, and reliable to detect the global or local damages in structures. This paper presents a non-destructive method in structural damage detection and assessment using GA and static data. Thus, a set of static forces is applied to some of degrees of freedom and the static responses (displacements) are measured at another set of DOFs. An analytical model of the truss structure is developed based on the available specification and the properties derived from static data. The damages in structure produce changes to its stiffness so this method used to determine damage based on change in the structural stiffness parameter. Changes in the static response which structural damage caused choose to produce some simultaneous equations. Genetic Algorithms are powerful tools for solving large optimization problems. Optimization is considered to minimize objective function involve difference between the static load vector of damaged and healthy structure. Several scenarios defined for damage detection (single scenario and multiple scenarios). The static damage identification methods have many advantages, but some difficulties still exist. So it is important to achieve the best damage identification and if the best result is obtained it means that the method is Reliable. This strategy is applied to a plane truss. This method is used for a plane truss. Numerical results demonstrate the ability of this method in detecting damage in given structures. Also figures show damage detections in multiple damage scenarios have really efficient answer. Even existence of noise in the measurements doesn’t reduce the accuracy of damage detections method in these structures.

Keywords: damage detection, finite element method, static data, non-destructive, genetic algorithm

Procedia PDF Downloads 237
694 Investigation of Shear Strength, and Dilative Behavior of Coarse-grained Samples Using Laboratory Test and Machine Learning Technique

Authors: Ehsan Mehryaar, Seyed Armin Motahari Tabari

Abstract:

Coarse-grained soils are known and commonly used in a wide range of geotechnical projects, including high earth dams or embankments for their high shear strength. The most important engineering property of these soils is friction angle which represents the interlocking between soil particles and can be applied widely in designing and constructing these earth structures. Friction angle and dilative behavior of coarse-grained soils can be estimated from empirical correlations with in-situ testing and physical properties of the soil or measured directly in the laboratory performing direct shear or triaxial tests. Unfortunately, large-scale testing is difficult, challenging, and expensive and is not possible in most soil mechanic laboratories. So, it is common to remove the large particles and do the tests, which cannot be counted as an exact estimation of the parameters and behavior of the original soil. This paper describes a new methodology to simulate particles grading distribution of a well-graded gravel sample to a smaller scale sample as it can be tested in an ordinary direct shear apparatus to estimate the stress-strain behavior, friction angle, and dilative behavior of the original coarse-grained soil considering its confining pressure, and relative density using a machine learning method. A total number of 72 direct shear tests are performed in 6 different sizes, 3 different confining pressures, and 4 different relative densities. Multivariate Adaptive Regression Spline (MARS) technique was used to develop an equation in order to predict shear strength and dilative behavior based on the size distribution of coarse-grained soil particles. Also, an uncertainty analysis was performed in order to examine the reliability of the proposed equation.

Keywords: MARS, coarse-grained soil, shear strength, uncertainty analysis

Procedia PDF Downloads 162
693 Facilitating Written Biology Assessment in Large-Enrollment Courses Using Machine Learning

Authors: Luanna B. Prevost, Kelli Carter, Margaurete Romero, Kirsti Martinez

Abstract:

Writing is an essential scientific practice, yet, in several countries, the increasing university science class-size limits the use of written assessments. Written assessments allow students to demonstrate their learning in their own words and permit the faculty to evaluate students’ understanding. However, the time and resources required to grade written assessments prohibit their use in large-enrollment science courses. This study examined the use of machine learning algorithms to automatically analyze student writing and provide timely feedback to the faculty about students' writing in biology. Written responses to questions about matter and energy transformation were collected from large-enrollment undergraduate introductory biology classrooms. Responses were analyzed using the LightSide text mining and classification software. Cohen’s Kappa was used to measure agreement between the LightSide models and human raters. Predictive models achieved agreement with human coding of 0.7 Cohen’s Kappa or greater. Models captured that when writing about matter-energy transformation at the ecosystem level, students focused on primarily on the concepts of heat loss, recycling of matter, and conservation of matter and energy. Models were also produced to capture writing about processes such as decomposition and biochemical cycling. The models created in this study can be used to provide automatic feedback about students understanding of these concepts to biology faculty who desire to use formative written assessments in larger enrollment biology classes, but do not have the time or personnel for manual grading.

Keywords: machine learning, written assessment, biology education, text mining

Procedia PDF Downloads 281