Search results for: mutant sets
392 House Price Index Predicts a Larger Impact of Habitat Loss than Primary Productivity on the Biodiversity of North American Avian Communities
Authors: Marlen Acosta Alamo, Lisa Manne, Richard Veit
Abstract:
Habitat loss due to land use change is one of the leading causes of biodiversity loss worldwide. This form of habitat loss is a non-random phenomenon since the same environmental factors that make an area suitable for supporting high local biodiversity overlap with those that make it attractive for urban development. We aimed to compare the effect of two non-random habitat loss predictors on the richness, abundance, and rarity of nature-affiliated and human-affiliated North American breeding birds. For each group of birds, we simulated the non-random habitat loss using two predictors: the House Price Index as a measure of the attractiveness of an area for humans and the Normalized Difference Vegetation Index as a proxy for primary productivity. We compared the results of the two non-random simulation sets and one set of random habitat loss simulations using an analysis of variance and followed up with a Tukey-Kramer test when appropriate. The attractiveness of an area for humans predicted estimates of richness loss and increase of rarity higher than primary productivity and random habitat loss for nature-affiliated and human-affiliated birds. For example, at 50% of habitat loss, the attractiveness of an area for humans produced estimates of richness at least 5% lower and of a rarity at least 40% higher than primary productivity and random habitat loss for both groups of birds. Only for the species abundance of nature-affiliated birds, the attractiveness of an area for humans did not outperform primary productivity as a predictor of biodiversity following habitat loss. We demonstrated the value of the House Price Index, which can be used in conservation assessments as an index of the risks of habitat loss for natural communities. Thus, our results have relevant implications for sustainable urban land-use planning practices and can guide stakeholders and developers in their efforts to conserve local biodiversity.Keywords: biodiversity loss, bird biodiversity, house price index, non-random habitat loss
Procedia PDF Downloads 87391 Pilot Study of Determining the Impact of Surface Subsidence at The Intersection of Cave Mining with the Surface Using an Electrical Impedance Tomography
Authors: Ariungerel Jargal
Abstract:
: Cave mining is a bulk underground mining method, which allows large low-grade deposits to be mined underground. This method involves undermining the orebody to make it collapse under its own weight into a series of chambers from which the ore extracted. It is a useful technique to extend the life of large deposits previously mined by open pits, and it is a method increasingly proposed for new mines around the world. We plan to conduct a feasibility study using Electrical impedance tomography (EIT) technology to show how much subsidence there is at the intersection with the cave mining surface. EIT is an imaging technique which uses electrical measurements at electrodes attached on the body surface to yield a cross-sectional image of conductivity changes within the object. EIT has been developed in several different applications areas as a simpler, cheaper alternative to many other imaging methods. A low frequency current is injected between pairs of electrodes while voltage measurements are collected at all other electrode pairs. In the difference EIT, images are reconstructed of the change in conductivity distribution (σ) between the acquisition of the two sets of measurements. Image reconstruction in EIT requires the solution of an ill-conditioned nonlinear inverse problem on noisy data, typically requiring make simpler assumptions or regularization. It is noted that the ratio of current to voltage represents a complex value according to Ohm’s law, and that it is theoretically possible to re-express EIT. The results of the experiment were presented on the simulation, and it was concluded that it is possible to conduct further real experiments. Drill a certain number of holes in the top wall of the cave to attach the electrodes, flow a current through them, and measure and acquire the potential through these electrodes. Appropriate values should be selected depending on the distance between the holes, the frequency and duration of the measurements, the surface characteristics and the size of the study area using an EIT device.Keywords: impedance tomography, cave mining, soil, EIT device
Procedia PDF Downloads 126390 Adopt and Apply Research-Supported Standards and Practices to Ensure Quality for Online Education and Digital Learning at Course, Program, and Institutional Levels
Authors: Yaping Gao
Abstract:
With the increasing globalization of education and the continued momentum and wider adoption of online education and digital learning all over the world, post pandemic, it is crucial that best practices and extensive experience and knowledge gained from the higher education community over the past few decades be adopted and adapted to benefit the broader international communities, which can be vastly different culturally and pedagogically. Schools and institutions worldwide should consider to adopt, adapt and apply these proven practices to develop strategic plans for digital transformation at institutional levels, and to improve or develop quality online or digital learning environments at course and program levels to help all students succeed. The presenter will introduce the primary components of the US-based quality assurance process, including: 1) five sets of research-supported standards to guide the design, development and review of online and hybrid courses; 2) professional development offerings and pathways for administrators, faculty and instructional support staff; 3) a peer-review process for course/program reviews resulting in constructive recommendations for continuous improvement, certification of quality and international recognition; and 4) implementation of the quality assurance process on a continuum to program excellence, achievement of institutional goals, and facilitation of accreditation process and success. Regardless language, culture, pedagogical practices, or technological infrastructure, the core elements of quality teaching and learning remain the same across all delivery formats. What is unique is how to ensure quality of teaching and learning in online education and digital learning. No one knows all the answers to everything but no one needs to reinvent the wheel either. Together the international education community can support and learn from each other to achieve institutional goals and ensure all students succeed in the digital learning environments.Keywords: online education, digital learning, quality standards, best practices, online teaching and learning
Procedia PDF Downloads 28389 Influence of HDI in the Spread of RSV Bronchiolitis in Children Aged 0 to 2 Years
Authors: Chloé Kernaléguen, Laura Kundun, Tessie Lery, Ryan Laleg, Zhangyun Tan
Abstract:
This study explores global disparities in respiratory syncytial virus (RSV) bronchiolitis incidence among children aged 0-2 years, focusing on the human development index (HDI) as a key determinant. RSV bronchiolitis poses a significant health risk to young children, influenced by factors, including socio-economic conditions captured by the HDI. Through a comprehensive systematic review and dataset selection (Switzerland, Brazil, United States of America), we formulated an HDI-SEIRS numerical model within the SEIRS framework. Results show variations in RSV bronchiolitis dynamics across countries, emphasizing the influence of HDI. Modelling reveals a correlation between higher HDI and increased bronchiolitis spread, notably in the USA and Switzerland. The ratios HDIcountry over HDImax strengthen this association, while climate disparities contribute to variations, especially in colder climates like the USA and Switzerland. The study raises the hypothesis of an indirect link between higher HDI and more frequent bronchiolitis, underlining the need for nuanced understanding. Factors like improved healthcare access, population density, mobility, and social behaviors in higher HDI countries might contribute to unexpected trends. Limitations include dataset quality and restricted RSV bronchiolitis data. Future research should encompass diverse HDI datasets to refine HDI's role in bronchiolitis dynamics. In conclusion, HDI-SEIRS models offer insights into factors influencing RSV bronchiolitis spread. While HDI is a significant indicator, its impact is indirect, necessitating a holistic approach to effective public health policies. This analysis sets the stage for further investigations into multifaceted interactions shaping bronchiolitis dynamics in diverse socio-economic contexts.Keywords: bronchiolitis propagation, HDI influence, respiratory syncytial virus, SEIRS model
Procedia PDF Downloads 67388 Adult Language Learning in the Institute of Technology Sector in the Republic of Ireland
Authors: Una Carthy
Abstract:
A recent study of third level institutions in Ireland reveals that both age and aptitude can be overcome by teaching methodologies to motivate second language learners. This PhD investigation gathered quantitative and qualitative data from 14 Institutes of Technology over a three years period from 2011 to 2014. The fundamental research question was to establish the impact of institutional language policy on attitudes towards language learning. However, other related issues around second language acquisition arose in the course of the investigation. Data were collected from both lectures and students, allowing interesting points of comparison to emerge from both datasets. Negative perceptions among lecturers regarding language provision were often associated with the view that language learning belongs to primary and secondary level and has no place in third level education. This perception was offset by substantial data showing positive attitudes towards adult language learning. Lenneberg’s Critical Age Theory postulated that the optimum age for learning a second language is before puberty. More recently, scholars have challenged this theory in their studies, revealing that mature learners can and do succeed at learning languages. With regard to aptitude, a preoccupation among lecturers regarding poor literacy skills among students emerged and was often associated with resistance to second language acquisition. This was offset by a preponderance of qualitative data from students highlighting the crucial role which teaching approaches play in the learning process. Interestingly, the data collected regarding learning disabilities reveals that, given the appropriate learning environments, individuals can be motivated to acquire second languages, and indeed succeed at learning them. These findings are in keeping with other recent studies regarding attitudes towards second language learning among students with learning disabilities. Both sets of findings reinforce the case for language policies in the Institute of Technology (IoTs). Supportive and positive learning environments can be created in third level institutions to motivate adult learners, thereby overcoming perceived obstacles relating to age and aptitude.Keywords: age, aptitude, second language acquisition, teaching methodologies
Procedia PDF Downloads 123387 Motif Search-Aided Screening of the Pseudomonas syringae pv. Maculicola Genome for Genes Encoding Tertiary Alcohol Ester Hydrolases
Authors: M. L. Mangena, N. Mokoena, K. Rashamuse, M. G. Tlou
Abstract:
Tertiary alcohol ester (TAE) hydrolases are a group of esterases (EC 3.1.1.-) that catalyze the kinetic resolution of TAEs and as a result, they are sought-after for the production of optically pure tertiary alcohols (TAs) which are useful as building blocks for number biologically active compounds. What sets these enzymes apart is, the presence of a GGG(A)X-motif in the active site which appears to be the main reason behind their activity towards the sterically demanding TAEs. The genome of Pseudomonas syringae pv. maculicola (Psm) comprises a multitude of genes that encode esterases. We therefore, hypothesize that some of these genes encode TAE hydrolases. In this study, Psm was screened for TAE hydrolase activity using the linalyl acetate (LA) plate assay and a positive reaction was observed. As a result, the genome of Psm was screened for esterases with a GGG(A)X-motif using the motif search tool and two potential TAE hydrolase genes (PsmEST1 and 2, 1100 and 1000bp, respectively) were identified, PsmEST1 was amplified by PCR and the gene sequenced for confirmation. Analysis of the sequence data with the SingnalP 4.1 server revealed that the protein comprises a signal peptide (22 amino acid residues) on the N-terminus. Primers specific for the gene encoding the mature protein (without the signal peptide) were designed such that they contain NdeI and XhoI restriction sites for directional cloning of the PCR products into pET28a. The gene was expressed in E. coli JM109 (DE3) and the clones screened for TAE hydrolase activity using the LA plate assay. A positive clone was selected, overexpressed and the protein purified using nickel affinity chromatography. The activity of the esterase towards LA was confirmed using thin layer chromatography.Keywords: hydrolases, tertiary alcohol esters, tertiary alcohols, screening, Pseudomonas syringae pv., maculicola genome, esterase activity, linalyl acetate
Procedia PDF Downloads 355386 Empowering Teachers to Bolster Vocational Education in Cameroon
Authors: Ambissah Asah Brigitte
Abstract:
This research is guided by observations in the types of education offered at the secondary level in Cameroon. The secondary education system in Cameroon comprises two types of education, including General Education and Technical and Vocational Education. Although General Education and, Technical and Vocational Education are given equal importance by public authorities, General Education remains on the thriving trend, enjoying the greatest enrolment. In the meantime, Technical and Vocational Education is still to reach the adequate momentum expected to fostering the country’s full-fledged development, as specified in the National Development Strategy, which is the blue print of State policies in Cameroon for the 2020-2030 decade. Vocational Education is credited for its ability to foster a country’s development, since it teaches students the precise skills and knowledge needed to carry out a specific craft, technical skill or trade. Yet, formal training on Vocational Education for teachers offers a pale face in secondary education. This limits the ability of the educational system to nurture vocations and provide the country’s economy with the manpower necessary to achieving development goals. This article seeks to analyse how concretely does the institutional framework spur vocational skills in secondary school teachers. It overviews the instruments instituting Vocational Education at the secondary level in Cameroon, then assesses their effective implementation on the ground. Questionnaires addressed to both active teachers and vocational education policy-makers serve to collect data which are analysed using descriptive statistics. The final objective is to contribute in the debate urging to rethink the role of teachers in bolstering Vocational Education, which is the cornerstone of industrial development. This is true everywhere in the world. In Cameroon and in Africa in general, teachers must be empowered in this field with specific sets of competencies they will need to pass on to learners. They equally need to be given opportunities to acquire and adapt their knowledge and teaching skills accordingly.Keywords: vocational education, cameroon, institutional framework, national development, competencies and skills
Procedia PDF Downloads 73385 Development of a Robust Protein Classifier to Predict EMT Status of Cervical Squamous Cell Carcinoma and Endocervical Adenocarcinoma (CESC) Tumors
Authors: ZhenlinJu, Christopher P. Vellano, RehanAkbani, Yiling Lu, Gordon B. Mills
Abstract:
The epithelial–mesenchymal transition (EMT) is a process by which epithelial cells acquire mesenchymal characteristics, such as profound disruption of cell-cell junctions, loss of apical-basolateral polarity, and extensive reorganization of the actin cytoskeleton to induce cell motility and invasion. A hallmark of EMT is its capacity to promote metastasis, which is due in part to activation of several transcription factors and subsequent downregulation of E-cadherin. Unfortunately, current approaches have yet to uncover robust protein marker sets that can classify tumors as possessing strong EMT signatures. In this study, we utilize reverse phase protein array (RPPA) data and consensus clustering methods to successfully classify a subset of cervical squamous cell carcinoma and endocervical adenocarcinoma (CESC) tumors into an EMT protein signaling group (EMT group). The overall survival (OS) of patients in the EMT group is significantly worse than those in the other Hormone and PI3K/AKT signaling groups. In addition to a shrinkage and selection method for linear regression (LASSO), we applied training/test set and Monte Carlo resampling approaches to identify a set of protein markers that predicts the EMT status of CESC tumors. We fit a logistic model to these protein markers and developed a classifier, which was fixed in the training set and validated in the testing set. The classifier robustly predicted the EMT status of the testing set with an area under the curve (AUC) of 0.975 by Receiver Operating Characteristic (ROC) analysis. This method not only identifies a core set of proteins underlying an EMT signature in cervical cancer patients, but also provides a tool to examine protein predictors that drive molecular subtypes in other diseases.Keywords: consensus clustering, TCGA CESC, Silhouette, Monte Carlo LASSO
Procedia PDF Downloads 468384 Deep Learning Prediction of Residential Radon Health Risk in Canada and Sweden to Prevent Lung Cancer Among Non-Smokers
Authors: Selim M. Khan, Aaron A. Goodarzi, Joshua M. Taron, Tryggve Rönnqvist
Abstract:
Indoor air quality, a prime determinant of health, is strongly influenced by the presence of hazardous radon gas within the built environment. As a health issue, dangerously high indoor radon arose within the 20th century to become the 2nd leading cause of lung cancer. While the 21st century building metrics and human behaviors have captured, contained, and concentrated radon to yet higher and more hazardous levels, the issue is rapidly worsening in Canada. It is established that Canadians in the Prairies are the 2nd highest radon-exposed population in the world, with 1 in 6 residences experiencing 0.2-6.5 millisieverts (mSv) radiation per week, whereas the Canadian Nuclear Safety Commission sets maximum 5-year occupational limits for atomic workplace exposure at only 20 mSv. This situation is also deteriorating over time within newer housing stocks containing higher levels of radon. Deep machine learning (LSTM) algorithms were applied to analyze multiple quantitative and qualitative features, determine the most important contributory factors, and predicted radon levels in the known past (1990-2020) and projected future (2021-2050). The findings showed gradual downwards patterns in Sweden, whereas it would continue to go from high to higher levels in Canada over time. The contributory factors found to be the basement porosity, roof insulation depthness, R-factor, and air dynamics of the indoor environment related to human window opening behaviour. Building codes must consider including these factors to ensure adequate indoor ventilation and healthy living that can prevent lung cancer in non-smokers.Keywords: radon, building metrics, deep learning, LSTM prediction model, lung cancer, canada, sweden
Procedia PDF Downloads 112383 Accumulation of Pollutants, Self-Purification and Impact on Peripheral Urban Areas: A Case Study in Shantytowns in Argentina
Authors: N. Porzionato, M. Mantiñan, E. Bussi, S. Grinberg, R. Gutierrez, G. Curutchet
Abstract:
This work sets out to debate the tensions involved in the processes of contamination and self-purification in the urban space, particularly in the streams that run through the Buenos Aires metropolitan area. For much of their course, those streams are piped; their waters do not come into contact with the outdoors until they have reached deeply impoverished urban areas with high levels of environmental contamination. These are peripheral zones that, until thirty years ago, were marshlands and fields. They are now densely populated areas largely lacking in urban infrastructure. The Cárcova neighborhood, where this project is underway, is in the José León Suárez section of General San Martín country, Buenos Aires province. A stretch of José León Suarez canal crosses the neighborhood. Starting upstream, this canal carries pollutants due to the sewage and industrial waste released into it. Further downstream, in the neighborhood, domestic drainage is poured into the stream. In this paper, we formulate a hypothesis diametrical to the one that holds that these neighborhoods are the primary source of contamination, suggesting instead that in the stretch of the canal that runs through the neighborhood the stream’s waters are actually cleaned and the sediments accumulate pollutants. Indeed, the stretches of water that runs through these neighborhoods act as water processing plants for the metropolis. This project has studied the different organic-load polluting contributions to the water in a certain stretch of the canal, the reduction of that load over the course of the canal, and the incorporation of pollutants into the sediments. We have found that the surface water has considerable ability to self-purify, mostly due to processes of sedimentation and adsorption. The polluting load is accumulated in the sediments where that load stabilizes slowly by means of anaerobic processes. In this study, we also investigated the risks of sediment management and the use of the processes studied here in controlled conditions as tools of environmental restoration.Keywords: bioremediation, pollutants, sediments, urban streams
Procedia PDF Downloads 441382 Investigating the Effectiveness of Multilingual NLP Models for Sentiment Analysis
Authors: Othmane Touri, Sanaa El Filali, El Habib Benlahmar
Abstract:
Natural Language Processing (NLP) has gained significant attention lately. It has proved its ability to analyze and extract insights from unstructured text data in various languages. It is found that one of the most popular NLP applications is sentiment analysis which aims to identify the sentiment expressed in a piece of text, such as positive, negative, or neutral, in multiple languages. While there are several multilingual NLP models available for sentiment analysis, there is a need to investigate their effectiveness in different contexts and applications. In this study, we aim to investigate the effectiveness of different multilingual NLP models for sentiment analysis on a dataset of online product reviews in multiple languages. The performance of several NLP models, including Google Cloud Natural Language API, Microsoft Azure Cognitive Services, Amazon Comprehend, Stanford CoreNLP, spaCy, and Hugging Face Transformers are being compared. The models based on several metrics, including accuracy, precision, recall, and F1 score, are being evaluated and compared to their performance across different categories of product reviews. In order to run the study, preprocessing of the dataset has been performed by cleaning and tokenizing the text data in multiple languages. Then training and testing each model has been applied using a cross-validation approach where randomly dividing the dataset into training and testing sets and repeating the process multiple times has been used. A grid search approach to optimize the hyperparameters of each model and select the best-performing model for each category of product reviews and language has been applied. The findings of this study provide insights into the effectiveness of different multilingual NLP models for Multilingual Sentiment Analysis and their suitability for different languages and applications. The strengths and limitations of each model were identified, and recommendations for selecting the most performant model based on the specific requirements of a project were provided. This study contributes to the advancement of research methods in multilingual NLP and provides a practical guide for researchers and practitioners in the field.Keywords: NLP, multilingual, sentiment analysis, texts
Procedia PDF Downloads 104381 A Tale of Seven Districts: Reviewing The Past, Present and Future of Patent Litigation Filings to Form a Two-Step Burden-Shifting Framework for 28 U.S.C. § 1404(a)
Authors: Timothy T. Hsieh
Abstract:
Current patent venue transfer laws under 28 U.S.C. § 1404(a) e.g., the Gilbert factors from Gulf Oil Corp. v. Gilbert, 330 U.S. 501 (1947) are too malleable in that they often lead to frequent mandamus orders from the U.S. Court of Appeals for the Federal Circuit (“Federal Circuit”) overturning district court rulings on venue transfer motions. Thus, this paper proposes a more robust two-step burden-shifting framework that replaces the eight Gilbert factors. Moreover, a brief history of venue transfer patterns in the seven most active federal patent district courts is covered, with special focus devoted to the venue transfer orders from Judge Alan D Albright of the U.S. District Court for the Western District of Texas. A comprehensive data summary of 45 case sets where the Federal Circuit ruled on writs of mandamus involving Judge Albright’s transfer orders is subsequently provided, with coverage summaries of certain cases including four precedential ones from the Federal Circuit. This proposed two-step burden shifting framework is then applied to these venue transfer cases, as well as Federal Circuit mandamus orders ruling on those decisions. Finally, alternative approaches to remedying the frequent reversals for venue transfer will be discussed, including potential legislative solutions, adjustments to common law framework approaches to venue transfer, deference to the inherent powers of Article III U.S. District Judge, and a unified federal patent district court. Overall, this paper seeks to offer a more robust and consistent three-step burden-shifting framework for venue transfer and for the Federal Circuit to follow in administering mandamus orders, which might change somewhat in light of Western District of Texas Chief Judge Orlando Garcia’s order on redistributing Judge Albright’s patent cases.Keywords: Patent law, venue, judge Alan Albright, minimum contacts, western district of Texas
Procedia PDF Downloads 109380 Grid and Market Integration of Large Scale Wind Farms using Advanced Predictive Data Mining Techniques
Authors: Umit Cali
Abstract:
The integration of intermittent energy sources like wind farms into the electricity grid has become an important challenge for the utilization and control of electric power systems, because of the fluctuating behaviour of wind power generation. Wind power predictions improve the economic and technical integration of large amounts of wind energy into the existing electricity grid. Trading, balancing, grid operation, controllability and safety issues increase the importance of predicting power output from wind power operators. Therefore, wind power forecasting systems have to be integrated into the monitoring and control systems of the transmission system operator (TSO) and wind farm operators/traders. The wind forecasts are relatively precise for the time period of only a few hours, and, therefore, relevant with regard to Spot and Intraday markets. In this work predictive data mining techniques are applied to identify a statistical and neural network model or set of models that can be used to predict wind power output of large onshore and offshore wind farms. These advanced data analytic methods helps us to amalgamate the information in very large meteorological, oceanographic and SCADA data sets into useful information and manageable systems. Accurate wind power forecasts are beneficial for wind plant operators, utility operators, and utility customers. An accurate forecast allows grid operators to schedule economically efficient generation to meet the demand of electrical customers. This study is also dedicated to an in-depth consideration of issues such as the comparison of day ahead and the short-term wind power forecasting results, determination of the accuracy of the wind power prediction and the evaluation of the energy economic and technical benefits of wind power forecasting.Keywords: renewable energy sources, wind power, forecasting, data mining, big data, artificial intelligence, energy economics, power trading, power grids
Procedia PDF Downloads 518379 High Input Driven Factors in Idea Campaigns in Large Organizations: A Case Depicting Best Practices
Authors: Babar Rasheed, Saad Ghafoor
Abstract:
Introduction: Idea campaigns are commonly held across organizations for generating employee engagement. The contribution is specifically designed to identify and solve prevalent issues. It is argued that numerous organizations fail to achieve their desired goals despite arranging for such campaigns and investing heavily in them. There are however practices that organizations use to achieve higher degree of effectiveness, and these practices may be up for exploration by research to make them usable for the other organizations. Purpose: The aim of this research is to surface the idea management practices of a leading electric company with global operations. The study involves a large sized, multi site organization that is attributed to have added challenges in terms of managing ideas from employees, in comparison to smaller organizations. The study aims to highlight the factors that are looked at as the idea management team strategies for the campaign, sets terms and rewards for it, makes follow up with the employees and lastly, evaluate and award ideas. Methodology: The study is conducted in a leading electric appliance corporation that has a large number of employees and is based in numerous regions of the world. A total of 7 interviews are carried out involving the chief innovation officer, innovation manager and members of idea management and evaluation teams. The interviews are carried out either on Skype or in-person based on the availability of the interviewee. Findings: While this being a working paper and while the study is under way, it is anticipated that valuable information is being achieved about specific details on how idea management systems are governed and how idea campaigns are carried out. The findings may be particularly useful for innovation consultants as resources they can use to promote idea campaigning. The usefulness of the best practices highlighted as a result is, in any case, the most valuable output of this study.Keywords: employee engagement, motivation, idea campaigns, large organizations, best practices, employees input, organizational output
Procedia PDF Downloads 174378 Fuzzy Logic Classification Approach for Exponential Data Set in Health Care System for Predication of Future Data
Authors: Manish Pandey, Gurinderjit Kaur, Meenu Talwar, Sachin Chauhan, Jagbir Gill
Abstract:
Health-care management systems are a unit of nice connection as a result of the supply a straightforward and fast management of all aspects relating to a patient, not essentially medical. What is more, there are unit additional and additional cases of pathologies during which diagnosing and treatment may be solely allotted by victimization medical imaging techniques. With associate ever-increasing prevalence, medical pictures area unit directly acquired in or regenerate into digital type, for his or her storage additionally as sequent retrieval and process. Data Mining is the process of extracting information from large data sets through using algorithms and Techniques drawn from the field of Statistics, Machine Learning and Data Base Management Systems. Forecasting may be a prediction of what's going to occur within the future, associated it's an unsure method. Owing to the uncertainty, the accuracy of a forecast is as vital because the outcome foretold by foretelling the freelance variables. A forecast management should be wont to establish if the accuracy of the forecast is within satisfactory limits. Fuzzy regression strategies have normally been wont to develop shopper preferences models that correlate the engineering characteristics with shopper preferences relating to a replacement product; the patron preference models offer a platform, wherever by product developers will decide the engineering characteristics so as to satisfy shopper preferences before developing the merchandise. Recent analysis shows that these fuzzy regression strategies area units normally will not to model client preferences. We tend to propose a Testing the strength of Exponential Regression Model over regression toward the mean Model.Keywords: health-care management systems, fuzzy regression, data mining, forecasting, fuzzy membership function
Procedia PDF Downloads 279377 Enhancing Learners' Metacognitive, Cultural and Linguistic Proficiency through Egyptian Series
Authors: Hanan Eltayeb, Reem Al Refaie
Abstract:
To be able to connect and relate to shows spoken in a foreign language, advanced learners must understand not only linguistics inferences but also cultural, metacognitive, and pragmatic connotations in colloquial Egyptian TV series. These connotations are needed to both understand the different facets of the dramas put before them, and they’re also consistently grown and formulated through watching these shows. The inferences have become a staple in the Egyptian colloquial culture over the years, making their way into day-to-day conversations as Egyptians use them to speak, relate, joke, and connect with each other, without having known one another from previous times. As for advanced learners, they need to understand these inferences not only to watch these shows, but also to be able to converse with Egyptians on a level that surpasses the formal, or standard. When faced with some of the somewhat recent shows on the Egyptian screens, learners faced challenges in understanding pragmatics, cultural, and religious background of the target language and consequently not able to interact effectively with a native speaker in real-life situations. This study aims to enhance the linguistic and cultural proficiency of learners through studying two genres of TV Colloquial Egyptian series. Study samples derived from two recent comedian and social Egyptian series ('The Seventh Neighbor' سابع جار, and 'Nelly and Sherihan' نيللي و شريهان). When learners watch such series, they are usually faced with a problem understanding inferences that have to do with social, religious, and political events that are addressed in the series. Using discourse analysis of the sematic, semantic, pragmatic, cultural, and linguistic characteristics of the target language, some major deductions were highlighted and repeated, showing a pattern in both. The research paper concludes that there are many sets of lingual and para-lingual phrases, idioms, and proverbs to be acquired and used effectively by teaching these series. The strategies adopted in the study can be applied to different types of media, like movies, TV shows, and even cartoons, to enhance student proficiency.Keywords: Egyptian series, culture, linguistic competence, pragmatics, semantics, social
Procedia PDF Downloads 143376 An Approach for Ensuring Data Flow in Freight Delivery and Management Systems
Authors: Aurelija Burinskienė, Dalė Dzemydienė, Arūnas Miliauskas
Abstract:
This research aims at developing the approach for more effective freight delivery and transportation process management. The road congestions and the identification of causes are important, as well as the context information recognition and management. The measure of many parameters during the transportation period and proper control of driver work became the problem. The number of vehicles per time unit passing at a given time and point for drivers can be evaluated in some situations. The collection of data is mainly used to establish new trips. The flow of the data is more complex in urban areas. Herein, the movement of freight is reported in detail, including the information on street level. When traffic density is extremely high in congestion cases, and the traffic speed is incredibly low, data transmission reaches the peak. Different data sets are generated, which depend on the type of freight delivery network. There are three types of networks: long-distance delivery networks, last-mile delivery networks and mode-based delivery networks; the last one includes different modes, in particular, railways and other networks. When freight delivery is switched from one type of the above-stated network to another, more data could be included for reporting purposes and vice versa. In this case, a significant amount of these data is used for control operations, and the problem requires an integrated methodological approach. The paper presents an approach for providing e-services for drivers by including the assessment of the multi-component infrastructure needed for delivery of freights following the network type. The construction of such a methodology is required to evaluate data flow conditions and overloads, and to minimize the time gaps in data reporting. The results obtained show the possibilities of the proposing methodological approach to support the management and decision-making processes with functionality of incorporating networking specifics, by helping to minimize the overloads in data reporting.Keywords: transportation networks, freight delivery, data flow, monitoring, e-services
Procedia PDF Downloads 126375 The Effect of Amendment of Soil with Rice Husk Charcoal Coated Urea and Rice Straw Compost on Nitrogen, Phosphorus and Potassium Leaching
Authors: D. A. S. Gamage, B. F. A. Basnayake, W. A. J. M. De Costa
Abstract:
Agriculture plays an important and strategic role in the performance of Sri Lankan national economy. Rice is the staple food of Sri Lankans thus; rice cultivation is the major agricultural activity of the country. In Sri Lanka, out of the total rice production, a considerable amount of rice straw and rice husk goes wasted. Hence, there is a great potential of production of quality compost and rice husk charcoal. The concept of making rice straw compost and rice husk charcoal is practicable in Sri Lanka, where more than 40% of the farmers are engaged in rice cultivation. The application of inorganic nitrogen fertilizer has become a burden to the country. Rice husk charcoal as a coating material to retain N fertilizer is a suitable solution to gradually release nitrogenous compounds. Objective of this study was to produce rice husk charcoal coated urea as a slow releasing fertilizer with rice straw compost and to compare the leaching losses of nitrogen, phosphorus and potassium using leaching columns. Leaching column studies were prepared using 1.2 m tall PVC pipes with a diameter of 15 cm and a sampling port was attached to the bottom end of the column-cap. Leachates (100 ml/leaching column) were obtained from two sets of (each set has four leaching columns) leaching columns. The sampling was done once a week for 3 month period. Rice husk charcoal coated urea can potentially be used as a slow releasing nitrogen fertilizer which reduces leaching losses of urea. It also helps reduce the phosphate and potassium leaching. The cyclic effect of phosphate release is an important finding which could be the central issue in defining microbial behavior in soils. The fluctuations of phosphate may have cyclic effects of 28 days. In addition, rice straw compost and rice husk charcoal coating is less costly and contribute to mitigate pollution of water bodies by inorganic fertilizers.Keywords: leaching, mitigate, rice husk charcoal, slow releasing fertilizer
Procedia PDF Downloads 326374 Sex and Sexuality Communication in African Families: The Dynamics of Openness and Closedness
Authors: Victorine Mbong Shu
Abstract:
Very little research exists on family sex and sexuality communication and identity formation and how communication is helping adolescents in forming their sexual identities in South Africa. This study is designed to explore the impact of sexual communication in African families and the dynamics that influence the openness or closedness of adolescent sexual identities. The primary objectives of this study are threefold; to understand how sexuality communication in African families impacts the closedness and/or openness of adolescent African identities; to explore the nature of African children's sexuality given the status of their families’ communications on sex; to describe how parental or adult sexual knowledge, attitudes, values of sex, etc. are translated to children in African families, if at all. This study seeks answers to challenges faced by African parents and caregivers of adolescent children in South Africa regarding sex-sexuality communication and their shifting identities in different spaces. Its outcome seeks to empower these families on how to continue to effectively communicate sex and sexuality at different stages and circumstances. Two sets of people are interviewed separately in order to explore issues of familial communication and how to understand how together with religion and culture, adolescents are socialised to form the social and gender identities that they take to adulthood. They were parents of adolescents and young adult children who spoke in retrospect on when they were adolescents. The results of this study will fill knowledge gaps considering the chosen theory of communication that gives clarity to the topic of sex and sexuality communication in African families in South Africa and the dynamics of privacy that influence identity formation. A subset of the 40 conversations, 5 female parents, 5 male parents, 5 young female adults, and 5 male young adults, was used for this analysis. The preliminary results revealed five emergent themes informed by research questions and the theoretical framework of this study: Open communication, Discomfort discussing sex and sexuality, The importance of sex communication to African parents, Factors influencing African families’ communication about sex and sexuality and Privacy and boundaries.Keywords: sex, sexuality, communication, African families, adolescents
Procedia PDF Downloads 79373 Structural Design Optimization of Reinforced Thin-Walled Vessels under External Pressure Using Simulation and Machine Learning Classification Algorithm
Authors: Lydia Novozhilova, Vladimir Urazhdin
Abstract:
An optimization problem for reinforced thin-walled vessels under uniform external pressure is considered. The conventional approaches to optimization generally start with pre-defined geometric parameters of the vessels, and then employ analytic or numeric calculations and/or experimental testing to verify functionality, such as stability under the projected conditions. The proposed approach consists of two steps. First, the feasibility domain will be identified in the multidimensional parameter space. Every point in the feasibility domain defines a design satisfying both geometric and functional constraints. Second, an objective function defined in this domain is formulated and optimized. The broader applicability of the suggested methodology is maximized by implementing the Support Vector Machines (SVM) classification algorithm of machine learning for identification of the feasible design region. Training data for SVM classifier is obtained using the Simulation package of SOLIDWORKS®. Based on the data, the SVM algorithm produces a curvilinear boundary separating admissible and not admissible sets of design parameters with maximal margins. Then optimization of the vessel parameters in the feasibility domain is performed using the standard algorithms for the constrained optimization. As an example, optimization of a ring-stiffened closed cylindrical thin-walled vessel with semi-spherical caps under high external pressure is implemented. As a functional constraint, von Mises stress criterion is used but any other stability constraint admitting mathematical formulation can be incorporated into the proposed approach. Suggested methodology has a good potential for reducing design time for finding optimal parameters of thin-walled vessels under uniform external pressure.Keywords: design parameters, feasibility domain, von Mises stress criterion, Support Vector Machine (SVM) classifier
Procedia PDF Downloads 327372 River Network Delineation from Sentinel 1 Synthetic Aperture Radar Data
Authors: Christopher B. Obida, George A. Blackburn, James D. Whyatt, Kirk T. Semple
Abstract:
In many regions of the world, especially in developing countries, river network data are outdated or completely absent, yet such information is critical for supporting important functions such as flood mitigation efforts, land use and transportation planning, and the management of water resources. In this study, a method was developed for delineating river networks using Sentinel 1 imagery. Unsupervised classification was applied to multi-temporal Sentinel 1 data to discriminate water bodies from other land covers then the outputs were combined to generate a single persistent water bodies product. A thinning algorithm was then used to delineate river centre lines, which were converted into vector features and built into a topologically structured geometric network. The complex river system of the Niger Delta was used to compare the performance of the Sentinel-based method against alternative freely available water body products from United States Geological Survey, European Space Agency and OpenStreetMap and a river network derived from a Shuttle Rader Topography Mission Digital Elevation Model. From both raster-based and vector-based accuracy assessments, it was found that the Sentinel-based river network products were superior to the comparator data sets by a substantial margin. The geometric river network that was constructed permitted a flow routing analysis which is important for a variety of environmental management and planning applications. The extracted network will potentially be applied for modelling dispersion of hydrocarbon pollutants in Ogoniland, a part of the Niger Delta. The approach developed in this study holds considerable potential for generating up to date, detailed river network data for the many countries where such data are deficient.Keywords: Sentinel 1, image processing, river delineation, large scale mapping, data comparison, geometric network
Procedia PDF Downloads 139371 E4D-MP: Time-Lapse Multiphysics Simulation and Joint Inversion Toolset for Large-Scale Subsurface Imaging
Authors: Zhuanfang Fred Zhang, Tim C. Johnson, Yilin Fang, Chris E. Strickland
Abstract:
A variety of geophysical techniques are available to image the opaque subsurface with little or no contact with the soil. It is common to conduct time-lapse surveys of different types for a given site for improved results of subsurface imaging. Regardless of the chosen survey methods, it is often a challenge to process the massive amount of survey data. The currently available software applications are generally based on the one-dimensional assumption for a desktop personal computer. Hence, they are usually incapable of imaging the three-dimensional (3D) processes/variables in the subsurface of reasonable spatial scales; the maximum amount of data that can be inverted simultaneously is often very small due to the capability limitation of personal computers. Presently, high-performance or integrating software that enables real-time integration of multi-process geophysical methods is needed. E4D-MP enables the integration and inversion of time-lapsed large-scale data surveys from geophysical methods. Using the supercomputing capability and parallel computation algorithm, E4D-MP is capable of processing data across vast spatiotemporal scales and in near real time. The main code and the modules of E4D-MP for inverting individual or combined data sets of time-lapse 3D electrical resistivity, spectral induced polarization, and gravity surveys have been developed and demonstrated for sub-surface imaging. E4D-MP provides capability of imaging the processes (e.g., liquid or gas flow, solute transport, cavity development) and subsurface properties (e.g., rock/soil density, conductivity) critical for successful control of environmental engineering related efforts such as environmental remediation, carbon sequestration, geothermal exploration, and mine land reclamation, among others.Keywords: gravity survey, high-performance computing, sub-surface monitoring, electrical resistivity tomography
Procedia PDF Downloads 157370 Kernel-Based Double Nearest Proportion Feature Extraction for Hyperspectral Image Classification
Authors: Hung-Sheng Lin, Cheng-Hsuan Li
Abstract:
Over the past few years, kernel-based algorithms have been widely used to extend some linear feature extraction methods such as principal component analysis (PCA), linear discriminate analysis (LDA), and nonparametric weighted feature extraction (NWFE) to their nonlinear versions, kernel principal component analysis (KPCA), generalized discriminate analysis (GDA), and kernel nonparametric weighted feature extraction (KNWFE), respectively. These nonlinear feature extraction methods can detect nonlinear directions with the largest nonlinear variance or the largest class separability based on the given kernel function. Moreover, they have been applied to improve the target detection or the image classification of hyperspectral images. The double nearest proportion feature extraction (DNP) can effectively reduce the overlap effect and have good performance in hyperspectral image classification. The DNP structure is an extension of the k-nearest neighbor technique. For each sample, there are two corresponding nearest proportions of samples, the self-class nearest proportion and the other-class nearest proportion. The term “nearest proportion” used here consider both the local information and other more global information. With these settings, the effect of the overlap between the sample distributions can be reduced. Usually, the maximum likelihood estimator and the related unbiased estimator are not ideal estimators in high dimensional inference problems, particularly in small data-size situation. Hence, an improved estimator by shrinkage estimation (regularization) is proposed. Based on the DNP structure, LDA is included as a special case. In this paper, the kernel method is applied to extend DNP to kernel-based DNP (KDNP). In addition to the advantages of DNP, KDNP surpasses DNP in the experimental results. According to the experiments on the real hyperspectral image data sets, the classification performance of KDNP is better than that of PCA, LDA, NWFE, and their kernel versions, KPCA, GDA, and KNWFE.Keywords: feature extraction, kernel method, double nearest proportion feature extraction, kernel double nearest feature extraction
Procedia PDF Downloads 344369 System Identification of Building Structures with Continuous Modeling
Authors: Ruichong Zhang, Fadi Sawaged, Lotfi Gargab
Abstract:
This paper introduces a wave-based approach for system identification of high-rise building structures with a pair of seismic recordings, which can be used to evaluate structural integrity and detect damage in post-earthquake structural condition assessment. The fundamental of the approach is based on wave features of generalized impulse and frequency response functions (GIRF and GFRF), i.e., wave responses at one structural location to an impulsive motion at another reference location in time and frequency domains respectively. With a pair of seismic recordings at the two locations, GFRF is obtainable as Fourier spectral ratio of the two recordings, and GIRF is then found with the inverse Fourier transformation of GFRF. With an appropriate continuous model for the structure, a closed-form solution of GFRF, and subsequent GIRF, can also be found in terms of wave transmission and reflection coefficients, which are related to structural physical properties above the impulse location. Matching the two sets of GFRF and/or GIRF from recordings and the model helps identify structural parameters such as wave velocity or shear modulus. For illustration, this study examines ten-story Millikan Library in Pasadena, California with recordings of Yorba Linda earthquake of September 3, 2002. The building is modelled as piecewise continuous layers, with which GFRF is derived as function of such building parameters as impedance, cross-sectional area, and damping. GIRF can then be found in closed form for some special cases and numerically in general. Not only does this study reveal the influential factors of building parameters in wave features of GIRF and GRFR, it also shows some system-identification results, which are consistent with other vibration- and wave-based results. Finally, this paper discusses the effectiveness of the proposed model in system identification.Keywords: wave-based approach, seismic responses of buildings, wave propagation in structures, construction
Procedia PDF Downloads 233368 Development and Validation of a Carbon Dioxide TDLAS Sensor for Studies on Fermented Dairy Products
Authors: Lorenzo Cocola, Massimo Fedel, Dragiša Savić, Bojana Danilović, Luca Poletto
Abstract:
An instrument for the detection and evaluation of gaseous carbon dioxide in the headspace of closed containers has been developed in the context of Packsensor Italian-Serbian joint project. The device is based on Tunable Diode Laser Absorption Spectroscopy (TDLAS) with a Wavelength Modulation Spectroscopy (WMS) technique in order to accomplish a non-invasive measurement inside closed containers of fermented dairy products (yogurts and fermented cheese in cups and bottles). The purpose of this instrument is the continuous monitoring of carbon dioxide concentration during incubation and storage of products over a time span of the whole shelf life of the product, in the presence of different microorganisms. The instrument’s optical front end has been designed to be integrated in a thermally stabilized incubator. An embedded computer provides processing of spectral artifacts and storage of an arbitrary set of calibration data allowing a properly calibrated measurement on many samples (cups and bottles) of different shapes and sizes commonly found in the retail distribution. A calibration protocol has been developed in order to be able to calibrate the instrument on the field also on containers which are notoriously difficult to seal properly. This calibration protocol is described and evaluated against reference measurements obtained through an industry standard (sampling) carbon dioxide metering technique. Some sets of validation test measurements on different containers are reported. Two test recordings of carbon dioxide concentration evolution are shown as an example of instrument operation. The first demonstrates the ability to monitor a rapid yeast growth in a contaminated sample through the increase of headspace carbon dioxide. Another experiment shows the dissolution transient with a non-saturated liquid medium in presence of a carbon dioxide rich headspace atmosphere.Keywords: TDLAS, carbon dioxide, cups, headspace, measurement
Procedia PDF Downloads 324367 Language Politics and Identity in Translation: From a Monolingual Text to Multilingual Text in Chinese Translations
Authors: Chu-Ching Hsu
Abstract:
This paper focuses on how the government-led language policies and the political changes in Taiwan manipulate the languages choice in translations and what translation strategies are employed by the translator to show his or her language ideology behind the power struggles and decision-making. Therefore, framed by Lefevere’s theoretical concept of translating as rewriting, and carried out a diachronic and chronological study, this paper specifically sets out to investigate the language ideology and translator’s idiolect of Chinese language translations of Anglo-American novels. The examples drawn to explore these issues were taken from different versions of Chinese renditions of Mark Twain’s English-language novel The Adventures of Huckleberry Finn in which there are several different dialogues originally written in the colloquial language and dialect used in the American state of Mississippi and reproduced in Mark Twain’s works. Also, adapted corpus methodology, many examples are extracted as instances from the translated texts and source text, to illuminate how the translators in Taiwan deal with the dialectal features encoded in Twain’s works, and how different versions of Chinese translations are employed by Taiwanese translators to confirm the language polices and to express their language identity textually in different periods of the past five decades, from the 1960s onward. The finding of this study suggests that the use of Taiwanese dialect and language patterns in translations does relate to the movement of the mother-tongue language and language ideology of the translator as well as to the issue of language identity raised in the island of Taiwan. Furthermore, this study confirms that the change of political power in Taiwan does bring significantly impact in language policy-- assimilationism, pluralism or multiculturalism, which also makes Taiwan from a monolingual to multilingual society, where the language ideology and identity can be revealed not only in people’s daily communication but also in written translations.Keywords: language politics and policies, literary translation, mother-tongue, multiculturalism, translator’s ideology
Procedia PDF Downloads 395366 Spatial Data Science for Data Driven Urban Planning: The Youth Economic Discomfort Index for Rome
Authors: Iacopo Testi, Diego Pajarito, Nicoletta Roberto, Carmen Greco
Abstract:
Today, a consistent segment of the world’s population lives in urban areas, and this proportion will vastly increase in the next decades. Therefore, understanding the key trends in urbanization, likely to unfold over the coming years, is crucial to the implementation of sustainable urban strategies. In parallel, the daily amount of digital data produced will be expanding at an exponential rate during the following years. The analysis of various types of data sets and its derived applications have incredible potential across different crucial sectors such as healthcare, housing, transportation, energy, and education. Nevertheless, in city development, architects and urban planners appear to rely mostly on traditional and analogical techniques of data collection. This paper investigates the prospective of the data science field, appearing to be a formidable resource to assist city managers in identifying strategies to enhance the social, economic, and environmental sustainability of our urban areas. The collection of different new layers of information would definitely enhance planners' capabilities to comprehend more in-depth urban phenomena such as gentrification, land use definition, mobility, or critical infrastructural issues. Specifically, the research results correlate economic, commercial, demographic, and housing data with the purpose of defining the youth economic discomfort index. The statistical composite index provides insights regarding the economic disadvantage of citizens aged between 18 years and 29 years, and results clearly display that central urban zones and more disadvantaged than peripheral ones. The experimental set up selected the city of Rome as the testing ground of the whole investigation. The methodology aims at applying statistical and spatial analysis to construct a composite index supporting informed data-driven decisions for urban planning.Keywords: data science, spatial analysis, composite index, Rome, urban planning, youth economic discomfort index
Procedia PDF Downloads 135365 Comparison of Multivariate Adaptive Regression Splines and Random Forest Regression in Predicting Forced Expiratory Volume in One Second
Authors: P. V. Pramila , V. Mahesh
Abstract:
Pulmonary Function Tests are important non-invasive diagnostic tests to assess respiratory impairments and provides quantifiable measures of lung function. Spirometry is the most frequently used measure of lung function and plays an essential role in the diagnosis and management of pulmonary diseases. However, the test requires considerable patient effort and cooperation, markedly related to the age of patients esulting in incomplete data sets. This paper presents, a nonlinear model built using Multivariate adaptive regression splines and Random forest regression model to predict the missing spirometric features. Random forest based feature selection is used to enhance both the generalization capability and the model interpretability. In the present study, flow-volume data are recorded for N= 198 subjects. The ranked order of feature importance index calculated by the random forests model shows that the spirometric features FVC, FEF 25, PEF,FEF 25-75, FEF50, and the demographic parameter height are the important descriptors. A comparison of performance assessment of both models prove that, the prediction ability of MARS with the `top two ranked features namely the FVC and FEF 25 is higher, yielding a model fit of R2= 0.96 and R2= 0.99 for normal and abnormal subjects. The Root Mean Square Error analysis of the RF model and the MARS model also shows that the latter is capable of predicting the missing values of FEV1 with a notably lower error value of 0.0191 (normal subjects) and 0.0106 (abnormal subjects). It is concluded that combining feature selection with a prediction model provides a minimum subset of predominant features to train the model, yielding better prediction performance. This analysis can assist clinicians with a intelligence support system in the medical diagnosis and improvement of clinical care.Keywords: FEV, multivariate adaptive regression splines pulmonary function test, random forest
Procedia PDF Downloads 310364 A Support Vector Machine Learning Prediction Model of Evapotranspiration Using Real-Time Sensor Node Data
Authors: Waqas Ahmed Khan Afridi, Subhas Chandra Mukhopadhyay, Bandita Mainali
Abstract:
The research paper presents a unique approach to evapotranspiration (ET) prediction using a Support Vector Machine (SVM) learning algorithm. The study leverages real-time sensor node data to develop an accurate and adaptable prediction model, addressing the inherent challenges of traditional ET estimation methods. The integration of the SVM algorithm with real-time sensor node data offers great potential to improve spatial and temporal resolution in ET predictions. In the model development, key input features are measured and computed using mathematical equations such as Penman-Monteith (FAO56) and soil water balance (SWB), which include soil-environmental parameters such as; solar radiation (Rs), air temperature (T), atmospheric pressure (P), relative humidity (RH), wind speed (u2), rain (R), deep percolation (DP), soil temperature (ST), and change in soil moisture (∆SM). The one-year field data are split into combinations of three proportions i.e. train, test, and validation sets. While kernel functions with tuning hyperparameters have been used to train and improve the accuracy of the prediction model with multiple iterations. This paper also outlines the existing methods and the machine learning techniques to determine Evapotranspiration, data collection and preprocessing, model construction, and evaluation metrics, highlighting the significance of SVM in advancing the field of ET prediction. The results demonstrate the robustness and high predictability of the developed model on the basis of performance evaluation metrics (R2, RMSE, MAE). The effectiveness of the proposed model in capturing complex relationships within soil and environmental parameters provide insights into its potential applications for water resource management and hydrological ecosystem.Keywords: evapotranspiration, FAO56, KNIME, machine learning, RStudio, SVM, sensors
Procedia PDF Downloads 69363 Application of Hydrological Engineering Centre – River Analysis System (HEC-RAS) to Estuarine Hydraulics
Authors: Julia Zimmerman, Gaurav Savant
Abstract:
This study aims to evaluate the efficacy of the U.S. Army Corp of Engineers’ River Analysis System (HEC-RAS) application to modeling the hydraulics of estuaries. HEC-RAS has been broadly used for a variety of riverine applications. However, it has not been widely applied to the study of circulation in estuaries. This report details the model development and validation of a combined 1D/2D unsteady flow hydraulic model using HEC-RAS for estuaries and they are associated with tidally influenced rivers. Two estuaries, Galveston Bay and Delaware Bay, were used as case studies. Galveston Bay, a bar-built, vertically mixed estuary, was modeled for the 2005 calendar year. Delaware Bay, a drowned river valley estuary, was modeled from October 22, 2019, to November 5, 2019. Water surface elevation was used to validate both models by comparing simulation results to NOAA’s Center for Operational Oceanographic Products and Services (CO-OPS) gauge data. Simulations were run using the Diffusion Wave Equations (DW), the Shallow Water Equations, Eulerian-Lagrangian Method (SWE-ELM), and the Shallow Water Equations Eulerian Method (SWE-EM) and compared for both accuracy and computational resources required. In general, the Diffusion Wave Equations results were found to be comparable to the two Shallow Water equations sets while requiring less computational power. The 1D/2D combined approach was valid for study areas within the 2D flow area, with the 1D flow serving mainly as an inflow boundary condition. Within the Delaware Bay estuary, the HEC-RAS DW model ran in 22 minutes and had an average R² value of 0.94 within the 2-D mesh. The Galveston Bay HEC-RAS DW ran in 6 hours and 47 minutes and had an average R² value of 0.83 within the 2-D mesh. The longer run time and lower R² for Galveston Bay can be attributed to the increased length of the time frame modeled and the greater complexity of the estuarine system. The models did not accurately capture tidal effects within the 1D flow area.Keywords: Delaware bay, estuarine hydraulics, Galveston bay, HEC-RAS, one-dimensional modeling, two-dimensional modeling
Procedia PDF Downloads 199