Search results for: gp-closed sets
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1227

Search results for: gp-closed sets

417 Comparison of Different Artificial Intelligence-Based Protein Secondary Structure Prediction Methods

Authors: Jamerson Felipe Pereira Lima, Jeane Cecília Bezerra de Melo

Abstract:

The difficulty and cost related to obtaining of protein tertiary structure information through experimental methods, such as X-ray crystallography or NMR spectroscopy, helped raising the development of computational methods to do so. An approach used in these last is prediction of tridimensional structure based in the residue chain, however, this has been proved an NP-hard problem, due to the complexity of this process, explained by the Levinthal paradox. An alternative solution is the prediction of intermediary structures, such as the secondary structure of the protein. Artificial Intelligence methods, such as Bayesian statistics, artificial neural networks (ANN), support vector machines (SVM), among others, were used to predict protein secondary structure. Due to its good results, artificial neural networks have been used as a standard method to predict protein secondary structure. Recent published methods that use this technique, in general, achieved a Q3 accuracy between 75% and 83%, whereas the theoretical accuracy limit for protein prediction is 88%. Alternatively, to achieve better results, support vector machines prediction methods have been developed. The statistical evaluation of methods that use different AI techniques, such as ANNs and SVMs, for example, is not a trivial problem, since different training sets, validation techniques, as well as other variables can influence the behavior of a prediction method. In this study, we propose a prediction method based on artificial neural networks, which is then compared with a selected SVM method. The chosen SVM protein secondary structure prediction method is the one proposed by Huang in his work Extracting Physico chemical Features to Predict Protein Secondary Structure (2013). The developed ANN method has the same training and testing process that was used by Huang to validate his method, which comprises the use of the CB513 protein data set and three-fold cross-validation, so that the comparative analysis of the results can be made comparing directly the statistical results of each method.

Keywords: artificial neural networks, protein secondary structure, protein structure prediction, support vector machines

Procedia PDF Downloads 594
416 INCIPIT-CRIS: A Research Information System Combining Linked Data Ontologies and Persistent Identifiers

Authors: David Nogueiras Blanco, Amir Alwash, Arnaud Gaudinat, René Schneider

Abstract:

At a time when the access to and the sharing of information are crucial in the world of research, the use of technologies such as persistent identifiers (PIDs), Current Research Information Systems (CRIS), and ontologies may create platforms for information sharing if they respond to the need of disambiguation of their data by assuring interoperability inside and between other systems. INCIPIT-CRIS is a continuation of the former INCIPIT project, whose goal was to set up an infrastructure for a low-cost attribution of PIDs with high granularity based on Archival Resource Keys (ARKs). INCIPIT-CRIS can be interpreted as a logical consequence and propose a research information management system developed from scratch. The system has been created on and around the Schema.org ontology with a further articulation of the use of ARKs. It is thus built upon the infrastructure previously implemented (i.e., INCIPIT) in order to enhance the persistence of URIs. As a consequence, INCIPIT-CRIS aims to be the hinge between previously separated aspects such as CRIS, ontologies and PIDs in order to produce a powerful system allowing the resolution of disambiguation problems using a combination of an ontology such as Schema.org and unique persistent identifiers such as ARK, allowing the sharing of information through a dedicated platform, but also the interoperability of the system by representing the entirety of the data as RDF triplets. This paper aims to present the implemented solution as well as its simulation in real life. We will describe the underlying ideas and inspirations while going through the logic and the different functionalities implemented and their links with ARKs and Schema.org. Finally, we will discuss the tests performed with our project partner, the Swiss Institute of Bioinformatics (SIB), by the use of large and real-world data sets.

Keywords: current research information systems, linked data, ontologies, persistent identifier, schema.org, semantic web

Procedia PDF Downloads 110
415 Artificial Neural Network Based Approach in Prediction of Potential Water Pollution Across Different Land-Use Patterns

Authors: M.Rüştü Karaman, İsmail İşeri, Kadir Saltalı, A.Reşit Brohi, Ayhan Horuz, Mümin Dizman

Abstract:

Considerable relations has recently been given to the environmental hazardous caused by agricultural chemicals such as excess fertilizers. In this study, a neural network approach was investigated in the prediction of potential nitrate pollution across different land-use patterns by using a feedforward multilayered computer model of artificial neural network (ANN) with proper training. Periodical concentrations of some anions, especially nitrate (NO3-), and cations were also detected in drainage waters collected from the drain pipes placed in irrigated tomato field, unirrigated wheat field, fallow and pasture lands. The soil samples were collected from the irrigated tomato field and unirrigated wheat field on a grid system with 20 m x 20 m intervals. Site specific nitrate concentrations in the soil samples were measured for ANN based simulation of nitrate leaching potential from the land profiles. In the application of ANN model, a multi layered feedforward was evaluated, and data sets regarding with training, validation and testing containing the measured soil nitrate values were estimated based on spatial variability. As a result of the testing values, while the optimal structures of 2-15-1 was obtained (R2= 0.96, P < 0.01) for unirrigated field, the optimal structures of 2-10-1 was obtained (R2= 0.96, P < 0.01) for irrigated field. The results showed that the ANN model could be successfully used in prediction of the potential leaching levels of nitrate, based on different land use patterns. However, for the most suitable results, the model should be calibrated by training according to different NN structures depending on site specific soil parameters and varied agricultural managements.

Keywords: artificial intelligence, ANN, drainage water, nitrate pollution

Procedia PDF Downloads 287
414 Instructional Leadership, Information and Communications Technology Competencies and Performance of Basic Education Teachers

Authors: Jay Martin L. Dionaldo

Abstract:

This study aimed to develop a causal model on the performance of the basic education teachers in the Division of Malaybalay City for the school year 2018-2019. This study used the responses of 300 randomly selected basic education teachers of Malaybalay City, Bukidnon. They responded to the three sets of questionnaires patterned from the National Education Association (2018) on instructional leadership of teachers, the questionnaire of Caluza et al., (2017) for information and communications technology competencies and the questionnaire on the teachers’ performance using the Individual Performance Commitment and Review Form (IPCRF) adopted by the Department of Education (DepEd). Descriptive statistics such as mean for the description, correlation for a relationship, regression for the extent influence, and path analysis for the model that best fits teachers’ performance were used. Result showed that basic education teachers have a very satisfactory level of performance. Also, the teachers highly practice instructional leadership practices in terms of coaching and mentoring, facilitating collaborative relationships, and community awareness and engagement. On the other hand, they are proficient users of ICT in terms of technology operations and concepts and basic users in terms of their pedagogical indicators. Furthermore, instructional leadership, coaching and mentoring, facilitating collaborative relationships and community awareness and engagement and information and communications technology competencies; technology operations and concept and pedagogy were significantly correlated toward teachers’ performance. Coaching and mentoring, community awareness and engagement, and technology operations and concept were the best predictors of teachers’ performance. The model that best fit teachers’ performance is anchored on coaching and mentoring of the teachers, embedded with facilitating collaborative relationships, community awareness, and engagement, technology operations, and concepts, and pedagogy.

Keywords: information and communications technology, instructional leadership, coaching and mentoring, collaborative relationship

Procedia PDF Downloads 99
413 Development of a Multi-Locus DNA Metabarcoding Method for Endangered Animal Species Identification

Authors: Meimei Shi

Abstract:

Objectives: The identification of endangered species, especially simultaneous detection of multiple species in complex samples, plays a critical role in alleged wildlife crime incidents and prevents illegal trade. This study was to develop a multi-locus DNA metabarcoding method for endangered animal species identification. Methods: Several pairs of universal primers were designed according to the mitochondria conserved gene regions. Experimental mixtures were artificially prepared by mixing well-defined species, including endangered species, e.g., forest musk, bear, tiger, pangolin, and sika deer. The artificial samples were prepared with 1-16 well-characterized species at 1% to 100% DNA concentrations. After multiplex-PCR amplification and parameter modification, the amplified products were analyzed by capillary electrophoresis and used for NGS library preparation. The DNA metabarcoding was carried out based on Illumina MiSeq amplicon sequencing. The data was processed with quality trimming, reads filtering, and OTU clustering; representative sequences were blasted using BLASTn. Results: According to the parameter modification and multiplex-PCR amplification results, five primer sets targeting COI, Cytb, 12S, and 16S, respectively, were selected as the NGS library amplification primer panel. High-throughput sequencing data analysis showed that the established multi-locus DNA metabarcoding method was sensitive and could accurately identify all species in artificial mixtures, including endangered animal species Moschus berezovskii, Ursus thibetanus, Panthera tigris, Manis pentadactyla, Cervus nippon at 1% (DNA concentration). In conclusion, the established species identification method provides technical support for customs and forensic scientists to prevent the illegal trade of endangered animals and their products.

Keywords: DNA metabarcoding, endangered animal species, mitochondria nucleic acid, multi-locus

Procedia PDF Downloads 114
412 Assessing Denitrification-Disintegration Model’s Efficacy in Simulating Greenhouse Gas Emissions, Crop Growth, Yield, and Soil Biochemical Processes in Moroccan Context

Authors: Mohamed Boullouz, Mohamed Louay Metougui

Abstract:

Accurate modeling of greenhouse gas (GHG) emissions, crop growth, soil productivity, and biochemical processes is crucial considering escalating global concerns about climate change and the urgent need to improve agricultural sustainability. The application of the denitrification-disintegration (DNDC) model in the context of Morocco's unique agro-climate is thoroughly investigated in this study. Our main research hypothesis is that the DNDC model offers an effective and powerful tool for precisely simulating a wide range of significant parameters, including greenhouse gas emissions, crop growth, yield potential, and complex soil biogeochemical processes, all consistent with the intricate features of environmental Moroccan agriculture. In order to verify these hypotheses, a vast amount of field data covering Morocco's various agricultural regions and encompassing a range of soil types, climatic factors, and crop varieties had to be gathered. These experimental data sets will serve as the foundation for careful model calibration and subsequent validation, ensuring the accuracy of simulation results. In conclusion, the prospective research findings add to the global conversation on climate-resilient agricultural practices while encouraging the promotion of sustainable agricultural models in Morocco. A policy architect's and an agricultural actor's ability to make informed decisions that not only advance food security but also environmental stability may be strengthened by the impending recognition of the DNDC model as a potent simulation tool tailored to Moroccan conditions.

Keywords: greenhouse gas emissions, DNDC model, sustainable agriculture, Moroccan cropping systems

Procedia PDF Downloads 47
411 Identification of the Microalgae Species in a Wild Mix Culture Acclimated to Landfill Leachate and Ammonia Removal Performances in a Microbubble Assisted Photobioreactor

Authors: Neslihan Ozman Say, Jim Gilmour, Pratik Desai, William Zimmerman

Abstract:

Landfill leachate treatment has been attracting researchers recently for various environmental and economical reasons. Leachate discharge to receiving waterbodies without treatment causes serious detrimental effects including partial oxygen depletion due to high biological oxygen demand (BOD) and chemical oxygen demand (COD) concentrations besides toxicity of heavy metals it contains and high ammonia concentrations. In this study, it is aimed to show microalgal ammonia removal performances of a wild microalgae consortia as an alternative treatment method and determine the dominant leachate tolerant species for this consortia. For the microalgae species identification experiments a microalgal consortium which has been isolated from a local pond in Sheffield inoculated in %5 diluted raw landfill leachate and acclimated to the leachate by batch feeding for a month. In order to determine the most tolerant microalgal consortium, four different untreated landfill leachate samples have been used as diluted in four different ratios as 5%, 10%, 20%, and 40%. Microalgae cell samples have been collected from all experiment sets and have been examined by using 18S rDNA sequencing and specialised gel electrophoresis which are adapted molecular biodiversity methods. The best leachate tolerant algal consortium is being used in order to determine ammonia removal performances of the culture in a microbubble assisted photobioreactor (PBR). A porous microbubble diffuser which is supported by a fluidic oscillator is being used for dosing CO₂ and air mixture in the PBR. It is known that high mass transfer performance of microbubble technology provides a better removal efficiency and a better mixing in the photobioreactor. Ammonia concentrations and microalgal growth are being monitored for PBR currently. It is aimed to present all the results of the study in final paper submission.

Keywords: ammonia removal from leachate, landfill leachate treatment, microalgae species identification, microbubble assisted photobioreactors

Procedia PDF Downloads 139
410 A Study of Learning Achievement for Heat Transfer by Using Experimental Sets of Convection with the Predict-Observe-Explain Teaching Technique

Authors: Wanlapa Boonsod, Nisachon Yangprasong, Udomsak Kitthawee

Abstract:

Thermal physics education is a complicated and challenging topic to discuss in any classroom. As a result, most students tend to be uninterested in learning this topic. In the current study, a convection experiment set was devised to show how heat can be transferred by a convection system to a thermoelectric plate until a LED flashes. This research aimed to 1) create a natural convection experimental set, 2) study learning achievement on the convection experimental set with the predict-observe-explain (POE) technique, and 3) study satisfaction for the convection experimental set with the predict-observe-explain (POE) technique. The samples were chosen by purposive sampling and comprised 28 students in grade 11 at Patumkongka School in Bangkok, Thailand. The primary research instrument was the plan for predict-observe-explain (POE) technique on heat transfer using a convection experimental set. Heat transfer experimental set by convection. The instruments used to collect data included a heat transfer achievement model by convection, a Satisfaction Questionnaire after the learning activity, and the predict-observe-explain (POE) technique for heat transfer using a convection experimental set. The research format comprised a one-group pretest-posttest design. The data was analyzed by GeoGebra program. The statistics used in the research were mean, standard deviation and t-test for dependent samples. The results of the research showed that achievement on heat transfer using convection experimental set was composed of thermo-electrics on the top side attached to the heat sink and another side attached to a stainless plate. Electrical current was displayed by the flashing of a 5v LED. The entire set of thermo-electrics was set up on the top of the box and heated by an alcohol burner. The achievement of learning was measured with the predict-observe-explain (POE) technique, with the natural convection experimental set statistically higher than before learning at a 0.01 level. Satisfaction with POE for physics learning of heat transfer by using convection experimental set was at a high level (4.83 from 5.00).

Keywords: convection, heat transfer, physics education, POE

Procedia PDF Downloads 194
409 In Silico Exploration of Quinazoline Derivatives as EGFR Inhibitors for Lung Cancer: A Multi-Modal Approach Integrating QSAR-3D, ADMET, Molecular Docking, and Molecular Dynamics Analyses

Authors: Mohamed Moussaoui

Abstract:

A series of thirty-one potential inhibitors targeting the epidermal growth factor receptor kinase (EGFR), derived from quinazoline, underwent 3D-QSAR analysis using CoMFA and CoMSIA methodologies. The training and test sets of quinazoline derivatives were utilized to construct and validate the QSAR models, respectively, with dataset alignment performed using the lowest energy conformer of the most active compound. The best-performing CoMFA and CoMSIA models demonstrated impressive determination coefficients, with R² values of 0.981 and 0.978, respectively, and Leave One Out cross-validation determination coefficients, Q², of 0.645 and 0.729, respectively. Furthermore, external validation using a test set of five compounds yielded predicted determination coefficients, R² test, of 0.929 and 0.909 for CoMFA and CoMSIA, respectively. Building upon these promising results, eighteen new compounds were designed and assessed for drug likeness and ADMET properties through in silico methods. Additionally, molecular docking studies were conducted to elucidate the binding interactions between the selected compounds and the enzyme. Detailed molecular dynamics simulations were performed to analyze the stability, conformational changes, and binding interactions of the quinazoline derivatives with the EGFR kinase. These simulations provided deeper insights into the dynamic behavior of the compounds within the active site. This comprehensive analysis enhances the understanding of quinazoline derivatives as potential anti-cancer agents and provides valuable insights for lead optimization in the early stages of drug discovery, particularly for developing highly potent anticancer therapeutics

Keywords: 3D-QSAR, CoMFA, CoMSIA, ADMET, molecular docking, quinazoline, molecular dynamic, egfr inhibitors, lung cancer, anticancer

Procedia PDF Downloads 22
408 Electrochemical Biosensor for the Detection of Botrytis spp. in Temperate Legume Crops

Authors: Marzia Bilkiss, Muhammad J. A. Shiddiky, Mostafa K. Masud, Prabhakaran Sambasivam, Ido Bar, Jeremy Brownlie, Rebecca Ford

Abstract:

A greater achievement in the Integrated Disease Management (IDM) to prevent the loss would result from early diagnosis and quantitation of the causal pathogen species for accurate and timely disease control. This could significantly reduce costs to the growers and reduce any flow on impacts to the environment from excessive chemical spraying. Necrotrophic fungal disease botrytis grey mould, caused by Botrytis cinerea and Botrytis fabae, significantly reduce temperate legume yield and grain quality during favourable environmental condition in Australia and worldwide. Several immunogenic and molecular probe-type protocols have been developed for their diagnosis, but these have varying levels of species-specificity, sensitivity, and consequent usefulness within the paddock. To substantially improve speed, accuracy, and sensitivity, advanced nanoparticle-based biosensor approaches have been developed. For this, two sets of primers were designed for both Botrytis cinerea and Botrytis fabae which have shown the species specificity with initial sensitivity of two genomic copies/µl in pure fungal backgrounds using multiplexed quantitative PCR. During further validation, quantitative PCR detected 100 spores on artificially infected legume leaves. Simultaneously an electro-catalytic assay was developed for both target fungal DNA using functionalised magnetic nanoparticles. This was extremely sensitive, able to detect a single spore within a raw total plant nucleic acid extract background. We believe that the translation of this technology to the field will enable quantitative assessment of pathogen load for future accurate decision support of informed botrytis grey mould management.

Keywords: biosensor, botrytis grey mould, sensitive, species specific

Procedia PDF Downloads 153
407 Integration of Big Data to Predict Transportation for Smart Cities

Authors: Sun-Young Jang, Sung-Ah Kim, Dongyoun Shin

Abstract:

The Intelligent transportation system is essential to build smarter cities. Machine learning based transportation prediction could be highly promising approach by delivering invisible aspect visible. In this context, this research aims to make a prototype model that predicts transportation network by using big data and machine learning technology. In detail, among urban transportation systems this research chooses bus system.  The research problem that existing headway model cannot response dynamic transportation conditions. Thus, bus delay problem is often occurred. To overcome this problem, a prediction model is presented to fine patterns of bus delay by using a machine learning implementing the following data sets; traffics, weathers, and bus statues. This research presents a flexible headway model to predict bus delay and analyze the result. The prototyping model is composed by real-time data of buses. The data are gathered through public data portals and real time Application Program Interface (API) by the government. These data are fundamental resources to organize interval pattern models of bus operations as traffic environment factors (road speeds, station conditions, weathers, and bus information of operating in real-time). The prototyping model is designed by the machine learning tool (RapidMiner Studio) and conducted tests for bus delays prediction. This research presents experiments to increase prediction accuracy for bus headway by analyzing the urban big data. The big data analysis is important to predict the future and to find correlations by processing huge amount of data. Therefore, based on the analysis method, this research represents an effective use of the machine learning and urban big data to understand urban dynamics.

Keywords: big data, machine learning, smart city, social cost, transportation network

Procedia PDF Downloads 238
406 The Effects of L2 Storybook Reading and Interactive Vocabulary Instruction on Vocabulary Acquisition

Authors: Lenore Van Den Berg

Abstract:

Vocabulary development is positively associated with reading development, reading comprehension, and academic achievement. It is frequently stated that South Africa is in the midst of a literacy crisis. The past 24 years since the first democratically elected government have not revolutionised the education system; rather, after various curriculum changes and continued struggles to incorporate all 11 official languages as languages of instruction, research shows that 78 per cent of South African Grade 4 learners are functionally illiterate. The study sets out to find solutions to this problem and to add to the research base on vocabulary acquisition by assessing the effect of integrating the principles of explicit, interactive vocabulary instruction, within the context of storybook reading, on Grade 1 vocabulary acquisition. Participants comprised of 69 Grade 1 English second language learners from three classes in two government primary schools. The two schools differ in socio-economic status (SES), with School A having a lower SES than School B. One Grade 1 class was randomly assigned to be the Experimental Group, while two other classes served as control groups. The intervention took place for a period of 18 weeks and consisted of 30-minute storybook reading sessions, accompanied by interactive vocabulary instruction, twice a week. The Peabody Picture Vocabulary Test IV (PPVT-IV) was the diagnostic test administered to all learners before the intervention, as a pre-test, and after the interventions as a post-test. Data regarding excising vocabulary instruction practices and approaches were also collected through classroom observations and individual, semi-structured interviews with the Experimental Group’s teacher. Findings suggest that second language storybook reading, accompanied by explicit, interactive vocabulary instruction, have a positive impact on Grade 1 vocabulary acquisition but that vocabulary teaching practices and socio-economic status also play a key role in vocabulary acquisition.

Keywords: interactive vocabulary instruction, second language vocabulary, storybook reading, vocabulary acquisition, reading development, PPVT

Procedia PDF Downloads 72
405 Research Trends in Using Virtual Reality for the Analysis and Treatment of Lower-Limb Musculoskeletal Injury of Athletes: A Literature Review

Authors: Hannah K. M. Tang, Muhammad Ateeq, Mark J. Lake, Badr Abdullah, Frederic A. Bezombes

Abstract:

There is little research applying virtual reality (VR) to the treatment of musculoskeletal injury in athletes. This is despite their prevalence, and the implications for physical and psychological health. Nevertheless, developments of wireless VR headsets better facilitate dynamic movement in VR environments (VREs), and more research is expected in this emerging field. This systematic review identified publications that used VR interventions for the analysis or treatment of lower-limb musculoskeletal injury of athletes. It established a search protocol, and through narrative discussion, identified existing trends. Database searches encompassed four term sets: 1) VR systems; 2) musculoskeletal injuries; 3) sporting population; 4) movement outcome analysis. Overall, a total of 126 publications were identified through database searching, and twelve were included in the final analysis and discussion. Many of the studies were pilot and proof of concept work. Seven of the twelve publications were observational studies. However, this may provide preliminary data from which clinical trials will branch. If specified, the focus of the literature was very narrow, with very similar population demographics and injuries. The trends in the literature findings emphasised the role of VR and attentional focus, the strategic manipulation of movement outcomes, and the transfer of skill to the real-world. Causal inferences may have been undermined by flaws, as most studies were limited by the practicality of conducting a two-factor clinical-VR-based study. In conclusion, by assessing the exploratory studies, and combining this with the use of numerous developments, techniques, and tools, a novel application could be established to utilise VR with dynamic movement, for the effective treatment of specific musculoskeletal injuries of athletes.

Keywords: athletes, lower-limb musculoskeletal injury, rehabilitation, return-to-sport, virtual reality

Procedia PDF Downloads 212
404 Management and Leadership Development at Higher Educational Institutions: A Case Study of a South African Management Development Program

Authors: Michael Naidoo

Abstract:

The purpose and functions of higher education institutions in the 21st century are evolving because of rapid changes in the global landscape. To remain germane, higher education institutions are in a period of swift and radical change. The success of these changes is highly dependent on the effective leadership of the institution. Consequently, many higher education institutions have invested time and finances into the management and leadership development of their staff. The development has taken many different forms and focus areas, depending on the societal and institutional needs, as well as available financial resources and infrastructural practices. South Africa has many public and private higher education institutions which are also undergoing significant changes to meet the contextual needs of the country. Many of these institutions have provided management and leadership development programs for their staff. This research aims at exploring the common, critical content, structure and practices of effective management and leadership development programs at higher educational institutions. This research will also examine a specific management development program (MDP) at a South African private higher educational institution. Finally, the research will review how organizational leadership is utilized in management and leadership development programs. The research is underpinned by the paradigm of interpretivism. This is because the aims of the research will be achieved by the collection of qualitative data. The qualitative data will be gathered through individual semi-structured interviews with the facilitators of the MDP program and some of the MDP candidates. The validity of the findings will be increased by the triangulation of data from both sets of interviews. An embedded, single case study design will be used. All ethical protocols will be followed throughout the research. The findings of the research should reveal more information about the key elements that should be incorporated into management and leadership development programs. These include crucial content, structure and practices. The research should also reveal how organizational leadership can be successfully incorporated into the programs. This research can then be used by higher educational institutions to strengthen their management and leadership development programs.

Keywords: managment, leadership, development, organizational leadership

Procedia PDF Downloads 35
403 The Effects of Prolonged Use of Caffeine on Thyroid and Adrenal Glands – A Retrospective Cohort Study

Authors: Vasishtha Avadhani Upadrasta, Mradul Kumar Daga, Smita Kaushik

Abstract:

Background: Caffeine consumption has skyrocketed in the recent decades as we try to match pace with the machines. Studies have been conducted on animals and a few on humans, mainly on the acute effects of high-dose caffeine intake. Almost none have been conducted on the chronic effects of caffeine consumption. This study involved Medical professionals as case subjects, who consumed caffeine daily. Methods: This study, over a period of 3 months, involved 96 volunteers (chosen randomly w.r.t. gender and field in medical fraternity), including people who drank >500mg of caffeine a day to people who consumed none. People with any co-morbidities at all were excluded straight away. Two sets of blood samples were drawn and assessed. Three groups were created, Group 1 (>200mg caffeine/day) and Group 2 (15-200 mg caffeine/day) and Group 3 (<200mg Caffeine/day). Results: The result of the study found that exposure to caffeine at doses >200mg/day for more than 6 months leads to a significant difference in circulating free T3 [(-0.96 pmol/L ± 0.07) = (-18.5%), CI 95%, p = .000024] and Cortisol [(-123 nmol/L ± 9.8) = (-46.8%), CI 95%, p = .00029] hormones but shows an insignificant effect on circulating TSH [0.4 mIU/L, CI 95%, p=.37] and ACTH [(-3.2 pg/ml ± 0.3), CI 95%, p = .53) hormones, which stay within normal physiological ranges, irrespective of the daily dose of consumption. Results also highlight that women are more susceptible to decrement in fT3 than men (Relative Risk =1.58, ANOVA F-static = 7.15, p = 0.0105). Conclusions: Caffeine consumption in excess of 200mg/day, for more than or equal to 6 months, causes significant derangement in basal fT3 and Cortisol hormone levels, without affecting the TSH and ACTH (regulatory) hormone levels, indicating disturbance of action at the peripheral and/or cellular levels, possibly via the Paraventricular Nucleus –Leptin-CAR-Adenosine interactions. Women are more susceptible to decrement in fT3 levels than men (at same dose of caffeine).

Keywords: ACTH, adrenals, caffeine, cortisol, thyroid, thyroxin, TSH

Procedia PDF Downloads 49
402 Optimizing Energy Efficiency: Leveraging Big Data Analytics and AWS Services for Buildings and Industries

Authors: Gaurav Kumar Sinha

Abstract:

In an era marked by increasing concerns about energy sustainability, this research endeavors to address the pressing challenge of energy consumption in buildings and industries. This study delves into the transformative potential of AWS services in optimizing energy efficiency. The research is founded on the recognition that effective management of energy consumption is imperative for both environmental conservation and economic viability. Buildings and industries account for a substantial portion of global energy use, making it crucial to develop advanced techniques for analysis and reduction. This study sets out to explore the integration of AWS services with big data analytics to provide innovative solutions for energy consumption analysis. Leveraging AWS's cloud computing capabilities, scalable infrastructure, and data analytics tools, the research aims to develop efficient methods for collecting, processing, and analyzing energy data from diverse sources. The core focus is on creating predictive models and real-time monitoring systems that enable proactive energy management. By harnessing AWS's machine learning and data analytics capabilities, the research seeks to identify patterns, anomalies, and optimization opportunities within energy consumption data. Furthermore, this study aims to propose actionable recommendations for reducing energy consumption in buildings and industries. By combining AWS services with metrics-driven insights, the research strives to facilitate the implementation of energy-efficient practices, ultimately leading to reduced carbon emissions and cost savings. The integration of AWS services not only enhances the analytical capabilities but also offers scalable solutions that can be customized for different building and industrial contexts. The research also recognizes the potential for AWS-powered solutions to promote sustainable practices and support environmental stewardship.

Keywords: energy consumption analysis, big data analytics, AWS services, energy efficiency

Procedia PDF Downloads 45
401 Colloquialism in Audiovisual Translation: English Subtitling of the Lebanese Film Capernaum as a Case Study

Authors: Fatima Saab

Abstract:

This paper attempts to study colloquialism in audio-visual translation, with particular emphasis given to investigating the difficulties and challenges encountered by subtitlers in translating Lebanese colloquial into English. To achieve the main objectives of this study, ample and thorough cultural and translational analysis of examples drawn from the subtitled movie Capernaum are presented in order to identify the strategies used to overcome cultural barriers and differences and to show the process of decision-making by the translator. Also, special attention is given to explain the technicalities in translating subtitles and how they affect the translation process. The research is a descriptive analytical study whereby the writer sets out empirical observations, consisting of descriptive and analytical examination of the difficulties and problems associated with translating Arabic colloquialisms, specifically Lebanese, into English in the subtitled film, Capernaum. The research methodology utilizes a qualitative approach to group the selected data into the subtitling strategies presented by Gottlieb under the domesticating or foreignizing strategies according to Venuti's Model. It is shown that producing the same meanings to a foreign audience is not an easy task. The background of cultural elements and the stories that make up the history and mindset of the Lebanese and Arabic peoples leads to the use of the transfer and paraphrase methodologies most of the time (81% of the sample used for analysis). The research shows that translating and subtitling colloquialism needs special skills by the translators to overcome the challenges imposed by the limited presentation space as well as cultural differences. Translation of colloquial Arabic/Lebanese can be achieved to a certain extent and delivering the meaning and effect of the source language culture is accomplished in as much as the translator investigates and relates to the target culture.

Keywords: Lebanese colloquial, audio-visual translation, subtitling, Capernaum

Procedia PDF Downloads 129
400 Analysis of a IncResU-Net Model for R-Peak Detection in ECG Signals

Authors: Beatriz Lafuente Alcázar, Yash Wani, Amit J. Nimunkar

Abstract:

Cardiovascular Diseases (CVDs) are the leading cause of death globally, and around 80% of sudden cardiac deaths are due to arrhythmias or irregular heartbeats. The majority of these pathologies are revealed by either short-term or long-term alterations in the electrocardiogram (ECG) morphology. The ECG is the main diagnostic tool in cardiology. It is a non-invasive, pain free procedure that measures the heart’s electrical activity and that allows the detecting of abnormal rhythms and underlying conditions. A cardiologist can diagnose a wide range of pathologies based on ECG’s form alterations, but the human interpretation is subjective and it is contingent to error. Moreover, ECG records can be quite prolonged in time, which can further complicate visual diagnosis, and deeply retard disease detection. In this context, deep learning methods have risen as a promising strategy to extract relevant features and eliminate individual subjectivity in ECG analysis. They facilitate the computation of large sets of data and can provide early and precise diagnoses. Therefore, the cardiology field is one of the areas that can most benefit from the implementation of deep learning algorithms. In the present study, a deep learning algorithm is trained following a novel approach, using a combination of different databases as the training set. The goal of the algorithm is to achieve the detection of R-peaks in ECG signals. Its performance is further evaluated in ECG signals with different origins and features to test the model’s ability to generalize its outcomes. Performance of the model for detection of R-peaks for clean and noisy ECGs is presented. The model is able to detect R-peaks in the presence of various types of noise, and when presented with data, it has not been trained. It is expected that this approach will increase the effectiveness and capacity of cardiologists to detect divergences in the normal cardiac activity of their patients.

Keywords: arrhythmia, deep learning, electrocardiogram, machine learning, R-peaks

Procedia PDF Downloads 154
399 Phylogenetic Relationships between the Whole Sets of Individual Flow Sorted U, M, S and C Chromosomes of Aegilops and Wheat as Revealed by COS Markers

Authors: András Farkas, István Molnár, Jan Vrána, Veronika Burešová, Petr Cápal, András Cseh, Márta Molnár-Láng, Jaroslav Doležel

Abstract:

Species of Aegilops played a central role in the evolution of wheat and are sources of traits related to yield quality and tolerance against biotic and abiotic stresses. These wild genes and alleles are desirable to use in crop improvement programs via introgressive hybridization. However, the success of chromosome mediated gene transfer to wheat are hampered by the pour knowledge on the genome structure of Aegilops relative to wheat and by the low number of cost-effective molecular markers specific for Aegilops chromosomes. The COS markers specific for genes conserved throughout evolution in both sequence and copy number between Triticeae/Aegilops taxa and define orthologous regions, thus enabling the comparison of regions on the chromosomes of related species. The present study compared individual chromosomes of Aegilops umbellulata (UU), Ae. comosa (MM), Ae. speltoides (SS) and Ae. caudata (CC) purified by flourescent labelling with oligonucleotid SSR repeats and biparametric flow cytometry with wheat by identifying orthologous chromosomal regions by COS markers. The linear order of bin-mapped COS markers along the wheat D chromosomes was identified by the use of chromosome-specific sequence data and virtual gene order. Syntenic regions of wheat identifying genome rearrangements differentiating the U, M, S or C genomes from the D genome of wheat were detected. The conserved orthologous set markers assigned to Aegilops chromosomes promise to accelerate gene introgression by facilitating the identification of alien chromatin. The syntenic relationships between the Aegilops species and wheat will facilitate the targeted development of new markers specific for U, M, S and C genomic regions and will contribute to the understanding of molecular processes related to the evolution of Aegilops.

Keywords: Aegilops, cos-markers, flow-sorting, wheat

Procedia PDF Downloads 481
398 Remote Vital Signs Monitoring in Neonatal Intensive Care Unit Using a Digital Camera

Authors: Fatema-Tuz-Zohra Khanam, Ali Al-Naji, Asanka G. Perera, Kim Gibson, Javaan Chahl

Abstract:

Conventional contact-based vital signs monitoring sensors such as pulse oximeters or electrocardiogram (ECG) may cause discomfort, skin damage, and infections, particularly in neonates with fragile, sensitive skin. Therefore, remote monitoring of the vital sign is desired in both clinical and non-clinical settings to overcome these issues. Camera-based vital signs monitoring is a recent technology for these applications with many positive attributes. However, there are still limited camera-based studies on neonates in a clinical setting. In this study, the heart rate (HR) and respiratory rate (RR) of eight infants at the Neonatal Intensive Care Unit (NICU) in Flinders Medical Centre were remotely monitored using a digital camera applying color and motion-based computational methods. The region-of-interest (ROI) was efficiently selected by incorporating an image decomposition method. Furthermore, spatial averaging, spectral analysis, band-pass filtering, and peak detection were also used to extract both HR and RR. The experimental results were validated with the ground truth data obtained from an ECG monitor and showed a strong correlation using the Pearson correlation coefficient (PCC) 0.9794 and 0.9412 for HR and RR, respectively. The RMSE between camera-based data and ECG data for HR and RR were 2.84 beats/min and 2.91 breaths/min, respectively. A Bland Altman analysis of the data also showed a close correlation between both data sets with a mean bias of 0.60 beats/min and 1 breath/min, and the lower and upper limit of agreement -4.9 to + 6.1 beats/min and -4.4 to +6.4 breaths/min for both HR and RR, respectively. Therefore, video camera imaging may replace conventional contact-based monitoring in NICU and has potential applications in other contexts such as home health monitoring.

Keywords: neonates, NICU, digital camera, heart rate, respiratory rate, image decomposition

Procedia PDF Downloads 92
397 An Experimental Investigation of the Surface Pressure on Flat Plates in Turbulent Boundary Layers

Authors: Azadeh Jafari, Farzin Ghanadi, Matthew J. Emes, Maziar Arjomandi, Benjamin S. Cazzolato

Abstract:

The turbulence within the atmospheric boundary layer induces highly unsteady aerodynamic loads on structures. These loads, if not accounted for in the design process, will lead to structural failure and are therefore important for the design of the structures. For an accurate prediction of wind loads, understanding the correlation between atmospheric turbulence and the aerodynamic loads is necessary. The aim of this study is to investigate the effect of turbulence within the atmospheric boundary layer on the surface pressure on a flat plate over a wide range of turbulence intensities and integral length scales. The flat plate is chosen as a fundamental geometry which represents structures such as solar panels and billboards. Experiments were conducted at the University of Adelaide large-scale wind tunnel. Two wind tunnel boundary layers with different intensities and length scales of turbulence were generated using two sets of spires with different dimensions and a fetch of roughness elements. Average longitudinal turbulence intensities of 13% and 26% were achieved in each boundary layer, and the longitudinal integral length scale within the three boundary layers was between 0.4 m and 1.22 m. The pressure distributions on a square flat plate at different elevation angles between 30° and 90° were measured within the two boundary layers with different turbulence intensities and integral length scales. It was found that the peak pressure coefficient on the flat plate increased with increasing turbulence intensity and integral length scale. For example, the peak pressure coefficient on a flat plate elevated at 90° increased from 1.2 to 3 with increasing turbulence intensity from 13% to 26%. Furthermore, both the mean and the peak pressure distribution on the flat plates varied with turbulence intensity and length scale. The results of this study can be used to provide a more accurate estimation of the unsteady wind loads on structures such as buildings and solar panels.

Keywords: atmospheric boundary layer, flat plate, pressure coefficient, turbulence

Procedia PDF Downloads 121
396 Upcoming Fight Simulation with Smart Shadow

Authors: Ramiz Kuliev, Fuad Kuliev-Smirnov

Abstract:

The 'Shadow Sparring' training exercise is widely used in the training of boxers and martial artists. The main disadvantage of the usual shadow sparring is that the trainer cannot fully control such training and evaluate its results. During the competition, the athlete, preparing for the upcoming fight, imagines the Shadow (upcoming opponent) in accordance with his own imagination. A ‘Smart-Shadow Sparring’ (SSS) is an innovative version of the ‘Shadow Sparring’. During SSS, the fighter will see the Shadow (virtual opponent that moves, defends, and punches) and understand when he misses the punches from the Shadow. The task of a real athlete is to spar with a virtual one, move around, punch in the direction of unprotected areas of the Shadow and dodge his punches. Moves and punches of Shadow are set up before each training. The system will give the coach full information about virtual sparring: (i) how many and what type of punches has the fighter landed, (ii) accuracy of these punches, (iii) how many and what type of virtual punches (punches of Smart-Shadow) has the fighter missed, etc. SSS will be recorded as animated fighting of two fighters and will help the coach to analyze past training. SSS can be configured to fit the physical and technical characteristics of the next real opponent (size, techniques, speed, missed and landed punches, etc.). This will allow to simulate and rehearse the upcoming fight and improve readiness for the next opponent. For amateur fighters, SSS will be reconfigured several times during a tournament, when the real opponent becomes known. SSS can be used in three versions: (1) Digital Shadow: the athlete will see a Shadow on a monitor (2) VR-Shadow: the athlete will see a Shadow in a VR-glasses (3) Smart Shadow: a Shadow will be controlled by artificial intelligence. These technologies are based on the ‘semi-real simulation’ method. The technology allows coaches to train athletes remotely. Simulation of different opponents will help the athletes better prepare for competition. Repeat rehearsals of the upcoming fight will help improve results. SSS can improve results in Boxing, Taekwondo, Karate, and Fencing. 41 sets of medals will be awarded in these sports at the 2020 Olympic Games.

Keywords: boxing, combat sports, fight simulation, shadow sparring

Procedia PDF Downloads 108
395 Determining the Extent and Direction of Relief Transformations Caused by Ski Run Construction Using LIDAR Data

Authors: Joanna Fidelus-Orzechowska, Dominika Wronska-Walach, Jaroslaw Cebulski

Abstract:

Mountain areas are very often exposed to numerous transformations connected with the development of tourist infrastructure. In mountain areas in Poland ski tourism is very popular, so agricultural areas are often transformed into tourist areas. The construction of new ski runs can change the direction and rate of slope development. The main aim of this research was to determine geomorphological and hydrological changes within slopes caused by ski run constructions. The study was conducted in the Remiaszów catchment in the Inner Polish Carpathians (southern Poland). The mean elevation of the catchment is 859 m a.s.l. and the maximum is 946 m a.s.l. The surface area of the catchment is 1.16 km2, of which 16.8% is the area of the two studied ski runs. The studied ski runs were constructed in 2014 and 2015. In order to determine the relief transformations connected with new ski run construction high resolution LIDAR data was analyzed. The general relief changes in the studied catchment were determined on the basis of ALS (Airborne Laser Scanning ) data obtained before (2013) and after (2016) ski run construction. Based on the two sets of ALS data a digital elevation models of differences (DoDs) was created, which made it possible to determine the quantitative relief changes in the entire studied catchment. Additionally, cross and longitudinal profiles were calculated within slopes where new ski runs were built. Detailed data on relief changes within selected test surfaces was obtained based on TLS (Terrestrial Laser Scanning). Hydrological changes within the analyzed catchment were determined based on the convergence and divergence index. The study shows that the construction of the new ski runs caused significant geomorphological and hydrological changes in the entire studied catchment. However, the most important changes were identified within the ski slopes. After the construction of ski runs the entire catchment area lowered about 0.02 m. Hydrological changes in the studied catchment mainly led to the interruption of surface runoff pathways and changes in runoff direction and geometry.

Keywords: hydrological changes, mountain areas, relief transformations, ski run construction

Procedia PDF Downloads 130
394 Selection of Optimal Reduced Feature Sets of Brain Signal Analysis Using Heuristically Optimized Deep Autoencoder

Authors: Souvik Phadikar, Nidul Sinha, Rajdeep Ghosh

Abstract:

In brainwaves research using electroencephalogram (EEG) signals, finding the most relevant and effective feature set for identification of activities in the human brain is a big challenge till today because of the random nature of the signals. The feature extraction method is a key issue to solve this problem. Finding those features that prove to give distinctive pictures for different activities and similar for the same activities is very difficult, especially for the number of activities. The performance of a classifier accuracy depends on this quality of feature set. Further, more number of features result in high computational complexity and less number of features compromise with the lower performance. In this paper, a novel idea of the selection of optimal feature set using a heuristically optimized deep autoencoder is presented. Using various feature extraction methods, a vast number of features are extracted from the EEG signals and fed to the autoencoder deep neural network. The autoencoder encodes the input features into a small set of codes. To avoid the gradient vanish problem and normalization of the dataset, a meta-heuristic search algorithm is used to minimize the mean square error (MSE) between encoder input and decoder output. To reduce the feature set into a smaller one, 4 hidden layers are considered in the autoencoder network; hence it is called Heuristically Optimized Deep Autoencoder (HO-DAE). In this method, no features are rejected; all the features are combined into the response of responses of the hidden layer. The results reveal that higher accuracy can be achieved using optimal reduced features. The proposed HO-DAE is also compared with the regular autoencoder to test the performance of both. The performance of the proposed method is validated and compared with the other two methods recently reported in the literature, which reveals that the proposed method is far better than the other two methods in terms of classification accuracy.

Keywords: autoencoder, brainwave signal analysis, electroencephalogram, feature extraction, feature selection, optimization

Procedia PDF Downloads 97
393 Assimilating Multi-Mission Satellites Data into a Hydrological Model

Authors: Mehdi Khaki, Ehsan Forootan, Joseph Awange, Michael Kuhn

Abstract:

Terrestrial water storage, as a source of freshwater, plays an important role in human lives. Hydrological models offer important tools for simulating and predicting water storages at global and regional scales. However, their comparisons with 'reality' are imperfect mainly due to a high level of uncertainty in input data and limitations in accounting for all complex water cycle processes, uncertainties of (unknown) empirical model parameters, as well as the absence of high resolution (both spatially and temporally) data. Data assimilation can mitigate this drawback by incorporating new sets of observations into models. In this effort, we use multi-mission satellite-derived remotely sensed observations to improve the performance of World-Wide Water Resources Assessment system (W3RA) hydrological model for estimating terrestrial water storages. For this purpose, we assimilate total water storage (TWS) data from the Gravity Recovery And Climate Experiment (GRACE) and surface soil moisture data from the Advanced Microwave Scanning Radiometer for the Earth Observing System (AMSR-E) into W3RA. This is done to (i) improve model estimations of water stored in ground and soil moisture, and (ii) assess the impacts of each satellite of data (from GRACE and AMSR-E) and their combination on the final terrestrial water storage estimations. These data are assimilated into W3RA using the Ensemble Square-Root Filter (EnSRF) filtering technique over Mississippi Basin (the United States) and Murray-Darling Basin (Australia) between 2002 and 2013. In order to evaluate the results, independent ground-based groundwater and soil moisture measurements within each basin are used.

Keywords: data assimilation, GRACE, AMSR-E, hydrological model, EnSRF

Procedia PDF Downloads 267
392 Seismic Inversion for Geothermal Exploration

Authors: E. N. Masri, E. Takács

Abstract:

Amplitude Versus Offset (AVO) and simultaneous model-based impedance inversion techniques have not been utilized for geothermal exploration commonly; however, some recent publications called the attention that they can be very useful in the geothermal investigations. In this study, we present rock physical attributes obtained from 3D pre-stack seismic data and well logs collected in a study area of the NW part of Pannonian Basin where the geothermal reservoir is located in the fractured zones of Triassic basement and it was hit by three productive-injection well pairs. The holes were planned very successfully based on the conventional 3D migrated stack volume prior to this study. Subsequently, the available geophysical-geological datasets provided a great opportunity to test modern inversion procedures in the same area. In this presentation, we provide a summary of the theory and application of the most promising seismic inversion techniques from the viewpoint of geothermal exploration. We demonstrate P- and S-wave impedance, as well as the velocity (Vp and Vs), the density, and the Vp/Vs ratio attribute volumes calculated from the seismic and well-logging data sets. After a detailed discussion, we conclude that P-wave impedance and Vp/Vp ratio are the most helpful parameters for lithology discrimination in the study area. They detect the hot water saturated fracture zone very well thus they can be very useful in mapping the investigated reservoir. Integrated interpretation of all the obtained rock-physical parameters is essential. We are extending the above discussed pre-stack seismic tools by studying the possibilities of Elastic Impedance Inversion (EII) for geothermal exploration. That procedure provides two other useful rock-physical properties, the compressibility and the rigidity (Lamé parameters). Results of those newly created elastic parameters will also be demonstrated in the presentation. Geothermal extraction is of great interest nowadays; and we can adopt several methods have been successfully applied in the hydrocarbon exploration for decades to discover new reservoirs and reduce drilling risk and cost.

Keywords: fractured zone, seismic, well-logging, inversion

Procedia PDF Downloads 108
391 Curvature Based-Methods for Automatic Coarse and Fine Registration in Dimensional Metrology

Authors: Rindra Rantoson, Hichem Nouira, Nabil Anwer, Charyar Mehdi-Souzani

Abstract:

Multiple measurements by means of various data acquisition systems are generally required to measure the shape of freeform workpieces for accuracy, reliability and holisticity. The obtained data are aligned and fused into a common coordinate system within a registration technique involving coarse and fine registrations. Standardized iterative methods have been established for fine registration such as Iterative Closest Points (ICP) and its variants. For coarse registration, no conventional method has been adopted yet despite a significant number of techniques which have been developed in the literature to supply an automatic rough matching between data sets. Two main issues are addressed in this paper: the coarse registration and the fine registration. For coarse registration, two novel automated methods based on the exploitation of discrete curvatures are presented: an enhanced Hough Transformation (HT) and an improved Ransac Transformation. The use of curvature features in both methods aims to reduce computational cost. For fine registration, a new variant of ICP method is proposed in order to reduce registration error using curvature parameters. A specific distance considering the curvature similarity has been combined with Euclidean distance to define the distance criterion used for correspondences searching. Additionally, the objective function has been improved by combining the point-to-point (P-P) minimization and the point-to-plane (P-Pl) minimization with automatic weights. These ones are determined from the preliminary calculated curvature features at each point of the workpiece surface. The algorithms are applied on simulated and real data performed by a computer tomography (CT) system. The obtained results reveal the benefit of the proposed novel curvature-based registration methods.

Keywords: discrete curvature, RANSAC transformation, hough transformation, coarse registration, ICP variant, point-to-point and point-to-plane minimization combination, computer tomography

Procedia PDF Downloads 408
390 Using Seismic and GPS Data for Hazard Estimation in Some Active Regions in Egypt

Authors: Abdel-Monem Sayed Mohamed

Abstract:

Egypt rapidly growing development is accompanied by increasing levels of standard living particular in its urban areas. However, there is a limited experience in quantifying the sources of risk management in Egypt and in designing efficient strategies to keep away serious impacts of earthquakes. From the historical point of view and recent instrumental records, there are some seismo-active regions in Egypt, where some significant earthquakes had occurred in different places. The special tectonic features in Egypt: Aswan, Greater Cairo, Red Sea and Sinai Peninsula regions are the territories of a high seismic risk, which have to be monitored by up-to date technologies. The investigations of the seismic events and interpretations led to evaluate the seismic hazard for disaster prevention and for the safety of the dense populated regions and the vital national projects as the High Dam. In addition to the monitoring of the recent crustal movements, the most powerful technique of satellite geodesy GPS are used where geodetic networks are covering such seismo-active regions. The results from the data sets are compared and combined in order to determine the main characteristics of the deformation and hazard estimation for specified regions. The final compiled output from the seismological and geodetic analysis threw lights upon the geodynamical regime of these seismo-active regions and put Aswan and Greater Cairo under the lowest class according to horizontal crustal strains classifications. This work will serve a basis for the development of so-called catastrophic models and can be further used for catastrophic risk management. Also, this work is trying to evaluate risk of large catastrophic losses within the important regions including the High Dam, strategic buildings and archeological sites. Studies on possible scenarios of earthquakes and losses are a critical issue for decision making in insurance as a part of mitigation measures.

Keywords: b-value, Gumbel distribution, seismic and GPS data, strain parameters

Procedia PDF Downloads 434
389 House Price Index Predicts a Larger Impact of Habitat Loss than Primary Productivity on the Biodiversity of North American Avian Communities

Authors: Marlen Acosta Alamo, Lisa Manne, Richard Veit

Abstract:

Habitat loss due to land use change is one of the leading causes of biodiversity loss worldwide. This form of habitat loss is a non-random phenomenon since the same environmental factors that make an area suitable for supporting high local biodiversity overlap with those that make it attractive for urban development. We aimed to compare the effect of two non-random habitat loss predictors on the richness, abundance, and rarity of nature-affiliated and human-affiliated North American breeding birds. For each group of birds, we simulated the non-random habitat loss using two predictors: the House Price Index as a measure of the attractiveness of an area for humans and the Normalized Difference Vegetation Index as a proxy for primary productivity. We compared the results of the two non-random simulation sets and one set of random habitat loss simulations using an analysis of variance and followed up with a Tukey-Kramer test when appropriate. The attractiveness of an area for humans predicted estimates of richness loss and increase of rarity higher than primary productivity and random habitat loss for nature-affiliated and human-affiliated birds. For example, at 50% of habitat loss, the attractiveness of an area for humans produced estimates of richness at least 5% lower and of a rarity at least 40% higher than primary productivity and random habitat loss for both groups of birds. Only for the species abundance of nature-affiliated birds, the attractiveness of an area for humans did not outperform primary productivity as a predictor of biodiversity following habitat loss. We demonstrated the value of the House Price Index, which can be used in conservation assessments as an index of the risks of habitat loss for natural communities. Thus, our results have relevant implications for sustainable urban land-use planning practices and can guide stakeholders and developers in their efforts to conserve local biodiversity.

Keywords: biodiversity loss, bird biodiversity, house price index, non-random habitat loss

Procedia PDF Downloads 64
388 Pilot Study of Determining the Impact of Surface Subsidence at The Intersection of Cave Mining with the Surface Using an Electrical Impedance Tomography

Authors: Ariungerel Jargal

Abstract:

: Cave mining is a bulk underground mining method, which allows large low-grade deposits to be mined underground. This method involves undermining the orebody to make it collapse under its own weight into a series of chambers from which the ore extracted. It is a useful technique to extend the life of large deposits previously mined by open pits, and it is a method increasingly proposed for new mines around the world. We plan to conduct a feasibility study using Electrical impedance tomography (EIT) technology to show how much subsidence there is at the intersection with the cave mining surface. EIT is an imaging technique which uses electrical measurements at electrodes attached on the body surface to yield a cross-sectional image of conductivity changes within the object. EIT has been developed in several different applications areas as a simpler, cheaper alternative to many other imaging methods. A low frequency current is injected between pairs of electrodes while voltage measurements are collected at all other electrode pairs. In the difference EIT, images are reconstructed of the change in conductivity distribution (σ) between the acquisition of the two sets of measurements. Image reconstruction in EIT requires the solution of an ill-conditioned nonlinear inverse problem on noisy data, typically requiring make simpler assumptions or regularization. It is noted that the ratio of current to voltage represents a complex value according to Ohm’s law, and that it is theoretically possible to re-express EIT. The results of the experiment were presented on the simulation, and it was concluded that it is possible to conduct further real experiments. Drill a certain number of holes in the top wall of the cave to attach the electrodes, flow a current through them, and measure and acquire the potential through these electrodes. Appropriate values should be selected depending on the distance between the holes, the frequency and duration of the measurements, the surface characteristics and the size of the study area using an EIT device.

Keywords: impedance tomography, cave mining, soil, EIT device

Procedia PDF Downloads 102