Search results for: ecological networks
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4064

Search results for: ecological networks

464 Ethnobotanical Medicines for Treating Snakebites among the Indigenous Maya Populations of Belize

Authors: Kerry Hull, Mark Wright

Abstract:

This paper brings light to ethnobotanical medicines used by the Maya of Belize to treat snake bites. The varying ecological zones of Belize boast over fifty species of snakes, nine of which are poisonous and dangerous to humans. Two distinct Maya groups occupy neighboring regions of Belize, the Q’eqchi’ and the Mopan. With Western medical care often far from their villages, what traditional methods are used to treat poisonous snake bites? Based primarily on data gathered with native consultants during the authors’ fieldwork with both groups, this paper details the ethnobotanical resources used by the Q’eqchi’ and Mopan traditional healers. The Q’eqchi’ and Mopan most commonly rely on traditional ‘bush doctors’ (ilmaj in Mopan), both male and female, and specialized ‘snake doctors’ to heal bites from venomous snakes. First, this paper presents each plant employed by healers for bites for the nine poisonous snakes in Belize along with the specific botanical recipes and methods of application for each remedy. Individual chemical and therapeutic qualities of some of those plants are investigated in an effort to explain their possible medicinal value for different toxins or the symptoms caused by those toxins. In addition, this paper explores mythological associations with certain snakes that inform local understanding regarding which plants are considered efficacious in each case, arguing that numerous oral traditions (recorded by the authors) help to link botanical medicines to episodes within their mythic traditions. Finally, the use of plants to counteract snakebites brought about through sorcery is discussed inasmuch as some snakes are seen as ‘helpers’ of sorcerers. Snake bites given under these circumstances can only be cured by those who know both the proper corresponding plant(s) and ritual prayer(s). This paper provides detailed documentation of traditional ethnomedicines and practices from the dying art of traditional Maya healers and argues for multi-faceted diagnostic techniques to determine toxin severity, the presence or absence of sorcery, and the appropriate botanical remedy.

Keywords: ethnobotany, Maya, medicine, snake bites

Procedia PDF Downloads 239
463 Performance Analysis of Three Absorption Heat Pump Cycles, Full and Partial Loads Operations

Authors: B. Dehghan, T. Toppi, M. Aprile, M. Motta

Abstract:

The environmental concerns related to global warming and ozone layer depletion along with the growing worldwide demand for heating and cooling have brought an increasing attention toward ecological and efficient Heating, Ventilation, and Air Conditioning (HVAC) systems. Furthermore, since space heating accounts for a considerable part of the European primary/final energy use, it has been identified as one of the sectors with the most challenging targets in energy use reduction. Heat pumps are commonly considered as a technology able to contribute to the achievement of the targets. Current research focuses on the full load operation and seasonal performance assessment of three gas-driven absorption heat pump cycles. To do this, investigations of the gas-driven air-source ammonia-water absorption heat pump systems for small-scale space heating applications are presented. For each of the presented cycles, both full-load under various temperature conditions and seasonal performances are predicted by means of numerical simulations. It has been considered that small capacity appliances are usually equipped with fixed geometry restrictors, meaning that the solution mass flow rate is driven by the pressure difference across the associated restrictor valve. Results show that gas utilization efficiency (GUE) of the cycles varies between 1.2 and 1.7 for both full and partial loads and vapor exchange (VX) cycle is found to achieve the highest efficiency. It is noticed that, for typical space heating applications, heat pumps operate over a wide range of capacities and thermal lifts. Thus, partially, the novelty introduced in the paper is the investigation based on a seasonal performance approach, following the method prescribed in a recent European standard (EN 12309). The overall result is a modest variation in the seasonal performance for analyzed cycles, from 1.427 (single-effect) to 1.493 (vapor-exchange).

Keywords: absorption cycles, gas utilization efficiency, heat pump, seasonal performance, vapor exchange cycle

Procedia PDF Downloads 114
462 In Search of Innovation: Exploring the Dynamics of Innovation

Authors: Michal Lysek, Mike Danilovic, Jasmine Lihua Liu

Abstract:

HMS Industrial Networks AB has been recognized as one of the most innovative companies in the industrial communication industry worldwide. The creation of their Anybus innovation during the 1990s contributed considerably to the company’s success. From inception, HMS’ employees were innovating for the purpose of creating new business (the creation phase). After the Anybus innovation, they began the process of internationalization (the commercialization phase), which in turn led them to concentrate on cost reduction, product quality, delivery precision, operational efficiency, and increasing growth (the growth phase). As a result of this transformation, performing new radical innovations have become more complicated. The purpose of our research was to explore the dynamics of innovation at HMS from the aspect of key actors, activities, and events, over the three phases, in order to understand what led to the creation of their Anybus innovation, and why it has become increasingly challenging for HMS to create new radical innovations for the future. Our research methodology was based on a longitudinal, retrospective study from the inception of HMS in 1988 to 2014, a single case study inspired by the grounded theory approach. We conducted 47 interviews and collected 1 024 historical documents for our research. Our analysis has revealed that HMS’ success in creating the Anybus, and developing a successful business around the innovation, was based on three main capabilities – cultivating customer relations on different managerial and organizational levels, inspiring business relations, and balancing complementary human assets for the purpose of business creation. The success of HMS has turned the management’s attention away from past activities of key actors, of their behavior, and how they influenced and stimulated the creation of radical innovations. Nowadays, they are rhetorically focusing on creativity and innovation. All the while, their real actions put emphasis on growth, cost reduction, product quality, delivery precision, operational efficiency, and moneymaking. In the process of becoming an international company, HMS gradually refocused. In so doing they became profitable and successful, but they also forgot what made them innovative in the first place. Fortunately, HMS’ management has come to realize that this is the case and they are now in search of recapturing innovation once again. Our analysis indicates that HMS’ management is facing several barriers to innovation related path dependency and other lock-in phenomena. HMS’ management has been captured, trapped in their mindset and actions, by the success of the past. But now their future has to be secured, and they have come to realize that moneymaking is not everything. In recent years, HMS’ management have begun to search for innovation once more, in order to recapture their past capabilities for creating radical innovations. In order to unlock their managerial perceptions of customer needs and their counter-innovation driven activities and events, to utilize the full potential of their employees and capture the innovation opportunity for the future.

Keywords: barriers to innovation, dynamics of innovation, in search of excellence and innovation, radical innovation

Procedia PDF Downloads 380
461 Characterization of Antibiotic Resistance in Cultivable Enterobacteriaceae Isolates from Different Ecological Niches in the Eastern Cape, South Africa

Authors: Martins A. Adefisoye, Mpaka Lindelwa, Fadare Folake, Anthony I. Okoh

Abstract:

Evolution and rapid dissemination of antibiotic resistance from one ecosystem to another has been responsible for wide-scale epidemic and endemic spreads of multi-drug resistance pathogens. This study assessed the prevalence of Enterobacteriaceae in different environmental samples, including river water, hospital effluents, abattoir wastewater, animal rectal swabs and faecal droppings, soil, and vegetables, using standard microbiological procedure. The identity of the isolates were confirmed using matrix-assisted laser desorption ionization-time of flight mass spectrophotometry (MALDI-TOF) while the isolates were profiled for resistance against a panel of 16 antibiotics using disc diffusion (DD) test, and the occurrence of resistance genes (ARG) was determined by polymerase chain reactions (PCR). Enterobacteriaceae counts in the samples range as follows: river water 4.0 × 101 – 2.0 × 104 cfu/100 ml, hospital effluents 1.5 × 103 – 3.0 × 107 cfu/100 ml, municipal wastewater 2.3 × 103 – 9.2 × 104 cfu/100 ml, faecal droppings 3.0 × 105 – 9.5 × 106 cfu/g, animal rectal swabs 3.0 × 102 – 2.9 × 107 cfu/ml, soil 0 – 1.2 × 105 cfu/g and vegetables 0 – 2.2 × 107 cfu/g. Of the 700 randomly selected presumptive isolates subjected to MALDI-TOF analysis, 129 (18.4%), 68 (9.7%), 67 (9.5%), 41 (5.9%) were E. coli, Klebsiella spp., Enterobacter spp., and Citrobacter spp. respectively while the remaining isolates belong to other genera not targeted in the study. The DD test shows resistance ranging between 91.6% (175/191) for cefuroxime and (15.2%, 29/191) for imipenem The predominant multiple antibiotic resistance phenotypes (MARP), (GM-AUG-AP-CTX-CXM-CIP-NOR-NI-C-NA-TS-T-DXT) occurred in 9 Klebsiella isolates. The multiple antibiotic resistance indices (MARI) the isolates (range 0.17–1.0) generally showed >95% had MARI above the 0.2 thresholds, suggesting that most of the isolates originate from high-risk environments with high antibiotic use and high selective pressure for the emergence of resistance. The associated ARG in the isolates include: bla TEM 61.9 (65), bla SHV 1.9 (2), bla OXA 8.6 (9), CTX-M-2 8.6 (9), CTX-M-9 6.7 (7), sul 2 26.7 (28), tet A 16.2 (17), tet M 17.1 (18), aadA 59.1 (62), strA 34.3 (36), aac(3)A 19.1 (20), (aa2)A 7.6 (8), and aph(3)-1A 10.5 (11). The results underscore the need for preventative measures to curb the proliferation of antibiotic-resistant bacteria including Enterobacteriaceae to protect public health.

Keywords: enterobacteriaceae, antibiotic-resistance, MALDI-TOF, resistance genes, MARP, MARI, public health

Procedia PDF Downloads 155
460 A Top-down vs a Bottom-up Approach on Lower Extremity Motor Recovery and Balance Following Acute Stroke: A Randomized Clinical Trial

Authors: Vijaya Kumar, Vidayasagar Pagilla, Abraham Joshua, Rakshith Kedambadi, Prasanna Mithra

Abstract:

Background: Post stroke rehabilitation are aimed to accelerate for optimal sensorimotor recovery, functional gain and to reduce long-term dependency. Intensive physical therapy interventions can enhance this recovery as experience-dependent neural plastic changes either directly act at cortical neural networks or at distal peripheral level (muscular components). Neuromuscular Electrical Stimulation (NMES), a traditional bottom-up approach, mirror therapy (MT), a relatively new top down approach have found to be an effective adjuvant treatment methods for lower extremity motor and functional recovery in stroke rehabilitation. However there is a scarcity of evidence to compare their therapeutic gain in stroke recovery.Aim: To compare the efficacy of neuromuscular electrical stimulation (NMES) and mirror therapy (MT) in very early phase of post stroke rehabilitation addressed to lower extremity motor recovery and balance. Design: observer blinded Randomized Clinical Trial. Setting: Neurorehabilitation Unit, Department of Physical Therapy, Tertiary Care Hospitals. Subjects: 32 acute stroke subjects with first episode of unilateral stroke with hemiparesis, referred for rehabilitation (onset < 3 weeks), Brunnstorm lower extremity recovery stages ≥3 and MMSE score more than 24 were randomized into two group [Group A-NMES and Group B-MT]. Interventions: Both the groups received eclectic approach to remediate lower extremity recovery which includes treatment components of Roods, Bobath and Motor learning approaches for 30 minutes a day for 6 days. Following which Group A (N=16) received 30 minutes of surface NMES training for six major paretic muscle groups (gluteus maximus and medius,quadriceps, hamstrings, tibialis anterior and gastrocnemius). Group B (N=16) was administered with 30 minutes of mirror therapy sessions to facilitate lower extremity motor recovery. Outcome measures: Lower extremity motor recovery, balance and activities of daily life (ADLs) were measured by Fugyl Meyer Assessment (FMA-LE), Berg Balance Scale (BBS), Barthel Index (BI) before and after intervention. Results: Pre Post analysis of either group across the time revealed statistically significant improvement (p < 0.001) for all the outcome variables for the either group. All parameters of NMES had greater change scores compared to MT group as follows: FMA-LE (25.12±3.01 vs. 23.31±2.38), BBS (35.12±4.61 vs. 34.68±5.42) and BI (40.00±10.32 vs. 37.18±7.73). Between the groups comparison of pre post values showed no significance with FMA-LE (p=0.09), BBS (p=0.80) and BI (p=0.39) respectively. Conclusion: Though either groups had significant improvement (pre to post intervention), none of them were superior to other in lower extremity motor recovery and balance among acute stroke subjects. We conclude that eclectic approach is an effective treatment irrespective of NMES or MT as an adjunct.

Keywords: balance, motor recovery, mirror therapy, neuromuscular electrical stimulation, stroke

Procedia PDF Downloads 283
459 Expression of Micro-RNA268 in Zinc Deficient Rice

Authors: Sobia Shafqat, Saeed Ahmad Qaisrani

Abstract:

MicroRNAs play an essential role in the regulation and development of all processes in most eukaryotes because of their prospective part as mediators controlling cell growth and differentiation towards the exact position of RNAs response in plants under biotic and abiotic factors or stressors. In a few cases, Zn is oblivious poisonous for plants due to its heavy metal status. Some other metals are extremely toxic, like Cd, Hg, and Pb, but these elements require in rice for the programming of genes under abiotic stress resembling Zn stress when micro RNAs268 was importantly introduced in rice. The micro RNAs overexpressed in transgenic plants with an accumulation of a large amount of melanin dialdehyde, hydrogen peroxide, and an excessive quantity of Zn in the seedlings stage. Let out results for rice pliability under Zn stress micro RNAs act as negative controllers. But the role of micro RNA268 act as a modulator in different ecological condition. It has been explained clearly with a long understanding of the role of micro RNA268 under stress conditions; pliability and practically showed outcome to increase plant sufferance under Zn stress because micro RNAs is an intervention technique for gene regulation in gene expression. The proposed study was experimented with by using genetic factors of Zn stress and toxicity effect on rice plants done at District Vehari, Pakistan. The trial was performed randomly with three replications in a complete block design (RCBD). These blocks were controlled with different concentrations of genetic factors. By overexpression of micro RNA268 rice, seedling growth was not stopped under Zn deficiency due to the accumulation of a large amount of melanin dialdehyde, hydrogen peroxide, and an excessive quantity of Zn in their seedlings. Results showed that micro RNA268 act as a negative controller under Zn stress. In the end, under stress conditions, micro RNA268 showed the necessary function in the tolerance of rice plants. The directorial work sketch gave out high agronomic applications and yield outcomes in rice with a specific amount of Zn application.

Keywords: micro RNA268, zinc, rice, agronomic approach

Procedia PDF Downloads 63
458 Understanding the Impact of Out-of-Sequence Thrust Dynamics on Earthquake Mitigation: Implications for Hazard Assessment and Disaster Planning

Authors: Rajkumar Ghosh

Abstract:

Earthquakes pose significant risks to human life and infrastructure, highlighting the importance of effective earthquake mitigation strategies. Traditional earthquake modelling and mitigation efforts have largely focused on the primary fault segments and their slip behaviour. However, earthquakes can exhibit complex rupture dynamics, including out-of-sequence thrust (OOST) events, which occur on secondary or subsidiary faults. This abstract examines the impact of OOST dynamics on earthquake mitigation strategies and their implications for hazard assessment and disaster planning. OOST events challenge conventional seismic hazard assessments by introducing additional fault segments and potential rupture scenarios that were previously unrecognized or underestimated. Consequently, these events may increase the overall seismic hazard in affected regions. The study reviews recent case studies and research findings that illustrate the occurrence and characteristics of OOST events. It explores the factors contributing to OOST dynamics, such as stress interactions between fault segments, fault geometry, and mechanical properties of fault materials. Moreover, it investigates the potential triggers and precursory signals associated with OOST events to enhance early warning systems and emergency response preparedness. The abstract also highlights the significance of incorporating OOST dynamics into seismic hazard assessment methodologies. It discusses the challenges associated with accurately modelling OOST events, including the need for improved understanding of fault interactions, stress transfer mechanisms, and rupture propagation patterns. Additionally, the abstract explores the potential for advanced geophysical techniques, such as high-resolution imaging and seismic monitoring networks, to detect and characterize OOST events. Furthermore, the abstract emphasizes the practical implications of OOST dynamics for earthquake mitigation strategies and urban planning. It addresses the need for revising building codes, land-use regulations, and infrastructure designs to account for the increased seismic hazard associated with OOST events. It also underscores the importance of public awareness campaigns to educate communities about the potential risks and safety measures specific to OOST-induced earthquakes. This sheds light on the impact of out-of-sequence thrust dynamics in earthquake mitigation. By recognizing and understanding OOST events, researchers, engineers, and policymakers can improve hazard assessment methodologies, enhance early warning systems, and implement effective mitigation measures. By integrating knowledge of OOST dynamics into urban planning and infrastructure development, societies can strive for greater resilience in the face of earthquakes, ultimately minimizing the potential for loss of life and infrastructure damage.

Keywords: earthquake mitigation, out-of-sequence thrust, seismic, satellite imagery

Procedia PDF Downloads 91
457 Contextual Toxicity Detection with Data Augmentation

Authors: Julia Ive, Lucia Specia

Abstract:

Understanding and detecting toxicity is an important problem to support safer human interactions online. Our work focuses on the important problem of contextual toxicity detection, where automated classifiers are tasked with determining whether a short textual segment (usually a sentence) is toxic within its conversational context. We use “toxicity” as an umbrella term to denote a number of variants commonly named in the literature, including hate, abuse, offence, among others. Detecting toxicity in context is a non-trivial problem and has been addressed by very few previous studies. These previous studies have analysed the influence of conversational context in human perception of toxicity in controlled experiments and concluded that humans rarely change their judgements in the presence of context. They have also evaluated contextual detection models based on state-of-the-art Deep Learning and Natural Language Processing (NLP) techniques. Counterintuitively, they reached the general conclusion that computational models tend to suffer performance degradation in the presence of context. We challenge these empirical observations by devising better contextual predictive models that also rely on NLP data augmentation techniques to create larger and better data. In our study, we start by further analysing the human perception of toxicity in conversational data (i.e., tweets), in the absence versus presence of context, in this case, previous tweets in the same conversational thread. We observed that the conclusions of previous work on human perception are mainly due to data issues: The contextual data available does not provide sufficient evidence that context is indeed important (even for humans). The data problem is common in current toxicity datasets: cases labelled as toxic are either obviously toxic (i.e., overt toxicity with swear, racist, etc. words), and thus context does is not needed for a decision, or are ambiguous, vague or unclear even in the presence of context; in addition, the data contains labeling inconsistencies. To address this problem, we propose to automatically generate contextual samples where toxicity is not obvious (i.e., covert cases) without context or where different contexts can lead to different toxicity judgements for the same tweet. We generate toxic and non-toxic utterances conditioned on the context or on target tweets using a range of techniques for controlled text generation(e.g., Generative Adversarial Networks and steering techniques). On the contextual detection models, we posit that their poor performance is due to limitations on both of the data they are trained on (same problems stated above) and the architectures they use, which are not able to leverage context in effective ways. To improve on that, we propose text classification architectures that take the hierarchy of conversational utterances into account. In experiments benchmarking ours against previous models on existing and automatically generated data, we show that both data and architectural choices are very important. Our model achieves substantial performance improvements as compared to the baselines that are non-contextual or contextual but agnostic of the conversation structure.

Keywords: contextual toxicity detection, data augmentation, hierarchical text classification models, natural language processing

Procedia PDF Downloads 174
456 Exploring Accessible Filmmaking and Video for Deafblind Audiences through Multisensory Participatory Design

Authors: Aikaterini Tavoulari, Mike Richardson

Abstract:

Objective: This abstract presents a multisensory participatory design project, inspired by a deafblind PhD student's ambition to climb Mount Everest. The project aims to explore accessible routes for filmmaking and video content creation, catering to the needs of individuals with hearing and sight loss. By engaging participants from the Southwest area of England, recruited through multiple networks, the project seeks to gather qualitative data and insights to inform the development of inclusive media practices. Design: It will be a community-based participatory research design. The workshop will feature various stations that stimulate different senses, such as scent, touch, sight, hearing as well as movement. Participants will have the opportunity to engage with these multisensory experiences, providing valuable feedback on their effectiveness and potential for enhancing accessibility in filmmaking and video content. Methods: Brief semi-structured interviews will be conducted to collect qualitative data, allowing participants to share their perspectives, challenges, and suggestions for improvement. The participatory design approach emphasizes the importance of involving the target audience in the creative process. By actively engaging individuals with hearing and sight loss, the project aims to ensure that their needs and preferences are central to the development of accessible filmmaking techniques and video content. This collaborative effort seeks to bridge the gap between content creators and diverse audiences, fostering a more inclusive media landscape. Results: The findings from this study will contribute to the growing body of research on accessible filmmaking and video content creation. Via inductive thematic analysis of the qualitative data collected through interviews and observations, the researchers aim to identify key themes, challenges, and opportunities for creating engaging and inclusive media experiences for deafblind audiences. The insights will inform the development of best practices and guidelines for accessible filmmaking, empowering content creators to produce more inclusive and immersive video content. Conclusion: The abstract targets the hybrid International Conference for Disability and Diversity in Canada (January 2025), as this platform provides an excellent opportunity to share the outcomes of the project with a global audience of researchers, practitioners, and advocates working towards inclusivity and accessibility in various disability domains. By presenting this research at the conference in person, the authors aim to contribute to the ongoing discourse on disability and diversity, highlighting the importance of multisensory experiences and participatory design in creating accessible media content for the deafblind community and the community with sensory impairments more broadly.

Keywords: vision impairment, hearing impairment, deafblindness, accessibility, filmmaking

Procedia PDF Downloads 48
455 Evolving Urban Landscapes: Smart Cities and Sustainable Futures

Authors: Mehrzad Soltani, Pegah Rezaei

Abstract:

In response to the escalating challenges posed by resource scarcity, urban congestion, and the dearth of green spaces, contemporary urban areas have undergone a remarkable transformation into smart cities. This evolution necessitates a strategic and forward-thinking approach to urban development, with the primary objective of diminishing and eventually eradicating dependence on non-renewable energy sources. This steadfast commitment to sustainable development is geared toward the continual enhancement of our global urban milieu, ensuring a healthier and more prosperous environment for forthcoming generations. This transformative vision has been meticulously shaped by an extensive research framework, incorporating in-depth field studies and investigations conducted at both neighborhood and city levels. Our holistic strategy extends its purview to encompass major cities and states, advocating for the realization of exceptional development firmly rooted in the principles of sustainable intelligence. At its core, this approach places a paramount emphasis on stringent pollution control measures, concurrently safeguarding ecological equilibrium and regional cohesion. Central to the realization of this vision is the widespread adoption of environmentally friendly materials and components, championing the cultivation of plant life and harmonious green spaces, and the seamless integration of intelligent lighting and irrigation systems. These systems, including solar panels and solar energy utilization, are deployed wherever feasible, effectively meeting the essential lighting and irrigation needs of these dynamic urban ecosystems. Overall, the transformation of urban areas into smart cities necessitates a holistic and innovative approach to urban development. By actively embracing sustainable intelligence and adhering to strict environmental standards, these cities pave the way for a brighter and more sustainable future, one that is marked by resilient, thriving, and eco-conscious urban communities.

Keywords: smart city, green urban, sustainability, urban management

Procedia PDF Downloads 76
454 Management of High Conservation Value Forests (HCVF) in Peninsular Malaysia as Part of Sustainable Forest Management Practices

Authors: Abu Samah Abdul Khalim, Hamzah Khali Aziz

Abstract:

Tropical forests in Malaysia safeguard enormous biological diversity while providing crucial benefits and services for the sustainable development of human communities. They are highly significant globally, both for their diverse and threatened species and as representative unique ecosystems. In order to promote the conservation and sustainable management of forest in this country, the Forestry Department (FD) is using ITTO guidelines on managing the forest under the Sustainable Forest Management practice (SFM). The fundamental principles of SFM are the sustained provision of products, goods and services; economic viability, social acceptability and the minimization of environmental/ecological impacts. With increased awareness and recognition of the importance of tropical forests and biodiversity in the global environment, efforts have been made to classify forests and natural areas with unique values or properties in a universally accepted scale. In line with that the concept of High Conservation Value Forest (HCVF) first used by the Forest Stewardship Council (FSC) in 1999, has been adopted and included as Principle ‘9’ in the Malaysia Criteria and Indicators for Forest Management Certification (MC&I 2002). The MC&I 2002 is a standard used for assessing forest management practices of the Forest Management Unit (FMU) level for purpose of certification. The key to the concept of HCVF is identification of HCVs of the forest. This paper highlighted initiative taken by the Forestry Department Peninsular Malaysia in establishing and managing HCVF areas within the Permanent Forest Reserves (PFE). To date almost all states forestry department in Peninsular Malaysia have established HCVFs in their respective states under different categories. Among others, the establishments of HCVF in this country are related to the importance of conserving biological diversity of the flora in the natural forest in particular endemic and threatened species such as Shorea bentongensis. As such it is anticipated that by taking this important initiatives, it will promote the conservation of biological diversity in the PFE of Peninsular Malaysia in line with the Sustainable Forest Management practice.

Keywords: high conservation value forest, sustainable forest management, forest management certification, Peninsular Malaysia

Procedia PDF Downloads 333
453 Artificial Neural Network Approach for GIS-Based Soil Macro-Nutrients Mapping

Authors: Shahrzad Zolfagharnassab, Abdul Rashid Mohamed Shariff, Siti Khairunniza Bejo

Abstract:

Conventional methods for nutrient soil mapping are based on laboratory tests of samples that are obtained from surveys. The time and cost involved in gathering and analyzing soil samples are the reasons that researchers use Predictive Soil Mapping (PSM). PSM can be defined as the development of a numerical or statistical model of the relationship among environmental variables and soil properties, which is then applied to a geographic database to create a predictive map. Kriging is a group of geostatistical techniques to spatially interpolate point values at an unobserved location from observations of values at nearby locations. The main problem with using kriging as an interpolator is that it is excessively data-dependent and requires a large number of closely spaced data points. Hence, there is a need to minimize the number of data points without sacrificing the accuracy of the results. In this paper, an Artificial Neural Networks (ANN) scheme was used to predict macronutrient values at un-sampled points. ANN has become a popular tool for prediction as it eliminates certain difficulties in soil property prediction, such as non-linear relationships and non-normality. Back-propagation multilayer feed-forward network structures were used to predict nitrogen, phosphorous and potassium values in the soil of the study area. A limited number of samples were used in the training, validation and testing phases of ANN (pattern reconstruction structures) to classify soil properties and the trained network was used for prediction. The soil analysis results of samples collected from the soil survey of block C of Sawah Sempadan, Tanjung Karang rice irrigation project at Selangor of Malaysia were used. Soil maps were produced by the Kriging method using 236 samples (or values) that were a combination of actual values (obtained from real samples) and virtual values (neural network predicted values). For each macronutrient element, three types of maps were generated with 118 actual and 118 virtual values, 59 actual and 177 virtual values, and 30 actual and 206 virtual values, respectively. To evaluate the performance of the proposed method, for each macronutrient element, a base map using 236 actual samples and test maps using 118, 59 and 30 actual samples respectively produced by the Kriging method. A set of parameters was defined to measure the similarity of the maps that were generated with the proposed method, termed the sample reduction method. The results show that the maps that were generated through the sample reduction method were more accurate than the corresponding base maps produced through a smaller number of real samples. For example, nitrogen maps that were produced from 118, 59 and 30 real samples have 78%, 62%, 41% similarity, respectively with the base map (236 samples) and the sample reduction method increased similarity to 87%, 77%, 71%, respectively. Hence, this method can reduce the number of real samples and substitute ANN predictive samples to achieve the specified level of accuracy.

Keywords: artificial neural network, kriging, macro nutrient, pattern recognition, precision farming, soil mapping

Procedia PDF Downloads 75
452 Assessment of DNA Sequence Encoding Techniques for Machine Learning Algorithms Using a Universal Bacterial Marker

Authors: Diego Santibañez Oyarce, Fernanda Bravo Cornejo, Camilo Cerda Sarabia, Belén Díaz Díaz, Esteban Gómez Terán, Hugo Osses Prado, Raúl Caulier-Cisterna, Jorge Vergara-Quezada, Ana Moya-Beltrán

Abstract:

The advent of high-throughput sequencing technologies has revolutionized genomics, generating vast amounts of genetic data that challenge traditional bioinformatics methods. Machine learning addresses these challenges by leveraging computational power to identify patterns and extract information from large datasets. However, biological sequence data, being symbolic and non-numeric, must be converted into numerical formats for machine learning algorithms to process effectively. So far, some encoding methods, such as one-hot encoding or k-mers, have been explored. This work proposes additional approaches for encoding DNA sequences in order to compare them with existing techniques and determine if they can provide improvements or if current methods offer superior results. Data from the 16S rRNA gene, a universal marker, was used to analyze eight bacterial groups that are significant in the pulmonary environment and have clinical implications. The bacterial genes included in this analysis are Prevotella, Abiotrophia, Acidovorax, Streptococcus, Neisseria, Veillonella, Mycobacterium, and Megasphaera. These data were downloaded from the NCBI database in Genbank file format, followed by a syntactic analysis to selectively extract relevant information from each file. For data encoding, a sequence normalization process was carried out as the first step. From approximately 22,000 initial data points, a subset was generated for testing purposes. Specifically, 55 sequences from each bacterial group met the length criteria, resulting in an initial sample of approximately 440 sequences. The sequences were encoded using different methods, including one-hot encoding, k-mers, Fourier transform, and Wavelet transform. Various machine learning algorithms, such as support vector machines, random forests, and neural networks, were trained to evaluate these encoding methods. The performance of these models was assessed using multiple metrics, including the confusion matrix, ROC curve, and F1 Score, providing a comprehensive evaluation of their classification capabilities. The results show that accuracies between encoding methods vary by up to approximately 15%, with the Fourier transform obtaining the best results for the evaluated machine learning algorithms. These findings, supported by the detailed analysis using the confusion matrix, ROC curve, and F1 Score, provide valuable insights into the effectiveness of different encoding methods and machine learning algorithms for genomic data analysis, potentially improving the accuracy and efficiency of bacterial classification and related genomic studies.

Keywords: DNA encoding, machine learning, Fourier transform, Fourier transformation

Procedia PDF Downloads 30
451 Design and Development of an Autonomous Beach Cleaning Vehicle

Authors: Mahdi Allaoua Seklab, Süleyman BaşTürk

Abstract:

In the quest to enhance coastal environmental health, this study introduces a fully autonomous beach cleaning machine, a breakthrough in leveraging green energy and advanced artificial intelligence for ecological preservation. Designed to operate independently, the machine is propelled by a solar-powered system, underscoring a commitment to sustainability and the use of renewable energy in autonomous robotics. The vehicle's autonomous navigation is achieved through a sophisticated integration of LIDAR and a camera system, utilizing an SSD MobileNet V2 object detection model for accurate and real-time trash identification. The SSD framework, renowned for its efficiency in detecting objects in various scenarios, is coupled with the lightweight and precise highly MobileNet V2 architecture, making it particularly suited for the computational constraints of on-board processing in mobile robotics. Training of the SSD MobileNet V2 model was conducted on Google Colab, harnessing cloud-based GPU resources to facilitate a rapid and cost-effective learning process. The model was refined with an extensive dataset of annotated beach debris, optimizing the parameters using the Adam optimizer and a cross-entropy loss function to achieve high-precision trash detection. This capability allows the machine to intelligently categorize and target waste, leading to more effective cleaning operations. This paper details the design and functionality of the beach cleaning machine, emphasizing its autonomous operational capabilities and the novel application of AI in environmental robotics. The results showcase the potential of such technology to fill existing gaps in beach maintenance, offering a scalable and eco-friendly solution to the growing problem of coastal pollution. The deployment of this machine represents a significant advancement in the field, setting a new standard for the integration of autonomous systems in the service of environmental stewardship.

Keywords: autonomous beach cleaning machine, renewable energy systems, coastal management, environmental robotics

Procedia PDF Downloads 34
450 Human Activities Damaging the Ecosystem of Isheri Ogun River, South West Nigeria

Authors: N. B. Ikenweiwe, A. A. Alimi, N. A. Bamidele, O. A. Ewumi, K. Fasina, S. O. Otubusin

Abstract:

A study on the physical, chemical and biological parameters of the lower course of Ogun River, Isheri-Olofin was carried out between January and December 2014 in order to determine the effects of the anthropogenic activities of the Kara abattoir and domestic waste depositions on the quality of the water. Water samples were taken twice each month at three selected stations A, B and C (based on characteristic features or activity levels) along the water course. Samples were analysed using standard methods for chemical and biological parameters the same day in the laboratory while physical parameters were determined in-situ with water parameters kit. Generally, results of Transparency, Dissolved Oxygen, Nitrates, TDS and Alkalinity fall below the permissible limits of WHO and FEPA standards for drinking and fish production. Results of phosphates, lead and cadmium were also low but still within the permissible limit. Only Temperature and pH were within limit. Low plankton community, (phytoplankton, zooplankton), which ranges from 3, 5 to 40, 23 were as a result of low levels of DO, transparency and phosphate. The presence of coliform bacteria of public health importance like Escherichia coli, Proteus vulgaris, Aeromonas sp., Shigella sp, Enterobacter aerogenes as well as gram negative bacteria Proteus morganii are mainly indicators of faecal pollution. Fish and other resources obtained from this water stand the risk of being contaminated with these organisms and man is at the receiving end. The results of the physical, chemical and some biological parameters of Isheri, Ogun River, according to this study showed that the live forms of aquatic and fisheries resources there are dwelling under stress as a result of deposition of bones, horns, faecal components, slurry of suspended solids, fat and blood into the water. Government should therefore establish good monitoring system against illegal waste depositions and create education programmes that will enlighten the community on the social, ecological and economic values of the river.

Keywords: damage, ecosystem, human activities, Isheri ogun river

Procedia PDF Downloads 549
449 Ecocentric Principles for the Change of the Anthropocentric Design Within the Other Species Related Fields

Authors: Armando Cuspinera

Abstract:

Humans are nature itself, being with non-human species part of the same ecosystem, but the praxis reflects that not all relations are the same. In fields of design such as Biomimicry, Biodesign, and Biophilic design exist different approaches towards nature, nevertheless, anthropocentric principles such as domination, objectivization, or exploitation are defined in the same as ecocentric principles of inherent importance in life itself. Anthropocentrism has showed humanity with pollution of the earth, water, air, and the destruction of whole ecosystems from monocultures and rampant production of useless objects that life cannot outstand this unaware rhythm of life focused only for the human benefits. Even if by nature the biosphere is resilient, studies showed in the Paris Agreement explain that humanity will perish in an unconscious way of praxis. This is why is important to develop a differentiation between anthropocentric and ecocentricprinciples in the praxis of design, in order to enhance respect, valorization, and positive affectivity towards other life forms is necessary to analyze what principles are reproduced from each practice of design. It is only from the study of immaterial dimensions of design such as symbolism, epistemology, and ontology that the relation towards nature can be redesigned, and in order to do so, it must be studies from the dimensions of ontological design what principles –anthropocentric or ecocentric- through what the objects enhance or focus the perception humans have to its surrounding. The things we design also design us is the principle of ontological design, and in order to develop a way of ecological design in which is possible to consider other species as users, designers or collaborators is important to extend the studies and relation to other living forms from a transdisciplinary perspective of techniques, knowledge, practice, and disciplines in general. Materials, technologies, and any kind of knowledge have the principle of a tool: is not good nor bad, but is in the way of using it the possibilities that exist within them. The collaboration of disciplines and fields of study gives the opportunity to connect principles from other cultures such as Deep Ecology and Environmental Humanities in the development of methodologies of design that study nature, integrates their strategies to our own species, and considers life of other species as important as human life, and is only form the studies of ontological design that material and immaterial dimensions can be analyzed and imbued with structures that already exist in other fields.

Keywords: design, antropocentrism, ecocentrism, ontological design

Procedia PDF Downloads 159
448 Clinical Staff Perceptions of the Quality of End-of-Life Care in an Acute Private Hospital: A Mixed Methods Design

Authors: Rosemary Saunders, Courtney Glass, Karla Seaman, Karen Gullick, Julie Andrew, Anne Wilkinson, Ashwini Davray

Abstract:

Current literature demonstrates that most Australians receive end-of-life care in a hospital setting, despite most hoping to die within their own home. The necessity for high quality end-of-life care has been emphasised by the Australian Commission on Safety and Quality in Health Care and the National Safety and Quality in Health Services Standards depict the requirement for comprehensive care at the end of life (Action 5.20), reinforcing the obligation for continual organisational assessment to determine if these standards are suitably achieved. Limited research exploring clinical staff perspectives of end-of-life care delivery has been conducted within an Australian private health context. This study aimed to investigate clinical staff member perceptions of end-of-life care delivery at a private hospital in Western Australia. The study comprised of a multi-faceted mixed-methods methodology, part of a larger study. Data was obtained from clinical staff utilising surveys and focus groups. A total of 133 questionnaires were completed by clinical staff, including registered nurses (61.4%), enrolled nurses (22.7%), allied health professionals (9.9%), non-palliative care consultants (3.8%) and junior doctors (2.2%). A total of 14.7% of respondents were palliative care ward staff members. Additionally, seven staff focus groups were conducted with physicians (n=3), nurses (n=26) and allied health professionals including social workers (n=1), dietitians (n=2), physiotherapists (n=5) and speech pathologists (n=3). Key findings from the surveys highlighted that the majority of staff agreed it was part of their role to talk to doctors about the care of patients who they thought may be dying, and recognised the importance of communication, appropriate training and support for clinical staff to provide quality end-of-life care. Thematic analysis of the qualitative data generated three key themes: creating the setting which highlighted the importance of adequate resourcing and conducive physical environments for end-of-life care and to support staff and families; planning and care delivery which emphasised the necessity for collaboration between staff, families and patients to develop care plans and treatment directives; and collaborating in end-of-life care, with effective communication and teamwork leading to achievable care delivery expectations. These findings contribute to health professionals better understanding of end-of-life care provision and the importance of collaborating with patients and families in care delivery. It is crucial that health care providers implement strategies to overcome gaps in care, so quality end-of-life care is provided. Findings from this study have been translated into practice, with the development and implementation of resources, training opportunities, support networks and guidelines for the delivery of quality end-of-life care.

Keywords: clinical staff, end-of-life care, mixed-methods, private hospital.

Procedia PDF Downloads 158
447 The Constitutional Rights of a Child to a Clean and Healthy Environment: A Case Study in the Vaal Triangle Region

Authors: Christiena Van Der Bank, Marjone Van Der Bank, Ronelle Prinsloo

Abstract:

The constitutional right to a healthy environment and the constitutional duty imposed on the state actively to protect the environment fulfill the specific duties to prevent pollution and ecological degradation and to promote conservation. The aim of this paper is to draw attention to the relationship between child rights and the environment. The focus is to analyse government’s responses as mandated with section 24 of the Bill of Rights for ensuring the right to a clean and healthy environment. The principle of sustainability of the environment encompasses the notion of equity and the harm to the environment affects the present as well as future generations. Section 24 obliges the state to ensure that the legacy of future generations is protected, an obligation that has been said to be part of the common law. The environment is an elusive and wide concept that can mean different things to different people depending on the context in which it is used for example clean drinking water or safe food. An extensive interpretation of the term environment would include almost everything that may positively or negatively influence the quality of human life. The analysis will include assessing policy measures, legislation, budgetary measures and other measures taken by the government in order to progressively meet its constitutional obligation. The opportunity of the child to grow up in a healthy and safe environment is extremely unjustly distributed. Without a realignment of political, legal and economic conditions this situation will not fundamentally change. South Africa as a developing country that needs to meet the demand of social transformation and economic growth whilst at the same time expediting its ability to compete in global markets, the country will inevitably embark on developmental programmes as a measure for sustainable development. The courts would have to inquire into the reasonableness of those measures. Environmental threats to children’s rights must be identified, taking into account children’s specific needs and vulnerabilities, their dependence and marginalisation. Obligations of states and violations of rights must be made more visible to the general public.

Keywords: environment, children rights, pollution, healthy, violation

Procedia PDF Downloads 176
446 Artificial Neural Network and Satellite Derived Chlorophyll Indices for Estimation of Wheat Chlorophyll Content under Rainfed Condition

Authors: Muhammad Naveed Tahir, Wang Yingkuan, Huang Wenjiang, Raheel Osman

Abstract:

Numerous models used in prediction and decision-making process but most of them are linear in natural environment, and linear models reach their limitations with non-linearity in data. Therefore accurate estimation is difficult. Artificial Neural Networks (ANN) found extensive acceptance to address the modeling of the complex real world for the non-linear environment. ANN’s have more general and flexible functional forms than traditional statistical methods can effectively deal with. The link between information technology and agriculture will become more firm in the near future. Monitoring crop biophysical properties non-destructively can provide a rapid and accurate understanding of its response to various environmental influences. Crop chlorophyll content is an important indicator of crop health and therefore the estimation of crop yield. In recent years, remote sensing has been accepted as a robust tool for site-specific management by detecting crop parameters at both local and large scales. The present research combined the ANN model with satellite-derived chlorophyll indices from LANDSAT 8 imagery for predicting real-time wheat chlorophyll estimation. The cloud-free scenes of LANDSAT 8 were acquired (Feb-March 2016-17) at the same time when ground-truthing campaign was performed for chlorophyll estimation by using SPAD-502. Different vegetation indices were derived from LANDSAT 8 imagery using ERADAS Imagine (v.2014) software for chlorophyll determination. The vegetation indices were including Normalized Difference Vegetation Index (NDVI), Green Normalized Difference Vegetation Index (GNDVI), Chlorophyll Absorbed Ratio Index (CARI), Modified Chlorophyll Absorbed Ratio Index (MCARI) and Transformed Chlorophyll Absorbed Ratio index (TCARI). For ANN modeling, MATLAB and SPSS (ANN) tools were used. Multilayer Perceptron (MLP) in MATLAB provided very satisfactory results. For training purpose of MLP 61.7% of the data, for validation purpose 28.3% of data and rest 10% of data were used to evaluate and validate the ANN model results. For error evaluation, sum of squares error and relative error were used. ANN model summery showed that sum of squares error of 10.786, the average overall relative error was .099. The MCARI and NDVI were revealed to be more sensitive indices for assessing wheat chlorophyll content with the highest coefficient of determination R²=0.93 and 0.90 respectively. The results suggested that use of high spatial resolution satellite imagery for the retrieval of crop chlorophyll content by using ANN model provides accurate, reliable assessment of crop health status at a larger scale which can help in managing crop nutrition requirement in real time.

Keywords: ANN, chlorophyll content, chlorophyll indices, satellite images, wheat

Procedia PDF Downloads 148
445 The Impact of Human Intervention on Net Primary Productivity for the South-Central Zone of Chile

Authors: Yannay Casas-Ledon, Cinthya A. Andrade, Camila E. Salazar, Mauricio Aguayo

Abstract:

The sustainable management of available natural resources is a crucial question for policy-makers, economists, and the research community. Among several, land constitutes one of the most critical resources, which is being intensively appropriated by human activities producing ecological stresses and reducing ecosystem services. In this context, net primary production (NPP) has been considered as a feasible proxy indicator for estimating the impacts of human interventions on land-uses intensity. Accordingly, the human appropriation of NPP (HANPP) was calculated for the south-central regions of Chile between 2007 and 2014. The HANPP was defined as the difference between the potential NPP of the naturally produced vegetation (NPP0, i.e., the vegetation that would exist without any human interferences) and the NPP remaining in the field after harvest (NPPeco), expressed in gC/m² yr. Other NPP flows taken into account in HANPP estimation were the harvested (NPPh) and the losses of NPP through land conversion (NPPluc). The ArcGIS 10.4 software was used for assessing the spatial and temporal HANPP changes. The differentiation of HANPP as % of NPP0 was estimated by each landcover type taken in 2007 and 2014 as the reference years. The spatial results depicted a negative impact on land use efficiency during 2007 and 2014, showing negative HANPP changes for the whole region. The harvest and biomass losses through land conversion components are the leading causes of loss of land-use efficiency. Furthermore, the study depicted higher HANPP in 2014 than in 2007, representing 50% of NPP0 for all landcover classes concerning 2007. This performance was mainly related to the higher volume of harvested biomass for agriculture. In consequence, the cropland depicted the high HANPP followed by plantation. This performance highlights the strong positive correlation between the economic activities developed into the region. This finding constitutes the base for a better understanding of the main driving force influencing biomass productivity and a powerful metric for supporting the sustainable management of land use.

Keywords: human appropriation, land-use changes, land-use impact, net primary productivity

Procedia PDF Downloads 145
444 Improving Exchange Rate Forecasting Accuracy Using Ensemble Learning Techniques: A Comparative Study

Authors: Gokcen Ogruk-Maz, Sinan Yildirim

Abstract:

Introduction: Exchange rate forecasting is pivotal for informed financial decision-making, encompassing risk management, investment strategies, and international trade planning. However, traditional forecasting models often fail to capture the complexity and volatility of currency markets. This study explores the potential of ensemble learning techniques such as Random Forest, Gradient Boosting, and AdaBoost to enhance the accuracy and robustness of exchange rate predictions. Research Objectives The primary objective is to evaluate the performance of ensemble methods in comparison to traditional econometric models such as Uncovered Interest Rate Parity, Purchasing Power Parity, and Monetary Models. By integrating advanced machine learning techniques with fundamental macroeconomic indicators, this research seeks to identify optimal approaches for predicting exchange rate movements across major currency pairs. Methodology: Using historical exchange rate data and economic indicators such as interest rates, inflation, money supply, and GDP, the study develops forecasting models leveraging ensemble techniques. Comparative analysis is performed against traditional models and hybrid approaches incorporating Facebook Prophet, Artificial Neural Networks, and XGBoost. The models are evaluated using statistical metrics like Mean Squared Error, Theil Ratio, and Diebold-Mariano tests across five currency pairs (JPY to USD, AUD to USD, CAD to USD, GBP to USD, and NZD to USD). Preliminary Results: Results indicate that ensemble learning models consistently outperform traditional methods in predictive accuracy. XGBoost shows the strongest performance among the techniques evaluated, achieving significant improvements in forecast precision with consistently low p-values and Theil Ratios. Hybrid models integrating macroeconomic fundamentals into machine learning frameworks further enhance predictive accuracy. Discussion: The findings show the potential of ensemble methods to address the limitations of traditional models by capturing non-linear relationships and complex dynamics in exchange rate movements. While Random Forest and Gradient Boosting are effective, the superior performance of XGBoost suggests that its capacity for handling sparse and irregular data offers a distinct advantage in financial forecasting. Conclusion and Implications: This research demonstrates that ensemble learning techniques, particularly when combined with traditional macroeconomic fundamentals, provide a robust framework for improving exchange rate forecasting. The study offers actionable insights for financial practitioners and policymakers, emphasizing the value of integrating machine learning approaches into predictive modeling for monetary economics.

Keywords: exchange rate forecasting, ensemble learning, financial modeling, machine learning, monetary economics, XGBoost

Procedia PDF Downloads 9
443 Change of Education Business in the Age of 5G

Authors: Heikki Ruohomaa, Vesa Salminen

Abstract:

Regions are facing huge competition to attract companies, businesses, inhabitants, students, etc. This way to improve living and business environment, which is rapidly changing due to digitalization. On the other hand, from the industry's point of view, the availability of a skilled labor force and an innovative environment are crucial factors. In this context, qualified staff has been seen to utilize the opportunities of digitalization and respond to the needs of future skills. World Manufacturing Forum has stated in the year 2019- report that in next five years, 40% of workers have to change their core competencies. Through digital transformation, new technologies like cloud, mobile, big data, 5G- infrastructure, platform- technology, data- analysis, and social networks with increasing intelligence and automation, enterprises can capitalize on new opportunities and optimize existing operations to achieve significant business improvement. Digitalization will be an important part of the everyday life of citizens and present in the working day of the average citizen and employee in the future. For that reason, the education system and education programs on all levels of education from diaper age to doctorate have been directed to fulfill this ecosystem strategy. Goal: The Fourth Industrial Revolution will bring unprecedented change to societies, education organizations and business environments. This article aims to identify how education, education content, the way education has proceeded, and overall whole the education business is changing. Most important is how we should respond to this inevitable co- evolution. Methodology: The study aims to verify how the learning process is boosted by new digital content, new learning software and tools, and customer-oriented learning environments. The change of education programs and individual education modules can be supported by applied research projects. You can use them in making proof- of- the concept of new technology, new ways to teach and train, and through the experiences gathered change education content, way to educate and finally education business as a whole. Major findings: Applied research projects can prove the concept- phases on real environment field labs to test technology opportunities and new tools for training purposes. Customer-oriented applied research projects are also excellent for students to make assignments and use new knowledge and content and teachers to test new tools and create new ways to educate. New content and problem-based learning are used in future education modules. This article introduces some case study experiences on customer-oriented digital transformation projects and how gathered knowledge on new digital content and a new way to educate has influenced education. The case study is related to experiences of research projects, customer-oriented field labs/learning environments and education programs of Häme University of Applied Sciences.

Keywords: education process, digitalization content, digital tools for education, learning environments, transdisciplinary co-operation

Procedia PDF Downloads 179
442 Evaluation of the Performance Measures of Two-Lane Roundabout and Turbo Roundabout with Varying Truck Percentages

Authors: Evangelos Kaisar, Anika Tabassum, Taraneh Ardalan, Majed Al-Ghandour

Abstract:

The economy of any country is dependent on its ability to accommodate the movement and delivery of goods. The demand for goods movement and services increases truck traffic on highways and inside the cities. The livability of most cities is directly affected by the congestion and environmental impacts of trucks, which are the backbone of the urban freight system. Better operation of heavy vehicles on highways and arterials could lead to the network’s efficiency and reliability. In many cases, roundabouts can respond better than at-level intersections to enable traffic operations with increased safety for both cars and heavy vehicles. Recently emerged, the concept of turbo-roundabout is a viable alternative to the two-lane roundabout aiming to improve traffic efficiency. The primary objective of this study is to evaluate the operation and performance level of an at-grade intersection, a conventional two-lane roundabout, and a basic turbo roundabout for freight movements. To analyze and evaluate the performances of the signalized intersections and the roundabouts, micro simulation models were developed PTV VISSIM. The networks chosen for this analysis in this study are to experiment and evaluate changes in the performance of the movement of vehicles with different geometric and flow scenarios. There are several scenarios that were examined when attempting to assess the impacts of various geometric designs on vehicle movements. The overall traffic efficiency depends on the geometric layout of the intersections, which consists of traffic congestion rate, hourly volume, frequency of heavy vehicles, type of road, and the ratio of major-street versus side-street traffic. The traffic performance was determined by evaluating the delay time, number of stops, and queue length of each intersection for varying truck percentages. The results indicate that turbo-roundabouts can replace signalized intersections and two-lane roundabouts only when the traffic demand is low, even with high truck volume. More specifically, it is clear that two-lane roundabouts are seen to have shorter queue lengths compared to signalized intersections and turbo-roundabouts. For instance, considering the scenario where the volume is highest, and the truck movement and left turn movement are maximum, the signalized intersection has 3 times, and the turbo-roundabout has 5 times longer queue length than a two-lane roundabout in major roads. Similarly, on minor roads, signalized intersections and turbo-roundabouts have 11 times longer queue lengths than two-lane roundabouts for the same scenario. As explained from all the developed scenarios, while the traffic demand lowers, the queue lengths of turbo-roundabouts shorten. This proves that turbo roundabouts perform well for low and medium traffic demand. The results indicate that turbo-roundabouts can replace signalized intersections and two-lane roundabouts only when the traffic demand is low, even with high truck volume. Finally, this study provides recommendations on the conditions under which different intersections perform better than each other.

Keywords: At-grade intersection, simulation, turbo-roundabout, two-lane roundabout

Procedia PDF Downloads 152
441 Spatial Mapping and Change Detection of a Coastal Woodland Mangrove Habitat in Fiji

Authors: Ashneel Ajay Singh, Anish Maharaj, Havish Naidu, Michelle Kumar

Abstract:

Mangrove patches are the foundation species located in the estuarine land areas. These patches provide a nursery, food source and protection for numerous aquatic, intertidal and well as land-based organisms. Mangroves also help in coastal protection, maintain water clarity and are one of the biggest sinks for blue carbon sequestration. In the Pacific Island countries, numerous coastal communities have a heavy socioeconomic dependence on coastal resources and mangroves play a key ecological and economical role in structuring the availability of these resources. Fiji has a large mangrove patch located in the Votua area of the Ba province. Globally, mangrove population continues to decline with the changes in climatic conditions and anthropogenic activities. Baseline information through wetland maps and time series change are essential references for development of effective mangrove management plans. These maps reveal the status of the resource and the effects arising from anthropogenic activities and climate change. In this study, we used remote sensing and GIS tools for mapping and temporal change detection over a period of >20 years in Votua, Fiji using Landsat imagery. Landsat program started in 1972 initially as Earth Resources Technology Satellite. Since then it has acquired millions of images of Earth. This archive allows mapping of temporal changes in mangrove forests. Mangrove plants consisted of the species Rhizophora stylosa, Rhizophora samoensis, Bruguiera gymnorrhiza, Lumnitzera littorea, Heritiera littoralis, Excoecaria agallocha and Xylocarpus granatum. Change detection analysis revealed significant reduction in the mangrove patch over the years. This information serves as a baseline for the development and implementation of effective management plans for one of Fiji’s biggest mangrove patches.

Keywords: climate change, GIS, Landsat, mangrove, temporal change

Procedia PDF Downloads 181
440 The Lived Experiences and Coping Strategies of Women with Attention Deficit and Hyperactivity Disorder (ADHD)

Authors: Oli Sophie Meredith, Jacquelyn Osborne, Sarah Verdon, Jane Frawley

Abstract:

PROJECT OVERVIEW AND BACKGROUND: Over one million Australians are affected by ADHD at an economic and social cost of over $20 billion per annum. Despite health outcomes being significantly worse compared with men, women have historically been overlooked in ADHD diagnosis and treatment. While research suggests physical activity and other non-prescription options can help with ADHD symptoms, the frontline response to ADHD remains expensive stimulant medications that can have adverse side effects. By interviewing women with ADHD, this research will examine women’s self-directed approaches to managing symptoms, including alternatives to prescription medications. It will investigate barriers and affordances to potentially helpful approaches and identify any concerning strategies pursued in lieu of diagnosis. SIGNIFICANCE AND INNOVATION: Despite the economic and societal impact of ADHD on women, research investigating how women manage their symptoms is scant. This project is significant because although women’s ADHD symptoms are markedly different to those of men, mainstream treatment has been based on the experiences of men. Further, it is thought that in developing nuanced coping strategies, women may have masked their symptoms. Thus, this project will highlight strategies which women deem effective in ‘thriving’ rather than just ‘hiding’. By investigating the health service use, self-care and physical activity of women with ADHD, this research aligns with a priority research areas as identified by the November 2023 senate ADHD inquiry report. APPROACH AND METHODS: Semi-structured interviews will be conducted with up to 20 women with ADHD. Interviews will be conducted in person and online to capture experience across rural and metropolitan Australia. Participants will be recruited in partnership with the peak representative body, ADHD Australia. The research will use an intersectional framework, and data will be analysed thematically. This project is led by an interdisciplinary and cross-institutional team of women with ADHD. Reflexive interviewing skills will be employed to help interviewees feel more comfortable disclosing their experiences, especially where they share common ground ENGAGEMENT, IMPACT AND BENEFIT: This research will benefit women with ADHD by increasing knowledge of strategies and alternative treatments to prescription medications, reducing the social and economic burden of ADHD on Australia and on individuals. It will also benefit women by identifying risks involved with some self-directed approaches in lieu of medical advice. The project has an accessible impact plan to directly benefit end-users, which includes the development of a podcast and a PDF resource translating findings. The resources will reach a wide audience through ADHD Australia’s extensive national networks. We will collaborate with Charles Sturt’s Accessibility and Inclusion Division of Safety, Security and Well-being to create a targeted resource for students with ADHD.

Keywords: ADHD, women's health, self-directed strategies, health service use, physical activity, public health

Procedia PDF Downloads 76
439 [Keynote Talk]: New Generations and Employment: An Exploratory Study about Tensions between the Psycho-Social Characteristics of the Generation Z and Expectations and Actions of Organizational Structures Related with Employment (CABA, 2016)

Authors: Esteban Maioli

Abstract:

Generational studies have an important research tradition in social and human sciences. On the one hand, the speed of social change in the context of globalization imposes the need to research the transformations are identified both the subjectivity of the agents involved and its inclusion in the institutional matrix, specifically employment. Generation Z, (generally considered as the population group whose birth occurs after 1995) have unique psycho-social characteristics. Gen Z is characterized by a different set of values, beliefs, attitudes and ambitions that impact in their concrete action in organizational structures. On the other hand, managers often have to deal with generational differences in the workplace. Organizations have members who belong to different generations; they had never before faced the challenge of having such a diverse group of members. The members of each historical generation are characterized by a different set of values, beliefs, attitudes and ambitions that are manifest in their concrete action in organizational structures. Gen Z it’s the only one who can fully be considered "global," while its members were born in the consolidated context of globalization. Some salient features of the Generation Z can be summarized as follows. They’re the first fully born into a digital world. Social networks and technology are integrated into their lives. They are concerned about the challenges of the modern world (poverty, inequality, climate change, among others). They are self-expressive, more liberal and open to change. They often bore easily, with short attention spans. They do not like routine tasks. They want to achieve a good life-work balance, and they are interested in a flexible work environment, as opposed to traditional work schedule. They are critical thinkers, who come with innovative and creative ideas to help. Research design considered methodological triangulation. Data was collected with two techniques: a self-administered survey with multiple choice questions and attitudinal scales applied over a non-probabilistic sample by reasoned decision. According to the multi-method strategy, also it was conducted in-depth interviews. Organizations constantly face new challenges. One of the biggest ones is to learn to manage a multi-generational scope of work. While Gen Z has not yet been fully incorporated (expected to do so in five years or so), many organizations have already begun to implement a series of changes in its recruitment and development. The main obstacle to retaining young talent is the gap between the expectations of iGen applicants and what companies offer. Members of the iGen expect not only a good salary and job stability but also a clear career plan. Generation Z needs to have immediate feedback on their tasks. However, many organizations have yet to improve both motivation and monitoring practices. It is essential for companies to take a review of organizational practices anchored in the culture of the organization.

Keywords: employment, expectations, generation Z, organizational culture, organizations, psycho-social characteristics

Procedia PDF Downloads 204
438 Bioinformatic Prediction of Hub Genes by Analysis of Signaling Pathways, Transcriptional Regulatory Networks and DNA Methylation Pattern in Colon Cancer

Authors: Ankan Roy, Niharika, Samir Kumar Patra

Abstract:

Anomalous nexus of complex topological assemblies and spatiotemporal epigenetic choreography at chromosomal territory may forms the most sophisticated regulatory layer of gene expression in cancer. Colon cancer is one of the leading malignant neoplasms of the lower gastrointestinal tract worldwide. There is still a paucity of information about the complex molecular mechanisms of colonic cancerogenesis. Bioinformatics prediction and analysis helps to identify essential genes and significant pathways for monitoring and conquering this deadly disease. The present study investigates and explores potential hub genes as biomarkers and effective therapeutic targets for colon cancer treatment. Colon cancer patient sample containing gene expression profile datasets, such as GSE44076, GSE20916, and GSE37364 were downloaded from Gene Expression Omnibus (GEO) database and thoroughly screened using the GEO2R tool and Funrich software to find out common 2 differentially expressed genes (DEGs). Other approaches, including Gene Ontology (GO) and KEGG pathway analysis, Protein-Protein Interaction (PPI) network construction and hub gene investigation, Overall Survival (OS) analysis, gene correlation analysis, methylation pattern analysis, and hub gene-Transcription factors regulatory network construction, were performed and validated using various bioinformatics tool. Initially, we identified 166 DEGs, including 68 up-regulated and 98 down-regulated genes. Up-regulated genes are mainly associated with the Cytokine-cytokine receptor interaction, IL17 signaling pathway, ECM-receptor interaction, Focal adhesion and PI3K-Akt pathway. Downregulated genes are enriched in metabolic pathways, retinol metabolism, Steroid hormone biosynthesis, and bile secretion. From the protein-protein interaction network, thirty hub genes with high connectivity are selected using the MCODE and cytoHubba plugin. Survival analysis, expression validation, correlation analysis, and methylation pattern analysis were further verified using TCGA data. Finally, we predicted COL1A1, COL1A2, COL4A1, SPP1, SPARC, and THBS2 as potential master regulators in colonic cancerogenesis. Moreover, our experimental data highlights that disruption of lipid raft and RAS/MAPK signaling cascade affects this gene hub at mRNA level. We identified COL1A1, COL1A2, COL4A1, SPP1, SPARC, and THBS2 as determinant hub genes in colon cancer progression. They can be considered as biomarkers for diagnosis and promising therapeutic targets in colon cancer treatment. Additionally, our experimental data advertise that signaling pathway act as connecting link between membrane hub and gene hub.

Keywords: hub genes, colon cancer, DNA methylation, epigenetic engineering, bioinformatic predictions

Procedia PDF Downloads 134
437 An in silico Approach for Exploring the Intercellular Communication in Cancer Cells

Authors: M. Cardenas-Garcia, P. P. Gonzalez-Perez

Abstract:

Intercellular communication is a necessary condition for cellular functions and it allows a group of cells to survive as a population. Throughout this interaction, the cells work in a coordinated and collaborative way which facilitates their survival. In the case of cancerous cells, these take advantage of intercellular communication to preserve their malignancy, since through these physical unions they can send signs of malignancy. The Wnt/β-catenin signaling pathway plays an important role in the formation of intercellular communications, being also involved in a large number of cellular processes such as proliferation, differentiation, adhesion, cell survival, and cell death. The modeling and simulation of cellular signaling systems have found valuable support in a wide range of modeling approaches, which cover a wide spectrum ranging from mathematical models; e.g., ordinary differential equations, statistical methods, and numerical methods– to computational models; e.g., process algebra for modeling behavior and variation in molecular systems. Based on these models, different simulation tools have been developed from mathematical ones to computational ones. Regarding cellular and molecular processes in cancer, its study has also found a valuable support in different simulation tools that, covering a spectrum as mentioned above, have allowed the in silico experimentation of this phenomenon at the cellular and molecular level. In this work, we simulate and explore the complex interaction patterns of intercellular communication in cancer cells using the Cellulat bioinformatics tool, a computational simulation tool developed by us and motivated by two key elements: 1) a biochemically inspired model of self-organizing coordination in tuple spaces, and 2) the Gillespie’s algorithm, a stochastic simulation algorithm typically used to mimic systems of chemical/biochemical reactions in an efficient and accurate way. The main idea behind the Cellulat simulation tool is to provide an in silico experimentation environment that complements and guides in vitro experimentation in intra and intercellular signaling networks. Unlike most of the cell signaling simulation tools, such as E-Cell, BetaWB and Cell Illustrator which provides abstractions to model only intracellular behavior, Cellulat is appropriate for modeling both intracellular signaling and intercellular communication, providing the abstractions required to model –and as a result, simulate– the interaction mechanisms that involve two or more cells, that is essential in the scenario discussed in this work. During the development of this work we made evident the application of our computational simulation tool (Cellulat) for the modeling and simulation of intercellular communication between normal and cancerous cells, and in this way, propose key molecules that may prevent the arrival of malignant signals to the cells that surround the tumor cells. In this manner, we could identify the significant role that has the Wnt/β-catenin signaling pathway in cellular communication, and therefore, in the dissemination of cancer cells. We verified, using in silico experiments, how the inhibition of this signaling pathway prevents that the cells that surround a cancerous cell are transformed.

Keywords: cancer cells, in silico approach, intercellular communication, key molecules, modeling and simulation

Procedia PDF Downloads 252
436 Geovisualisation for Defense Based on a Deep Learning Monocular Depth Reconstruction Approach

Authors: Daniel R. dos Santos, Mateus S. Maldonado, Estevão J. R. Batista

Abstract:

The military commanders increasingly dependent on spatial awareness, as knowing where enemy are, understanding how war battle scenarios change over time, and visualizing these trends in ways that offer insights for decision-making. Thanks to advancements in geospatial technologies and artificial intelligence algorithms, the commanders are now able to modernize military operations on a universal scale. Thus, geovisualisation has become an essential asset in the defense sector. It has become indispensable for better decisionmaking in dynamic/temporal scenarios, operation planning and management for the war field, situational awareness, effective planning, monitoring, and others. For example, a 3D visualization of war field data contributes to intelligence analysis, evaluation of postmission outcomes, and creation of predictive models to enhance decision-making and strategic planning capabilities. However, old-school visualization methods are slow, expensive, and unscalable. Despite modern technologies in generating 3D point clouds, such as LIDAR and stereo sensors, monocular depth values based on deep learning can offer a faster and more detailed view of the environment, transforming single images into visual information for valuable insights. We propose a dedicated monocular depth reconstruction approach via deep learning techniques for 3D geovisualisation of satellite images. It introduces scalability in terrain reconstruction and data visualization. First, a dataset with more than 7,000 satellite images and associated digital elevation model (DEM) is created. It is based on high resolution optical and radar imageries collected from Planet and Copernicus, on which we fuse highresolution topographic data obtained using technologies such as LiDAR and the associated geographic coordinates. Second, we developed an imagery-DEM fusion strategy that combine feature maps from two encoder-decoder networks. One network is trained with radar and optical bands, while the other is trained with DEM features to compute dense 3D depth. Finally, we constructed a benchmark with sparse depth annotations to facilitate future research. To demonstrate the proposed method's versatility, we evaluated its performance on no annotated satellite images and implemented an enclosed environment useful for Geovisualisation applications. The algorithms were developed in Python 3.0, employing open-source computing libraries, i.e., Open3D, TensorFlow, and Pythorch3D. The proposed method provides fast and accurate decision-making with GIS for localization of troops, position of the enemy, terrain and climate conditions. This analysis enhances situational consciousness, enabling commanders to fine-tune the strategies and distribute the resources proficiently.

Keywords: depth, deep learning, geovisualisation, satellite images

Procedia PDF Downloads 16
435 Classification of ECG Signal Based on Mixture of Linear and Non-Linear Features

Authors: Mohammad Karimi Moridani, Mohammad Abdi Zadeh, Zahra Shahiazar Mazraeh

Abstract:

In recent years, the use of intelligent systems in biomedical engineering has increased dramatically, especially in the diagnosis of various diseases. Also, due to the relatively simple recording of the electrocardiogram signal (ECG), this signal is a good tool to show the function of the heart and diseases associated with it. The aim of this paper is to design an intelligent system for automatically detecting a normal electrocardiogram signal from abnormal one. Using this diagnostic system, it is possible to identify a person's heart condition in a very short time and with high accuracy. The data used in this article are from the Physionet database, available in 2016 for use by researchers to provide the best method for detecting normal signals from abnormalities. Data is of both genders and the data recording time varies between several seconds to several minutes. All data is also labeled normal or abnormal. Due to the low positional accuracy and ECG signal time limit and the similarity of the signal in some diseases with the normal signal, the heart rate variability (HRV) signal was used. Measuring and analyzing the heart rate variability with time to evaluate the activity of the heart and differentiating different types of heart failure from one another is of interest to the experts. In the preprocessing stage, after noise cancelation by the adaptive Kalman filter and extracting the R wave by the Pan and Tampkinz algorithm, R-R intervals were extracted and the HRV signal was generated. In the process of processing this paper, a new idea was presented that, in addition to using the statistical characteristics of the signal to create a return map and extraction of nonlinear characteristics of the HRV signal due to the nonlinear nature of the signal. Finally, the artificial neural networks widely used in the field of ECG signal processing as well as distinctive features were used to classify the normal signals from abnormal ones. To evaluate the efficiency of proposed classifiers in this paper, the area under curve ROC was used. The results of the simulation in the MATLAB environment showed that the AUC of the MLP and SVM neural network was 0.893 and 0.947, respectively. As well as, the results of the proposed algorithm in this paper indicated that the more use of nonlinear characteristics in normal signal classification of the patient showed better performance. Today, research is aimed at quantitatively analyzing the linear and non-linear or descriptive and random nature of the heart rate variability signal, because it has been shown that the amount of these properties can be used to indicate the health status of the individual's heart. The study of nonlinear behavior and dynamics of the heart's neural control system in the short and long-term provides new information on how the cardiovascular system functions, and has led to the development of research in this field. Given that the ECG signal contains important information and is one of the common tools used by physicians to diagnose heart disease, but due to the limited accuracy of time and the fact that some information about this signal is hidden from the viewpoint of physicians, the design of the intelligent system proposed in this paper can help physicians with greater speed and accuracy in the diagnosis of normal and patient individuals and can be used as a complementary system in the treatment centers.

Keywords: neart rate variability, signal processing, linear and non-linear features, classification methods, ROC Curve

Procedia PDF Downloads 265