Search results for: continuous machine condition monitoring
100 Observing Teaching Practices Through the Lenses of Self-Regulated Learning: A Study Within the String Instrument Individual Context
Authors: Marija Mihajlovic Pereira
Abstract:
Teaching and learning a musical instrument is challenging for both teachers and students. Teachers generally use diverse strategies to resolve students' particular issues in a one-to-one context. Considering individual sessions as a supportive educational context, the teacher can play a decisive role in stimulating and promoting self-regulated learning strategies, especially with beginning learners. The teachers who promote self-controlling behaviors, strategic monitoring, and regulation of actions toward goals could expect their students to practice more qualitatively and consciously. When encouraged to adopt self-regulation habits, students' could benefit from greater productivity on a longer path. Founded on Bary Zimmerman's cyclical model that comprehends three phases - forethought, performance, and self-reflection, this work aims to articulate self-regulated and music learning. Self-regulated learning appeals to the individual's attitude in planning, controlling, and reflecting on their performance. Furthermore, this study aimed to present an observation grid for perceiving teaching instructions that encourage students' controlling cognitive behaviors in light of the belief that conscious promotion of self-regulation may motivate strategic actions toward goals in musical performance. The participants, two teachers, and two students have been involved in the social inclusion project in Lisbon (Portugal). The author and one independent inter-observer analyzed six video-recorded string instrument lessons. The data correspond to three sessions per teacher lectured to one (different) student. Violin (f) and violoncello (m) teachers hold a Master's degree in music education and approximately five years of experience. In their second year of learning an instrument, students have acquired reasonable skills in musical reading, posture, and sound quality until then. The students also manifest positive learning behaviors, interest in learning a musical instrument, although their study habits are still inconsistent. According to the grid's four categories (parent codes), in-class rehearsal frames were coded using MaxQda software, version 20, according to the grid's four categories (parent codes): self-regulated learning, teaching verbalizations, teaching strategies, and students' in-class performance. As a result, selected rehearsal frames qualitatively describe teaching instructions that might promote students' body and hearing awareness, such as "close the eyes while playing" or "sing to internalize the pitch." Another analysis type, coding the short video events according to the observation grid's subcategories (child codes), made it possible to perceive the time teachers dedicate to specific verbal or non-verbal strategies. Furthermore, a coding overlay analysis indicated that teachers tend to stimulate. (i) Forethought – explain tasks, offer feedback and ensure that students identify a goal, (ii) Performance – teach study strategies and encourage students to sing and use vocal abilities to ensure inner audition, (iii) Self-reflection – frequent inquiring and encouraging the student to verbalize their perception of performance. Although developed in the context of individual string instrument lessons, this classroom observation grid brings together essential variables in a one-to-one lesson. It may find utility in a broader context of music education due to the possibility to organize, observe and evaluate teaching practices. Besides that, this study contributes to cognitive development by suggesting a practical approach to fostering self-regulated learning.Keywords: music education, observation grid, self-regulated learning, string instruments, teaching practices
Procedia PDF Downloads 9899 Role of Dedicated Medical Social Worker in Fund Mobilisation and Economic Evaluation in Ovarian Cancer: Experience from a Tertiary Referral Centre in Eastern India
Authors: Aparajita Bhattacharya, Mousumi Dutta, Zakir Husain, Dionne Sequeira, Asima Mukhopadhyay
Abstract:
Background: Tata Medical Centre (TMC), Kolkata is a major cancer referral centre in Eastern India and neighbouring countries providing state of the art facilities; however, it is a non-profit organisation with patients requiring to pay at subsidised rates. Although a system for social assessment and applying for governmental/ non-governmental (NGO) funds is in place, access is challenging. Amongst gynaecological cancers (GC), ovarian cancer (OC) is associated with the highest treatment cost; majority of which is required at the beginning when complex surgery is performed and funding arrangements cannot be made in time. We therefore appointed a dedicated Medical Social Worker (MSW) in 2016, supported by NGO for GC patients in order to assist patients/family members to access/avail these funds more readily and assist in economic evaluation for both direct and opportunity costs. Objectives: To reflect on our experience and challenges in collecting data on economic evaluation of cancer patients and compare success rates in achieving fund mobilization after introduction of MSW. Methods: A Retrospective survey. Patients with OC and their relatives were seen by the MSW during the initial outpatients department visit and followed though till discharge from the hospital and during follow-up visits. Assistance was provided in preparing the essential documents/paperwork/contacts for the funding agencies including both governmental (Chief-Minister/Prime-Minister/President) and NGO sources. In addition, a detailed questionnaire was filled up for economic assessment of direct/opportunity costs during the entire treatment and 12 months follow up period which forms a part of the study called HEPTROC (Health economic evaluation of primary treatment for ovarian cancer) developed in collaboration with economics departments of Universities. Results: In 2015, 102 patients were operated for OC; only 16 patients (15.68 %) had availed funding of a total sum of INR 1640000 through the hospital system for social assessment. Following challenges were faced by majority of the relatives: 1. Gathering important documents/proper contact details for governmental funding bodies and difficulty in following up the current status 3. Late arrival of funds. In contrast in 2016, 104 OC patients underwent surgery; the direct cost of treatment was significantly higher (median, INR 300000- 400000) compared to other GCs (n=274). 98/104 (94.23%) OC patients could be helped to apply for funds and 90/104(86.56%) patients received funding amounting to a total of INR 10897000. There has been a tenfold increase in funds mobilized in 2016 after the introduction of dedicated MSW in GC. So far, in 2017 (till June), 46/54(85.18%) OC patients applied for funds and 37/54(68.51%) patients have received funding. In a qualitative survey, all patients appreciated the role of the MSW who subsequently became the key worker for patient follow up and the chief portal for patient reported outcome monitoring. Data collection quality for the HEPTROC study was improved when questionnaires were administered by the MSW compared to researchers. Conclusion: Introduction of cancer specific MSW can expedite the availability of funds required for cancer patients and it can positively impact on patient satisfaction and outcome reporting. The economic assessment will influence fund allocation and decision for policymaking in ovarian cancer. Acknowledgement: Jivdaya Foundation Dallas, Texas.Keywords: economic evaluation, funding, medical social worker, ovarian cancer
Procedia PDF Downloads 15498 EEG and DC-Potential Level Сhanges in the Elderly
Authors: Irina Deputat, Anatoly Gribanov, Yuliya Dzhos, Alexandra Nekhoroshkova, Tatyana Yemelianova, Irina Bolshevidtseva, Irina Deryabina, Yana Kereush, Larisa Startseva, Tatyana Bagretsova, Irina Ikonnikova
Abstract:
In the modern world the number of elderly people increases. Preservation of functionality of an organism in the elderly becomes very important now. During aging the higher cortical functions such as feelings, perception, attention, memory, and ideation are gradual decrease. It is expressed in the rate of information processing reduction, volume of random access memory loss, ability to training and storing of new information decrease. Perspective directions in studying of aging neurophysiological parameters are brain imaging: computer electroencephalography, neuroenergy mapping of a brain, and also methods of studying of a neurodynamic brain processes. Research aim – to study features of a brain aging in elderly people by electroencephalogram (EEG) and the DC-potential level. We examined 130 people aged 55 - 74 years that did not have psychiatric disorders and chronic states in a decompensation stage. EEG was recorded with a 128-channel GES-300 system (USA). EEG recordings are collected while the participant sits at rest with their eyes closed for 3 minutes. For a quantitative assessment of EEG we used the spectral analysis. The range was analyzed on delta (0,5–3,5 Hz), a theta - (3,5–7,0 Hz), an alpha 1-(7,0–11,0 Hz) an alpha 2-(11–13,0 Hz), beta1-(13–16,5 Hz) and beta2-(16,5–20 Hz) ranges. In each frequency range spectral power was estimated. The 12-channel hardware-software diagnostic ‘Neuroenergometr-KM’ complex was applied for registration, processing and the analysis of a brain constant potentials level. The DC-potential level registered in monopolar leads. It is revealed that the EEG of elderly people differ in higher rates of spectral power in the range delta (р < 0,01) and a theta - (р < 0,05) rhythms, especially in frontal areas in aging. By results of the comparative analysis it is noted that elderly people 60-64 aged differ in higher values of spectral power alfa-2 range in the left frontal and central areas (р < 0,05) and also higher values beta-1 range in frontal and parieto-occipital areas (р < 0,05). Study of a brain constant potential level distribution revealed increase of total energy consumption on the main areas of a brain. In frontal leads we registered the lowest values of constant potential level. Perhaps it indicates decrease in an energy metabolism in this area and difficulties of executive functions. The comparative analysis of a potential difference on the main assignments testifies to unevenness of a lateralization of a brain functions at elderly people. The results of a potential difference between right and left hemispheres testify to prevalence of the left hemisphere activity. Thus, higher rates of functional activity of a cerebral cortex are peculiar to people of early advanced age (60-64 years) that points to higher reserve opportunities of central nervous system. By 70 years there are age changes of a cerebral power exchange and level of electrogenesis of a brain which reflect deterioration of a condition of homeostatic mechanisms of self-control and the program of processing of the perceptual data current flow.Keywords: brain, DC-potential level, EEG, elderly people
Procedia PDF Downloads 48497 A Modular Solution for Large-Scale Critical Industrial Scheduling Problems with Coupling of Other Optimization Problems
Authors: Ajit Rai, Hamza Deroui, Blandine Vacher, Khwansiri Ninpan, Arthur Aumont, Francesco Vitillo, Robert Plana
Abstract:
Large-scale critical industrial scheduling problems are based on Resource-Constrained Project Scheduling Problems (RCPSP), that necessitate integration with other optimization problems (e.g., vehicle routing, supply chain, or unique industrial ones), thus requiring practical solutions (i.e., modular, computationally efficient with feasible solutions). To the best of our knowledge, the current industrial state of the art is not addressing this holistic problem. We propose an original modular solution that answers the issues exhibited by the delivery of complex projects. With three interlinked entities (project, task, resources) having their constraints, it uses a greedy heuristic with a dynamic cost function for each task with a situational assessment at each time step. It handles large-scale data and can be easily integrated with other optimization problems, already existing industrial tools and unique constraints as required by the use case. The solution has been tested and validated by domain experts on three use cases: outage management in Nuclear Power Plants (NPPs), planning of future NPP maintenance operation, and application in the defense industry on supply chain and factory relocation. In the first use case, the solution, in addition to the resources’ availability and tasks’ logical relationships, also integrates several project-specific constraints for outage management, like, handling of resource incompatibility, updating of tasks priorities, pausing tasks in a specific circumstance, and adjusting dynamic unit of resources. With more than 20,000 tasks and multiple constraints, the solution provides a feasible schedule within 10-15 minutes on a standard computer device. This time-effective simulation corresponds with the nature of the problem and requirements of several scenarios (30-40 simulations) before finalizing the schedules. The second use case is a factory relocation project where production lines must be moved to a new site while ensuring the continuity of their production. This generates the challenge of merging job shop scheduling and the RCPSP with location constraints. Our solution allows the automation of the production tasks while considering the rate expectation. The simulation algorithm manages the use and movement of resources and products to respect a given relocation scenario. The last use case establishes a future maintenance operation in an NPP. The project contains complex and hard constraints, like on Finish-Start precedence relationship (i.e., successor tasks have to start immediately after predecessors while respecting all constraints), shareable coactivity for managing workspaces, and requirements of a specific state of "cyclic" resources (they can have multiple states possible with only one at a time) to perform tasks (can require unique combinations of several cyclic resources). Our solution satisfies the requirement of minimization of the state changes of cyclic resources coupled with the makespan minimization. It offers a solution of 80 cyclic resources with 50 incompatibilities between levels in less than a minute. Conclusively, we propose a fast and feasible modular approach to various industrial scheduling problems that were validated by domain experts and compatible with existing industrial tools. This approach can be further enhanced by the use of machine learning techniques on historically repeated tasks to gain further insights for delay risk mitigation measures.Keywords: deterministic scheduling, optimization coupling, modular scheduling, RCPSP
Procedia PDF Downloads 19896 A Review of Data Visualization Best Practices: Lessons for Open Government Data Portals
Authors: Bahareh Ansari
Abstract:
Background: The Open Government Data (OGD) movement in the last decade has encouraged many government organizations around the world to make their data publicly available to advance democratic processes. But current open data platforms have not yet reached to their full potential in supporting all interested parties. To make the data useful and understandable for everyone, scholars suggested that opening the data should be supplemented by visualization. However, different visualizations of the same information can dramatically change an individual’s cognitive and emotional experience in working with the data. This study reviews the data visualization literature to create a list of the methods empirically tested to enhance users’ performance and experience in working with a visualization tool. This list can be used in evaluating the OGD visualization practices and informing the future open data initiatives. Methods: Previous reviews of visualization literature categorized the visualization outcomes into four categories including recall/memorability, insight/comprehension, engagement, and enjoyment. To identify the papers, a search for these outcomes was conducted in the abstract of the publications of top-tier visualization venues including IEEE Transactions for Visualization and Computer Graphics, Computer Graphics, and proceedings of the CHI Conference on Human Factors in Computing Systems. The search results are complemented with a search in the references of the identified articles, and a search for 'open data visualization,' and 'visualization evaluation' keywords in the IEEE explore and ACM digital libraries. Articles are included if they provide empirical evidence through conducting controlled user experiments, or provide a review of these empirical studies. The qualitative synthesis of the studies focuses on identification and classifying the methods, and the conditions under which they are examined to positively affect the visualization outcomes. Findings: The keyword search yields 760 studies, of which 30 are included after the title/abstract review. The classification of the included articles shows five distinct methods: interactive design, aesthetic (artistic) style, storytelling, decorative elements that do not provide extra information including text, image, and embellishment on the graphs), and animation. Studies on decorative elements show consistency on the positive effects of these elements on user engagement and recall but are less consistent in their examination of the user performance. This inconsistency could be attributable to the particular data type or specific design method used in each study. The interactive design studies are consistent in their findings of the positive effect on the outcomes. Storytelling studies show some inconsistencies regarding the design effect on user engagement, enjoyment, recall, and performance, which could be indicative of the specific conditions required for the use of this method. Last two methods, aesthetics and animation, have been less frequent in the included articles, and provide consistent positive results on some of the outcomes. Implications for e-government: Review of the visualization best-practice methods show that each of these methods is beneficial under specific conditions. By using these methods in a potentially beneficial condition, OGD practices can promote a wide range of individuals to involve and work with the government data and ultimately engage in government policy-making procedures.Keywords: best practices, data visualization, literature review, open government data
Procedia PDF Downloads 10595 Examining Influence of The Ultrasonic Power and Frequency on Microbubbles Dynamics Using Real-Time Visualization of Synchrotron X-Ray Imaging: Application to Membrane Fouling Control
Authors: Masoume Ehsani, Ning Zhu, Huu Doan, Ali Lohi, Amira Abdelrasoul
Abstract:
Membrane fouling poses severe challenges in membrane-based wastewater treatment applications. Ultrasound (US) has been considered an effective fouling remediation technique in filtration processes. Bubble cavitation in the liquid medium results from the alternating rarefaction and compression cycles during the US irradiation at sufficiently high acoustic pressure. Cavitation microbubbles generated under US irradiation can cause eddy current and turbulent flow within the medium by either oscillating or discharging energy to the system through microbubble explosion. Turbulent flow regime and shear forces created close to the membrane surface cause disturbing the cake layer and dislodging the foulants, which in turn improve the cleaning efficiency and filtration performance. Therefore, the number, size, velocity, and oscillation pattern of the microbubbles created in the liquid medium play a crucial role in foulant detachment and permeate flux recovery. The goal of the current study is to gain in depth understanding of the influence of the US power intensity and frequency on the microbubble dynamics and its characteristics generated under US irradiation. In comparison with other imaging techniques, the synchrotron in-line Phase Contrast Imaging technique at the Canadian Light Source (CLS) allows in-situ observation and real-time visualization of microbubble dynamics. At CLS biomedical imaging and therapy (BMIT) polychromatic beamline, the effective parameters were optimized to enhance the contrast gas/liquid interface for the accuracy of the qualitative and quantitative analysis of bubble cavitation within the system. With the high flux of photons and the high-speed camera, a typical high projection speed was achieved; and each projection of microbubbles in water was captured in 0.5 ms. ImageJ software was used for post-processing the raw images for the detailed quantitative analyses of microbubbles. The imaging has been performed under the US power intensity levels of 50 W, 60 W, and 100 W, in addition to the US frequency levels of 20 kHz, 28 kHz, and 40 kHz. For the duration of 2 seconds of imaging, the effect of the US power and frequency on the average number, size, and fraction of the area occupied by bubbles were analyzed. Microbubbles’ dynamics in terms of their velocity in water was also investigated. For the US power increase of 50 W to 100 W, the average bubble number and the average bubble diameter were increased from 746 to 880 and from 36.7 µm to 48.4 µm, respectively. In terms of the influence of US frequency, a fewer number of bubbles were created at 20 kHz (average of 176 bubbles rather than 808 bubbles at 40 kHz), while the average bubble size was significantly larger than that of 40 kHz (almost seven times). The majority of bubbles were captured close to the membrane surface in the filtration unit. According to the study observations, membrane cleaning efficiency is expected to be improved at higher US power and lower US frequency due to the higher energy release to the system by increasing the number of bubbles or growing their size during oscillation (optimum condition is expected to be at 20 kHz and 100 W).Keywords: bubble dynamics, cavitational bubbles, membrane fouling, ultrasonic cleaning
Procedia PDF Downloads 14994 Manual Wheelchair Propulsion Efficiency on Different Slopes
Authors: A. Boonpratatong, J. Pantong, S. Kiattisaksophon, W. Senavongse
Abstract:
In this study, an integrated sensing and modeling system for manual wheelchair propulsion measurement and propulsion efficiency calculation was used to indicate the level of overuse. Seven subjects participated in the measurement. On the level surface, the propulsion efficiencies were not different significantly as the riding speed increased. By contrast, the propulsion efficiencies on the 15-degree incline were restricted to around 0.5. The results are supported by previously reported wheeling resistance and propulsion torque relationships implying margin of the overuse. Upper limb musculoskeletal injuries and syndromes in manual wheelchair riders are common, chronic, and may be caused at different levels by the overuse i.e. repetitive riding on steep incline. The qualitative analysis such as the mechanical effectiveness on manual wheeling to establish the relationship between the riding difficulties, mechanical efforts and propulsion outputs is scarce, possibly due to the challenge of simultaneous measurement of those factors in conventional manual wheelchairs and everyday environments. In this study, the integrated sensing and modeling system were used to measure manual wheelchair propulsion efficiency in conventional manual wheelchairs and everyday environments. The sensing unit is comprised of the contact pressure and inertia sensors which are portable and universal. Four healthy male and three healthy female subjects participated in the measurement on level and 15-degree incline surface. Subjects were asked to perform manual wheelchair ridings with three different self-selected speeds on level surface and only preferred speed on the 15-degree incline. Five trials were performed in each condition. The kinematic data of the subject’s dominant hand and a spoke and the trunk of the wheelchair were collected through the inertia sensors. The compression force applied from the thumb of the dominant hand to the push rim was collected through the contact pressure sensors. The signals from all sensors were recorded synchronously. The subject-selected speeds for slow, preferred and fast riding on level surface and subject-preferred speed on 15-degree incline were recorded. The propulsion efficiency as a ratio between the pushing force in tangential direction to the push rim and the net force as a result of the three-dimensional riding motion were derived by inverse dynamic problem solving in the modeling unit. The intra-subject variability of the riding speed was not different significantly as the self-selected speed increased on the level surface. Since the riding speed on the 15-degree incline was difficult to regulate, the intra-subject variability was not applied. On the level surface, the propulsion efficiencies were not different significantly as the riding speed increased. However, the propulsion efficiencies on the 15-degree incline were restricted to around 0.5 for all subjects on their preferred speed. The results are supported by the previously reported relationship between the wheeling resistance and propulsion torque in which the wheelchair axle torque increased but the muscle activities were not increased when the resistance is high. This implies the margin of dynamic efforts on the relatively high resistance being similar to the margin of the overuse indicated by the restricted propulsion efficiency on the 15-degree incline.Keywords: contact pressure sensor, inertia sensor, integrating sensing and modeling system, manual wheelchair propulsion efficiency, manual wheelchair propulsion measurement, tangential force, resultant force, three-dimensional riding motion
Procedia PDF Downloads 29093 The Applications of Zero Water Discharge (ZWD) Systems for Environmental Management
Authors: Walter W. Loo
Abstract:
China declared the “zero discharge rules which leave no toxics into our living environment and deliver blue sky, green land and clean water to many generations to come”. The achievement of ZWD will provide conservation of water, soil and energy and provide drastic increase in Gross Domestic Products (GDP). Our society’s engine needs a major tune up; it is sputtering. ZWD is achieved in world’s space stations – no toxic air emission and the water is totally recycled and solid wastes all come back to earth. This is all done with solar power. These are all achieved under extreme temperature, pressure and zero gravity in space. ZWD can be achieved on earth under much less fluctuations in temperature, pressure and normal gravity environment. ZWD systems are not expensive and will have multiple beneficial returns on investment which are both financially and environmentally acceptable. The paper will include successful case histories since the mid-1970s. ZWD discharge can be applied to the following types of projects: nuclear and coal fire power plants with a closed loop system that will eliminate thermal water discharge; residential communities with wastewater treatment sump and recycle the water use as a secondary water supply; waste water treatment Plants with complete water recycling including water distillation to produce distilled water by very economical 24-hours solar power plant. Landfill remediation is based on neutralization of landfilled gas odor and preventing anaerobic leachate formation. It is an aerobic condition which will render landfill gas emission explosion proof. Desert development is the development of recovering soil moisture from soil and completing a closed loop water cycle by solar energy within and underneath an enclosed greenhouse. Salt-alkali land development can be achieved by solar distillation of salty shallow water into distilled water. The distilled water can be used for soil washing and irrigation and complete a closed loop water cycle with energy and water conservation. Heavy metals remediation can be achieved by precipitation of dissolved toxic metals below the plant or vegetation root zone by solar electricity without pumping and treating. Soil and groundwater remediation - abandoned refineries, chemical and pesticide factories can be remediated by in-situ electrobiochemical and bioventing treatment method without pumping or excavation. Toxic organic chemicals are oxidized into carbon dioxide and heavy metals precipitated below plant and vegetation root zone. New water sources: low temperature distilled water can be recycled for repeated use within a greenhouse environment by solar distillation; nano bubble water can be made from the distilled water with nano bubbles of oxygen, nitrogen and carbon dioxide from air (fertilizer water) and also eliminate the use of pesticides because the nano oxygen will break the insect growth chain in the larvae state. Three dimensional high yield greenhouses can be constructed by complete water recycling using the vadose zone soil as a filter with no farming wastewater discharge.Keywords: greenhouses, no discharge, remediation of soil and water, wastewater
Procedia PDF Downloads 34492 Bridging Minds and Nature: Revolutionizing Elementary Environmental Education Through Artificial Intelligence
Authors: Hoora Beheshti Haradasht, Abooali Golzary
Abstract:
Environmental education plays a pivotal role in shaping the future stewards of our planet. Leveraging the power of artificial intelligence (AI) in this endeavor presents an innovative approach to captivate and educate elementary school children about environmental sustainability. This paper explores the application of AI technologies in designing interactive and personalized learning experiences that foster curiosity, critical thinking, and a deep connection to nature. By harnessing AI-driven tools, virtual simulations, and personalized content delivery, educators can create engaging platforms that empower children to comprehend complex environmental concepts while nurturing a lifelong commitment to protecting the Earth. With the pressing challenges of climate change and biodiversity loss, cultivating an environmentally conscious generation is imperative. Integrating AI in environmental education revolutionizes traditional teaching methods by tailoring content, adapting to individual learning styles, and immersing students in interactive scenarios. This paper delves into the potential of AI technologies to enhance engagement, comprehension, and pro-environmental behaviors among elementary school children. Modern AI technologies, including natural language processing, machine learning, and virtual reality, offer unique tools to craft immersive learning experiences. Adaptive platforms can analyze individual learning patterns and preferences, enabling real-time adjustments in content delivery. Virtual simulations, powered by AI, transport students into dynamic ecosystems, fostering experiential learning that goes beyond textbooks. AI-driven educational platforms provide tailored content, ensuring that environmental lessons resonate with each child's interests and cognitive level. By recognizing patterns in students' interactions, AI algorithms curate customized learning pathways, enhancing comprehension and knowledge retention. Utilizing AI, educators can develop virtual field trips and interactive nature explorations. Children can navigate virtual ecosystems, analyze real-time data, and make informed decisions, cultivating an understanding of the delicate balance between human actions and the environment. While AI offers promising educational opportunities, ethical concerns must be addressed. Safeguarding children's data privacy, ensuring content accuracy, and avoiding biases in AI algorithms are paramount to building a trustworthy learning environment. By merging AI with environmental education, educators can empower children not only with knowledge but also with the tools to become advocates for sustainable practices. As children engage in AI-enhanced learning, they develop a sense of agency and responsibility to address environmental challenges. The application of artificial intelligence in elementary environmental education presents a groundbreaking avenue to cultivate environmentally conscious citizens. By embracing AI-driven tools, educators can create transformative learning experiences that empower children to grasp intricate ecological concepts, forge an intimate connection with nature, and develop a strong commitment to safeguarding our planet for generations to come.Keywords: artificial intelligence, environmental education, elementary children, personalized learning, sustainability
Procedia PDF Downloads 8291 A Corpus-Based Study on the Lexical, Syntactic and Sequential Features across Interpreting Types
Authors: Qianxi Lv, Junying Liang
Abstract:
Among the various modes of interpreting, simultaneous interpreting (SI) is regarded as a ‘complex’ and ‘extreme condition’ of cognitive tasks while consecutive interpreters (CI) do not have to share processing capacity between tasks. Given that SI exerts great cognitive demand, it makes sense to posit that the output of SI may be more compromised than that of CI in the linguistic features. The bulk of the research has stressed the varying cognitive demand and processes involved in different modes of interpreting; however, related empirical research is sparse. In keeping with our interest in investigating the quantitative linguistic factors discriminating between SI and CI, the current study seeks to examine the potential lexical simplification, syntactic complexity and sequential organization mechanism with a self-made inter-model corpus of transcribed simultaneous and consecutive interpretation, translated speech and original speech texts with a total running word of 321960. The lexical features are extracted in terms of the lexical density, list head coverage, hapax legomena, and type-token ratio, as well as core vocabulary percentage. Dependency distance, an index for syntactic complexity and reflective of processing demand is employed. Frequency motif is a non-grammatically-bound sequential unit and is also used to visualize the local function distribution of interpreting the output. While SI is generally regarded as multitasking with high cognitive load, our findings evidently show that CI may impose heavier or taxing cognitive resource differently and hence yields more lexically and syntactically simplified output. In addition, the sequential features manifest that SI and CI organize the sequences from the source text in different ways into the output, to minimize the cognitive load respectively. We reasoned the results in the framework that cognitive demand is exerted both on maintaining and coordinating component of Working Memory. On the one hand, the information maintained in CI is inherently larger in volume compared to SI. On the other hand, time constraints directly influence the sentence reformulation process. The temporal pressure from the input in SI makes the interpreters only keep a small chunk of information in the focus of attention. Thus, SI interpreters usually produce the output by largely retaining the source structure so as to relieve the information from the working memory immediately after formulated in the target language. Conversely, CI interpreters receive at least a few sentences before reformulation, when they are more self-paced. CI interpreters may thus tend to retain and generate the information in a way to lessen the demand. In other words, interpreters cope with the high demand in the reformulation phase of CI by generating output with densely distributed function words, more content words of higher frequency values and fewer variations, simpler structures and more frequently used language sequences. We consequently propose a revised effort model based on the result for a better illustration of cognitive demand during both interpreting types.Keywords: cognitive demand, corpus-based, dependency distance, frequency motif, interpreting types, lexical simplification, sequential units distribution, syntactic complexity
Procedia PDF Downloads 17890 Emerging Identities: A Transformative ‘Green Zone’
Authors: Alessandra Swiny, Yiorgos Hadjichristou
Abstract:
There exists an on-going geographical scar creating a division through the Island of Cyprus and its capital, Nicosia. The currently amputated city center is accessed legally by the United Nations convoys, infiltrated only by Turkish and Greek Cypriot army scouts and illegal traders and scavengers. On Christmas day 1963 in Nicosia, Captain M. Hobden of the British Army took a green chinagraph pencil and on a large scale Joint Army-RAF map ‘marked’ the division. From then on this ‘buffer zone’ was called the ‘green line.' This once dividing form, separating the main communities of Greek and Turkish Cypriots from one another, has now been fully reclaimed by an autonomous intruder. It's currently most captivating inhabitant is nature. She keeps taking over, for the past fifty years indigenous and introduced fauna and flora thrive; trees emerge from rooftops and plants, bushes and flowers grow randomly through the once bustling market streets, allowing this ‘no man’s land’ to teem with wildlife. And where are its limits? The idea of fluidity is ever present; it encroaches into the urban and built environment that surrounds it, and notions of ownership and permanence are questioned. Its qualities have contributed significantly in the search for new ‘identities,' expressed in the emergence of new living conditions, be they real or surreal. Without being physically reachable, it can be glimpsed at through punctured peepholes, military bunker windows that act as enticing portals into an emotional and conceptual level of inhabitation. The zone is mystical and simultaneously suspended in time, it triggers people’s imagination, not just that of the two prevailing communities but also of immigrants, refugees, and visitors; it mesmerizes all who come within its proximity. The paper opens a discussion on the issues and the binary questions raised. What is natural and artificial; what is private and public; what is ephemeral and permanent? The ‘green line’ exists in a central fringe condition and can serve in mixing generations and groups of people; mingling functions of living with work and social interaction; merging nature and the human being in a new-found synergy of human hope and survival, allowing thus for new notions of place to be introduced. Questions seek to be answered, such as, “Is the impossibility of dwelling made possible, by interweaving these ‘in-between conditions’ into eloquently traced spaces?” The methodologies pursued are developed through academic research, professional practice projects, and students’ research/design work. Realized projects, case studies and other examples cited both nationally and internationally hold global and local applications. Both paths of the research deal with the explorative understanding of the impossibility of dwelling, testing the limits of its autonomy. The expected outcome of the experience evokes in the user a sense of a new urban landscape, created from human topographies that echo the voice of an emerging identity.Keywords: urban wildlife, human topographies, buffer zone, no man’s land
Procedia PDF Downloads 19889 Lessons Learnt from Industry: Achieving Net Gain Outcomes for Biodiversity
Authors: Julia Baker
Abstract:
Development plays a major role in stopping biodiversity loss. But the ‘silo species’ protection of legislation (where certain species are protected while many are not) means that development can be ‘legally compliant’ and result in biodiversity loss. ‘Net Gain’ (NG) policies can help overcome this by making it an absolute requirement that development causes no overall loss of biodiversity and brings a benefit. However, offsetting biodiversity losses in one location with gains elsewhere is controversial because people suspect ‘offsetting’ to be an easy way for developers to buy their way out of conservation requirements. Yet the good practice principles (GPP) of offsetting provide several advantages over existing legislation for protecting biodiversity from development. This presentation describes the learning from implementing NG approaches based on GPP. It regards major upgrades of the UK’s transport networks, which involved removing vegetation in order to construct and safely operate new infrastructure. While low-lying habitats were retained, trees and other habitats disrupting the running or safety of transport networks could not. Consequently, achieving NG within the transport corridor was not possible and offsetting was required. The first ‘lessons learnt’ were on obtaining a commitment from business leaders to go beyond legislative requirements and deliver NG, and on the institutional change necessary to embed GPP within daily operations. These issues can only be addressed when the challenges that biodiversity poses for business are overcome. These challenges included: biodiversity cannot be measured easily unlike other sustainability factors like carbon and water that have metrics for target-setting and measuring progress; and, the mindset that biodiversity costs money and does not generate cash in return, which is the opposite of carbon or waste for example, where people can see how ‘sustainability’ actions save money. The challenges were overcome by presenting the GPP of NG as a cost-efficient solution to specific, critical risks facing the business that also boost industry recognition, and by using government-issued NG metrics to develop business-specific toolkits charting their NG progress whilst ensuring that NG decision-making was based on rich ecological data. An institutional change was best achieved by supporting, mentoring and training sustainability/environmental managers for these ‘frontline’ staff to embed GPP within the business. The second learning was from implementing the GPP where business partnered with local governments, wildlife groups and land owners to support their priorities for nature conservation, and where these partners had a say in decisions about where and how best to achieve NG. From this inclusive approach, offsetting contributed towards conservation priorities when all collaborated to manage trade-offs between: -Delivering ecologically equivalent offsets or compensating for losses of one type of biodiversity by providing another. -Achieving NG locally to the development whilst contributing towards national conservation priorities through landscape-level planning. -Not just protecting the extent and condition of existing biodiversity but ‘doing more’. -The multi-sector collaborations identified practical, workable solutions to ‘in perpetuity’. But key was strengthening linkages between biodiversity measures implemented for development and conservation work undertaken by local organizations so that developers support NG initiatives that really count.Keywords: biodiversity offsetting, development, nature conservation planning, net gain
Procedia PDF Downloads 19588 Quasi-Photon Monte Carlo on Radiative Heat Transfer: An Importance Sampling and Learning Approach
Authors: Utkarsh A. Mishra, Ankit Bansal
Abstract:
At high temperature, radiative heat transfer is the dominant mode of heat transfer. It is governed by various phenomena such as photon emission, absorption, and scattering. The solution of the governing integrodifferential equation of radiative transfer is a complex process, more when the effect of participating medium and wavelength properties are taken into consideration. Although a generic formulation of such radiative transport problem can be modeled for a wide variety of problems with non-gray, non-diffusive surfaces, there is always a trade-off between simplicity and accuracy of the problem. Recently, solutions of complicated mathematical problems with statistical methods based on randomization of naturally occurring phenomena have gained significant importance. Photon bundles with discrete energy can be replicated with random numbers describing the emission, absorption, and scattering processes. Photon Monte Carlo (PMC) is a simple, yet powerful technique, to solve radiative transfer problems in complicated geometries with arbitrary participating medium. The method, on the one hand, increases the accuracy of estimation, and on the other hand, increases the computational cost. The participating media -generally a gas, such as CO₂, CO, and H₂O- present complex emission and absorption spectra. To model the emission/absorption accurately with random numbers requires a weighted sampling as different sections of the spectrum carries different importance. Importance sampling (IS) was implemented to sample random photon of arbitrary wavelength, and the sampled data provided unbiased training of MC estimators for better results. A better replacement to uniform random numbers is using deterministic, quasi-random sequences. Halton, Sobol, and Faure Low-Discrepancy Sequences are used in this study. They possess better space-filling performance than the uniform random number generator and gives rise to a low variance, stable Quasi-Monte Carlo (QMC) estimators with faster convergence. An optimal supervised learning scheme was further considered to reduce the computation costs of the PMC simulation. A one-dimensional plane-parallel slab problem with participating media was formulated. The history of some randomly sampled photon bundles is recorded to train an Artificial Neural Network (ANN), back-propagation model. The flux was calculated using the standard quasi PMC and was considered to be the training target. Results obtained with the proposed model for the one-dimensional problem are compared with the exact analytical and PMC model with the Line by Line (LBL) spectral model. The approximate variance obtained was around 3.14%. Results were analyzed with respect to time and the total flux in both cases. A significant reduction in variance as well a faster rate of convergence was observed in the case of the QMC method over the standard PMC method. However, the results obtained with the ANN method resulted in greater variance (around 25-28%) as compared to the other cases. There is a great scope of machine learning models to help in further reduction of computation cost once trained successfully. Multiple ways of selecting the input data as well as various architectures will be tried such that the concerned environment can be fully addressed to the ANN model. Better results can be achieved in this unexplored domain.Keywords: radiative heat transfer, Monte Carlo Method, pseudo-random numbers, low discrepancy sequences, artificial neural networks
Procedia PDF Downloads 22387 Solid Polymer Electrolyte Membranes Based on Siloxane Matrix
Authors: Natia Jalagonia, Tinatin Kuchukhidze
Abstract:
Polymer electrolytes (PE) play an important part in electrochemical devices such as batteries and fuel cells. To achieve optimal performance, the PE must maintain a high ionic conductivity and mechanical stability at both high and low relative humidity. The polymer electrolyte also needs to have excellent chemical stability for long and robustness. According to the prevailing theory, ionic conduction in polymer electrolytes is facilitated by the large-scale segmental motion of the polymer backbone, and primarily occurs in the amorphous regions of the polymer electrolyte. Crystallinity restricts polymer backbone segmental motion and significantly reduces conductivity. Consequently, polymer electrolytes with high conductivity at room temperature have been sought through polymers which have highly flexible backbones and have largely amorphous morphology. The interest in polymer electrolytes was increased also by potential applications of solid polymer electrolytes in high energy density solid state batteries, gas sensors and electrochromic windows. Conductivity of 10-3 S/cm is commonly regarded as a necessary minimum value for practical applications in batteries. At present, polyethylene oxide (PEO)-based systems are most thoroughly investigated, reaching room temperature conductivities of 10-7 S/cm in some cross-linked salt in polymer systems based on amorphous PEO-polypropylene oxide copolymers.. It is widely accepted that amorphous polymers with low glass transition temperatures Tg and a high segmental mobility are important prerequisites for high ionic conductivities. Another necessary condition for high ionic conductivity is a high salt solubility in the polymer, which is most often achieved by donors such as ether oxygen or imide groups on the main chain or on the side groups of the PE. It is well established also that lithium ion coordination takes place predominantly in the amorphous domain, and that the segmental mobility of the polymer is an important factor in determining the ionic mobility. Great attention was pointed to PEO-based amorphous electrolyte obtained by synthesis of comb-like polymers, by attaching short ethylene oxide unit sequences to an existing amorphous polymer backbone. The aim of presented work is to obtain of solid polymer electrolyte membranes using PMHS as a matrix. For this purpose the hydrosilylation reactions of α,ω-bis(trimethylsiloxy)methyl¬hydrosiloxane with allyl triethylene-glycol mo¬nomethyl ether and vinyltriethoxysilane at 1:28:7 ratio of initial com¬pounds in the presence of Karstedt’s catalyst, platinum hydrochloric acid (0.1 M solution in THF) and platinum on the carbon catalyst in 50% solution of anhydrous toluene have been studied. The synthesized olygomers are vitreous liquid products, which are well soluble in organic solvents with specific viscosity ηsp ≈ 0.05 - 0.06. The synthesized olygomers were analysed with FTIR, 1H, 13C, 29Si NMR spectroscopy. Synthesized polysiloxanes were investigated with wide-angle X-ray, gel-permeation chromatography, and DSC analyses. Via sol-gel processes of doped with lithium trifluoromethylsulfonate (triflate) or lithium bis¬(trifluoromethylsulfonyl)¬imide polymer systems solid polymer electrolyte membranes have been obtained. The dependence of ionic conductivity as a function of temperature and salt concentration was investigated and the activation energies of conductivity for all obtained compounds are calculatedKeywords: synthesis, PMHS, membrane, electrolyte
Procedia PDF Downloads 25786 Comparing Perceived Restorativeness in Natural and Urban Environment: A Meta-Analysis
Authors: Elisa Menardo, Margherita Pasini, Margherita Brondino
Abstract:
A growing body of empirical research from different areas of inquiry suggests that brief contact with natural environment restore mental resources. The Attention Restoration Theory (ART) is the widespread used and empirical founded theory developed to explain why exposure to nature helps people to recovery cognitive resources. It assumes that contact with nature allows people to free (and then recovery) voluntary attention resources and thus allows them to recover from a cognitive fatigue situation. However, it was suggested that some people could have more cognitive benefit after exposure to urban environment. The objective of this study is to report the results of a meta-analysis on studies (peer-reviewed articles) comparing the restorativeness (the quality to be restorative) perceived in natural environments than those perceived in urban environments. This meta-analysis intended to estimate how much nature environments (forests, parks, boulevards) are perceived to be more restorativeness than urban ones (i.e., the magnitude of the perceived restorativeness’ difference). Moreover, given the methodological difference between study, it studied the potential role of moderator variables as participants (student or other), instrument used (Perceived Restorativeness Scale or other), and procedure (in laboratory or in situ). PsycINFO, PsycARTICLES, Scopus, SpringerLINK, Web of Science online database were used to identify all peer-review articles on restorativeness published to date (k = 167). Reference sections of obtained papers were examined for additional studies. Only 22 independent studies (with a total of 1371 participants) met inclusion criteria (direct exposure to environment, comparison between one outdoor environment with natural element and one without natural element, and restorativeness measured by self-report scale) and were included in meta-analysis. To estimate the average effect size, a random effect model (Restricted Maximum-likelihood estimator) was used because the studies included in the meta-analysis were conducted independently and using different methods in different populations, so no common effect-size was expected. The presence of publication bias was checked using trim and fill approach. Univariate moderator analysis (mixed effect model) were run to determine whether the variable coded moderated the perceived restorativeness difference. Results show that natural environments are perceived to be more restorativeness than urban environments, confirming from an empirical point of view what is now considered a knowledge gained in environmental psychology. The relevant information emerging from this study is the magnitude of the estimated average effect size, which is particularly high (d = 1.99) compared to those that are commonly observed in psychology. Significant heterogeneity between study was found (Q(19) = 503.16, p < 0.001;) and studies’ variability was very high (I2[C.I.] = 96.97% [94.61 - 98.62]). Subsequent univariate moderator analyses were not significant. Methodological difference (participants, instrument, and procedure) did not explain variability between study. Other methodological difference (e.g., research design, environment’s characteristics, light’s condition) could explain this variability between study. In the mine while, studies’ variability could be not due to methodological difference but to individual difference (age, gender, education level) and characteristics (connection to nature, environmental attitude). Furthers moderator analysis are working in progress.Keywords: meta-analysis, natural environments, perceived restorativeness, urban environments
Procedia PDF Downloads 16985 Towards Better Integration: Qualitative Study on Perceptions of Russian-Speaking Immigrants in Australia
Authors: Oleg Shovkovyy
Abstract:
This research conducted in response to one of the most pressing questions on the agenda of many public administration offices around the world: “What could be done for better integration and assimilation of immigrants into hosting communities?” In author’s view, the answer could be suggested by immigrants themselves. They, often ‘bogged down in the past,’ snared by own idols and demons, perceive things differently, which, in turn, may result in their inability to integrate smoothly into hosting communities. Brief literature review suggests that perceptions of immigrants are completely neglected or something unsought in the current research on migrants, which, often, based on opinion polls by members of hosting communities themselves or superficial research data by various research organizations. Even those specimens that include voices of immigrants, unlikely to shed any additional light onto the problem simply because certain things are not made to speak out loud, especially to those in whose hands immigrants’ fate is (authorities). In this regard, this qualitative study, conducted by an insider to a few Russian-speaking communities, represents a unique opportunity for all stakeholders to look at the question of integration through the eyes of immigrants, from a different perspective and thus, makes research findings especially valuable for better understanding of the problem. Case study research employed ethnographic methods of gathering data where, approximately 200 Russian-speaking immigrants of first and second generations were closely observed by the Russian-speaking researcher in their usual setting, for eight months, and at different venues. The number of informal interviews with 27 key informants, with whom the researcher managed to establish a good rapport and who were keen enough to share their experiences voluntarily, were conducted. The field notes were taken at 14 locations (study sites) within the Brisbane region of Queensland, Australia. Moreover, all this time, researcher lived in dwelling of one of the immigrants and was an active participant in the social life (worship, picnics, dinners, weekend schools, concerts, cultural events, social gathering, etc.) of observed communities, whose members, to a large extent, belong to various religious lines of the Russian and Protestant Church. It was found that the majority of immigrants had experienced some discrimination in matters of hiring, employment, recognition of educational qualifications from home countries, and simply felt a sort of dislike from society in various everyday situations. Many noted complete absences or very limited state assistance in terms of employment, training, education, and housing. For instance, the Australian Government Department of Human Services not only does not stimulate job search but, on the contrary, encourages to refuse short-term works and employment. On the other hand, offered free courses on adaptation, and the English language proved to be ineffective and unpopular amongst immigrants. Many interviewees have reported overstated requirements for English proficiency and local work experience, whereas it was not critical for the given task or job. Based on the result of long-term monitoring, the researcher also had the courage to assert the negative and decelerating roles of immigrants’ communities, particularly religious communities, on processes of integration and assimilation. The findings suggest that governments should either change current immigration policies in the direction of their toughening or to take more proactive and responsible role in dealing with immigrant-related issues; for instance, increasing assistance and support to all immigrants and probably, paying more attention to and taking stake in managing and organizing lives of immigrants’ communities rather, simply leaving it all to chance.Keywords: Australia, immigration, integration, perceptions
Procedia PDF Downloads 22084 Predicting Open Chromatin Regions in Cell-Free DNA Whole Genome Sequencing Data by Correlation Clustering
Authors: Fahimeh Palizban, Farshad Noravesh, Amir Hossein Saeidian, Mahya Mehrmohamadi
Abstract:
In the recent decade, the emergence of liquid biopsy has significantly improved cancer monitoring and detection. Dying cells, including those originating from tumors, shed their DNA into the blood and contribute to a pool of circulating fragments called cell-free DNA. Accordingly, identifying the tissue origin of these DNA fragments from the plasma can result in more accurate and fast disease diagnosis and precise treatment protocols. Open chromatin regions are important epigenetic features of DNA that reflect cell types of origin. Profiling these features by DNase-seq, ATAC-seq, and histone ChIP-seq provides insights into tissue-specific and disease-specific regulatory mechanisms. There have been several studies in the area of cancer liquid biopsy that integrate distinct genomic and epigenomic features for early cancer detection along with tissue of origin detection. However, multimodal analysis requires several types of experiments to cover the genomic and epigenomic aspects of a single sample, which will lead to a huge amount of cost and time. To overcome these limitations, the idea of predicting OCRs from WGS is of particular importance. In this regard, we proposed a computational approach to target the prediction of open chromatin regions as an important epigenetic feature from cell-free DNA whole genome sequence data. To fulfill this objective, local sequencing depth will be fed to our proposed algorithm and the prediction of the most probable open chromatin regions from whole genome sequencing data can be carried out. Our method integrates the signal processing method with sequencing depth data and includes count normalization, Discrete Fourie Transform conversion, graph construction, graph cut optimization by linear programming, and clustering. To validate the proposed method, we compared the output of the clustering (open chromatin region+, open chromatin region-) with previously validated open chromatin regions related to human blood samples of the ATAC-DB database. The percentage of overlap between predicted open chromatin regions and the experimentally validated regions obtained by ATAC-seq in ATAC-DB is greater than 67%, which indicates meaningful prediction. As it is evident, OCRs are mostly located in the transcription start sites (TSS) of the genes. In this regard, we compared the concordance between the predicted OCRs and the human genes TSS regions obtained from refTSS and it showed proper accordance around 52.04% and ~78% with all and the housekeeping genes, respectively. Accurately detecting open chromatin regions from plasma cell-free DNA-seq data is a very challenging computational problem due to the existence of several confounding factors, such as technical and biological variations. Although this approach is in its infancy, there has already been an attempt to apply it, which leads to a tool named OCRDetector with some restrictions like the need for highly depth cfDNA WGS data, prior information about OCRs distribution, and considering multiple features. However, we implemented a graph signal clustering based on a single depth feature in an unsupervised learning manner that resulted in faster performance and decent accuracy. Overall, we tried to investigate the epigenomic pattern of a cell-free DNA sample from a new computational perspective that can be used along with other tools to investigate genetic and epigenetic aspects of a single whole genome sequencing data for efficient liquid biopsy-related analysis.Keywords: open chromatin regions, cancer, cell-free DNA, epigenomics, graph signal processing, correlation clustering
Procedia PDF Downloads 15083 Spinetoram10% WG+Sulfoxaflor 30% WG: A Promising Green Chemistry to Manage Pest Complex in Bt Cotton
Authors: Siddharudha B. Patil
Abstract:
Cotton is a premier commercial fibre crop of India subjected to ravages of insect pests. Sucking pests viz thrips, Thrips tabaci,(lind) leaf hopper Amrsca devastance,(dist) miridbug, Poppiocapsidea beseratense (Dist) and bollworms continue to inflict damage Bt Cotton right from seeding stage. Their infestation impact cotton yield to an extent of 30-40 percent. Chemical control is still adoptable as one of the techniques for combating these pests. Presently, growers have many challenges in selecting effective chemicals which fit in with an integrated pest management. Spinetoram has broad spectrum with excellent insecticidal activity against both sucking pests and bollworms. Hence, it is expected to make a great contribution to stable production and quality improvement of agricultural products. Spinetoram is a derivative of biologically active substances (Spinosyns) produced by soil actinomycetes, Saccharopolypara spinosa which is semi synthetic active ingredient representing Spinosyn chemical class of insecticide and has demonstrated higher level of efficacy with reduced risk on beneficial arthropods. The efforts were made in the present study to test the efficacy of Spinetoram against sucking pests and bollworms in comparison with other insecticides in Bt Cotton under field condition. Field experiment was laid out during 2013-14 and 2014-15 at Agricultural Research station Dharwad (Karnataka-India) in a randomized block design comprising eight treatments and three replications. Bt cotton genotype, Bunny BG-II was sown in a plot size of 5.4 m x5.4 m. Recommend agronomical practices were followed. The Spinetoram 12% SC alone and incombination with sulfaxaflore with varied dosages against pest complex was tested. Performance was compared with Spinosad 45% SC and thiamethoxam 25% WG. The results of consecutive seasons revealed that nonsignificant difference in thrips and leafhopper population and varied significantly after 3 days of imposition. Among the treatments, combiproduct, Spinetoram 10%WG + Sulfoxaflor 30% WG@ 140 gai/ha registered lowest population of thrips (3.91/3 leaves) and leaf hoppers (1.08/3 leaves) followed by its lower dosages viz 120 gai/ha (4.86/3 leaves and 1.14/3 leaves of thrips and leaf hoppers, respectively) and 100 gai/ha (6.02 and 1.23./3 leaves of thrips and leaf hoppers respectively) being at par, significantly superior to rest of the treatments. On the contrary, the population of thrips, leaf hopper and miridbugs in untreated control was on higher side. Similarly the higher dosage of Spinetoram 10% WG+ Sulfoxaflor 30% WG (140 gai/ha) proved its bioefficacy by registering lowest miridbug incidence of 1.70/25 squares, followed by its lower dosage (1.78 and 1.83/25 squares respectively) Further observation made on bollworms incidence revealed that the higher dosage of Spinetoram 10% WG+Sulfoxaflor 30% WG (140 gai/ha) registered lowest percentage of boll damage (7.22%), more number of good opened bolls (36.89/plant) and higher seed cotton yield (19.45q/ha) followed by rest of its lower dosages, Spinetoram 12% SC alone and Spinosad 45% SC being at par significantly superior to rest of the treatments. However, significantly higher boll damage (15.13%) and lower seed cotton yield (14.45 q/ha) was registered in untreated control. Thus Spinetoram10% WG+Sulfoxaflor 30% WG can be a promising option for pest management in Bt Cotton.Keywords: Spinetoram10% WG+Sulfoxaflor 30% WG, sucking pests, bollworms, Bt cotton, management
Procedia PDF Downloads 25082 Mining and Ecological Events and its Impact on the Genesis and Geo-Distribution of Ebola Outbreaks in Africa
Authors: E Tambo, O. O. Olalubi, E. C. Ugwu, J. Y. Ngogang
Abstract:
Despite the World Health Organization (WHO) declaration of international health emergency concern, the status quo of responses and efforts to stem the worst-recorded Ebola epidemic Ebola outbreak is still precariously inadequate in most of the affected in West. Mining natural resources have been shown to play a key role in both motivating and fuelling ethnic, civil and armed conflicts that have plagued a number of African countries over the last decade. Revenues from the exploitation of natural resources are not only used in sustaining the national economy but also armies, personal enrichment and building political support. Little is documented on the mining and ecological impact on the emergence and geographical distribution of Ebola in Africa over time and space. We aimed to provide a better understanding of the interconnectedness among issues of mining natural, resource management, mining conflict and post-conflict on Ebola outbreak and how wealth generated from abundant natural resources could be better managed in promoting research and development towards strengthening environmental, socioeconomic and health systems sustainability on Ebola outbreak and other emerging diseases surveillance and responses systems prevention and control, early warning alert, durable peace and sustainable development rather than to fuel conflicts, resurgence and emerging diseases epidemics in the perspective of community and national/regional approach. Our results showed the first assessment of systematic impact of all major minerals conflict events diffusion over space and time and mining activities on nine Ebola genesis and geo-distribution in affected countries across Africa. We demonstrate how, where and when mining activities in Africa increase ecological degradation, conflicts at the local level and then spreads violence across territory and time by enhancing the financial capacities of fighting groups/ethnics and diseases onset. In addition, led process of developing minimum standards for natural resource governance; improving governmental and civil society capacity for natural resource management, including the strengthening of monitoring and enforcement mechanisms; understanding the post-mining and conflicts community or national reconstruction and rehabilitation programmes in strengthening or developing community health systems and regulatory mechanisms. In addition the quest for the control over these resources and illegal mining across the landscape forest incursion provided increase environmental and ecological instability and displacement and disequilibrium, therefore affecting the intensity and duration of mining and conflict/wars and episode of Ebola outbreaks over time and space. We highlight the key findings and lessons learnt in promoting country or community-led process in transforming natural resource wealth from a peace liability to a peace asset. The imperative necessity for advocacy and through facilitating intergovernmental deliberations on critical issues and challenges affecting Africa community transforming exploitation of natural resources from a peace liability to outbreak prevention and control. The vital role of mining in increasing government revenues and expenditures, equitable distribution of wealth and health to all stakeholders, in particular local communities requires coordination, cooperative leadership and partnership in fostering sustainable developmental initiatives from mining context to outbreak and other infectious diseases surveillance responses systems in prevention and control, and judicious resource management.Keywords: mining, mining conflicts, mines, ecological, Ebola, outbreak, mining companies, miners, impact
Procedia PDF Downloads 30281 Evaluation of Physical Parameters and in-Vitro and in-Vivo Antidiabetic Activity of a Selected Combined Medicinal Plant Extracts Mixture
Authors: S. N. T. I. Sampath, J. M. S. Jayasinghe, A. P. Attanayake, V. Karunaratne
Abstract:
Diabetes mellitus is one of the major public health posers throughout the world today that incidence and associated with increasing mortality. Insufficient regulation of the blood glucose level might be serious effects for health and its necessity to identify new therapeutics that have ability to reduce hyperglycaemic condition in the human body. Even though synthetic antidiabetic drugs are more effective to control diabetes mellitus, there are considerable side effects have been reported. Thus, there is an increasing demand for searching new natural products having high antidiabetic activity with lesser side effects. The purposes of the present study were to evaluate different physical parameters and in-vitro and in-vivo antidiabetic potential of the selected combined medicinal plant extracts mixture composed of leaves of Murraya koenigii, cloves of Allium sativum, fruits of Garcinia queasita and seeds of Piper nigrum. The selected plants parts were mixed and ground together and extracted sequentially into the hexane, ethyl acetate and methanol. Solvents were evaporated and they were further dried by freeze-drying to obtain a fine powder of each extract. Various physical parameters such as moisture, total ash, acid insoluble ash and water soluble ash were evaluated using standard test procedures. In-vitro antidiabetic activity of combined plant extracts mixture was screened using enzyme assays such as α-amylase inhibition assay and α-glucosidase inhibition assay. The acute anti-hyperglycaemic activity was performed using oral glucose tolerance test for the streptozotocin induced diabetic Wistar rats to find out in-vivo antidiabetic activity of combined plant extracts mixture and it was assessed through total oral glucose tolerance curve (TAUC) values. The percentage of moisture content, total ash content, acid insoluble ash content and water soluble ash content were ranged of 7.6-17.8, 8.1-11.78, 0.019-0.134 and 6.2-9.2 respectively for the plant extracts and those values were less than standard values except the methanol extract. The hexane and ethyl acetate extracts exhibited highest α-amylase (IC50 = 25.7 ±0.6; 27.1 ±1.2 ppm) and α-glucosidase (IC50 = 22.4 ±0.1; 33.7 ±0.2 ppm) inhibitory activities than methanol extract (IC50 = 360.2 ±0.6; 179.6 ±0.9 ppm) when compared with the acarbose positive control (IC50 = 5.7 ±0.4; 17.1 ±0.6 ppm). The TAUC values for hexane, ethyl acetate, and methanol extracts and glibenclamide (positive control) treated rats were 8.01 ±0.66; 8.05 ±1.07; 8.40±0.50; 5.87 ±0.93 mmol/L.h respectively, whereas in diabetic control rats the TAUC value was 13.22 ±1.07 mmol/L.h. Administration of plant extracts treated rats significantly suppressed (p<0.05) the rise in plasma blood glucose levels compared to control rats but less significant than glibenclamide. The obtained results from in-vivo and in-vitro antidiabetic study showed that the hexane and ethyl acetate extracts of selected combined plant mixture might be considered as a potential source to isolate natural antidiabetic agents and physical parameters of hexane and ethyl acetate extracts will helpful to develop antidiabetic drug with further standardize properties.Keywords: diabetes mellitus, in-vitro antidiabetic assays, medicinal plants, standardization
Procedia PDF Downloads 13180 Accurate Energy Assessment Technique for Mine-Water District Heat Network
Authors: B. Philip, J. Littlewood, R. Radford, N. Evans, T. Whyman, D. P. Jones
Abstract:
UK buildings and energy infrastructures are heavily dependent on natural gas, a large proportion of which is used for domestic space heating. However, approximately half of the gas consumed in the UK is imported. Improving energy security and reducing carbon emissions are major government drivers for reducing gas dependency. In order to do so there needs to be a wholesale shift in the energy provision to householders without impacting on thermal comfort levels, convenience or cost of supply to the end user. Heat pumps are seen as a potential alternative in modern well insulated homes, however, can the same be said of older homes? A large proportion of housing stock in Britain was built prior to 1919. The age of the buildings bears testimony to the quality of construction; however, their thermal performance falls far below the minimum currently set by UK building standards. In recent years significant sums of money have been invested to improve energy efficiency and combat fuel poverty in some of the most deprived areas of Wales. Increasing energy efficiency of older properties remains a significant challenge, which cannot be achieved through insulation and air-tightness interventions alone, particularly when alterations to historically important architectural features of the building are not permitted. This paper investigates the energy demand of pre-1919 dwellings in a former Welsh mining village, the feasibility of meeting that demand using water from the disused mine workings to supply a district heat network and potential barriers to success of the scheme. The use of renewable solar energy generation and storage technologies, both thermal and electrical, to reduce the load and offset increased electricity demand, are considered. A wholistic surveying approach to provide a more accurate assessment of total household heat demand is proposed. Several surveying techniques, including condition surveys, air permeability, heat loss calculations, and thermography were employed to provide a clear picture of energy demand. Additional insulation can bring unforeseen consequences which are detrimental to the fabric of the building, potentially leading to accelerated dilapidation of the asset being ‘protected’. Increasing ventilation should be considered in parallel, to compensate for the associated reduction in uncontrolled infiltration. The effectiveness of thermal performance improvements are demonstrated and the detrimental effects of incorrect material choice and poor installation are highlighted. The findings show estimated heat demand to be in close correlation to household energy bills. Major areas of heat loss were identified such that improvements to building thermal performance could be targeted. The findings demonstrate that the use of heat pumps in older buildings is viable, provided sufficient improvement to thermal performance is possible. Addition of passive solar thermal and photovoltaic generation can help reduce the load and running cost for the householder. The results were used to predict future heat demand following energy efficiency improvements, thereby informing the size of heat pumps required.Keywords: heat demand, heat pump, renewable energy, retrofit
Procedia PDF Downloads 9279 Evaluation of Forensic Pathology Practice Outside Germany – Experiences From 20 Years of Second Look Autopsies in Cooperation with the Institute of Legal Medicine Munich
Authors: Michael Josef Schwerer, Oliver Peschel
Abstract:
Background: The sense and purpose of forensic postmortem examinations are undoubtedly the same in Institutes of Legal Medicine all over the world. Cause and manner of death must be determined, persons responsible for unnatural death must be brought to justice, and accidents demand changes in the respective scenarios to avoid future mishaps. The latter particularly concerns aircraft accidents, not only regarding consequences from criminal or civil law but also in pursuance of the International Civil Aviation Authority’s regulations, which demand lessons from mishap investigations to improve flight safety. Irrespective of the distinct circumstances of a given casualty or the respective questions in subsequent death investigations, a forensic autopsy is the basis for all further casework, the clue to otherwise hidden solutions, and the crucial limitation for final success when not all possible findings have been properly collected. This also implies that the targeted work of police forces and expert witnesses strongly depends on the quality of forensic pathology practice. Deadly events in foreign countries, which lead to investigations not only abroad but also in Germany, can be challenging in this context. Frequently, second-look autopsies after the repatriation of the deceased to Germany are requested by the legal authorities to ensure proper and profound documentation of all relevant findings. Aims and Methods: To validate forensic postmortem practice abroad, a retrospective study using the findings in the corresponding second-look autopsies in the Institute of Legal Medicine Munich over the last 20 years was carried out. New findings unreported in the previous autopsy were recorded and judged for their relevance to solving the respective case. Further, the condition of the corpse at the time of the second autopsy was rated to discuss artifacts mimicking evidence or the possibility of lost findings resulting from, e.g., decomposition. Recommendations for future handling of death cases abroad and efficient autopsy practice were pursued. Results and Discussion: Our re-evaluation confirmed a high quality of autopsy practice abroad in the vast majority of cases. However, in some casework, incomplete documentation of pathology findings was revealed along with either insufficient or misconducted dissection of organs. Further, some of the bodies showed missing parts of some organs, most probably resulting from sampling for histology studies during the first postmortem. For the aeromedical evaluation of a decedent’s health status prior to an aviation mishap, particularly lost or obscured findings in the heart, lungs, and brain impeded expert testimony. Moreover, incomplete fixation of the body or body parts for repatriation was seen in several cases. This particularly involved previously dissected organs deposited back into the body cavities at the end of the first autopsy. Conclusions and Recommendations: Detailed preparation in the first forensic autopsy avoids the necessity of a second-look postmortem in the majority of cases. To limit decomposition changes during repatriation from abroad, special care must be taken to include pre-dissected organs in the chemical fixation process, particularly when they are separated from the blood vessels and just deposited back into the body cavities.Keywords: autopsy practice, second-look autopsy, retrospective study, quality standards, decomposition changes, repatriation
Procedia PDF Downloads 4978 Use of Zikani’s Ribosome Modulating Agents for Treating Recessive Dystrophic & Junctional Epidermolysis Bullosa with Nonsense Mutations
Authors: Mei Chen, Yingping Hou, Michelle Hao, Soheil Aghamohammadzadeh, Esteban Terzo, Roger Clark, Vijay Modur
Abstract:
Background: Recessive Dystrophic Epidermolysis Bullosa (RDEB) is a genetic skin condition characterized by skin tearing and unremitting blistering upon minimal trauma. Repeated blistering, fibrosis, and scarring lead to aggressive squamous cell carcinoma later in life. RDEB is caused by mutations in the COL7A1 gene encoding collagen type VII (C7), the major component of anchoring fibrils mediating epidermis-dermis adherence. Nonsense mutations in the COL7A1 gene of a subset of RDEB patients leads to premature termination codons (PTC). Similarly, most Junctional Epidermolysis Bullosa (JEB) cases are caused by nonsense mutations in the LAMB3 gene encoding the β3 subunit of laminin 332. Currently, there is an unmet need for the treatment of RDEB and JEB. Zikani Therapeutics has discovered an array of macrocyclic compounds with ring structures similar to macrolide antibiotics that can facilitate readthrough activity of nonsense mutations in the COL7A1 and LAMB3 genes by acting as Ribosome Modulating Agents (RMAs). The medicinal chemistry synthetic advancements of these macrocyclic compounds have allowed targeting the human ribosome while preserving the structural elements responsible for the safety and pharmacokinetic profile of clinically used macrolide antibiotics. Methods: C7 expression was used as a measure of readthrough activity by immunoblot assays in two primary human fibroblasts from RDEB patients (R578X/R578X and R163X/R1683X-COL7A1). Similarly, immunoblot assays in C325X/c.629-12T > A-LAMB3 keratinocytes were used to measure readthrough activity for JEB. The relative readthrough activity of each compound was measured relative to Gentamicin. An imaging-based fibroblast migration assay was used as an assessment of C7 functionality in RDEB-fibroblasts over 16-20 hrs. The incubation period for the above experiments was 48 hrs for RDEB fibroblasts and 72 hours for JEB keratinocytes. Results: 9 RMAs demonstrated increased protein expression in both patient RDEB fibroblasts. The highest readthrough activity at tested concentrations without cytotoxicities increased protein expression up to 179% of Gentamicin (400 µg/ml), with favored readthrough activity in R163X/R1683X-COL7A1 fibroblasts. Concurrent with protein expression, fibroblast hypermotility phenotype observed in RDEB was rescued by reducing motility by ~35% to WT levels (the same level as 690 µM Gentamicin treated cells). Laminin β3 expression was also shown to be increased by 6 RMAs in keratinocytes to 33-83% of (400 µg/ml) Gentamicin. Conclusions: To date, 9 RMAs have been identified that enhance the expression of functional C7 in a mutation-dependent manner in two different RDEB patient fibroblast backgrounds (R578X/R578X and R163X/R1683X-COL7A1). A further 6 RMAs have been identified that enhance the readthrough of C325X-LAMB3 in JEB patient keratinocytes. Based on the clinical trial conducted by us with topical gentamycin in 2017, Zikani’s RMAs achieve clinically significant levels of read-through for the treatment of recessive dystrophic and Junctional Epidermolysis Bullosa.Keywords: epidermolysis bullosa, nonsense mutation, readthrough, ribosome modulation
Procedia PDF Downloads 9877 Braille Lab: A New Design Approach for Social Entrepreneurship and Innovation in Assistive Tools for the Visually Impaired
Authors: Claudio Loconsole, Daniele Leonardis, Antonio Brunetti, Gianpaolo Francesco Trotta, Nicholas Caporusso, Vitoantonio Bevilacqua
Abstract:
Unfortunately, many people still do not have access to communication, with specific regard to reading and writing. Among them, people who are blind or visually impaired, have several difficulties in getting access to the world, compared to the sighted. Indeed, despite technology advancement and cost reduction, nowadays assistive devices are still expensive such as Braille-based input/output systems which enable reading and writing texts (e.g., personal notes, documents). As a consequence, assistive technology affordability is fundamental in supporting the visually impaired in communication, learning, and social inclusion. This, in turn, has serious consequences in terms of equal access to opportunities, freedom of expression, and actual and independent participation to a society designed for the sighted. Moreover, the visually impaired experience difficulties in recognizing objects and interacting with devices in any activities of daily living. It is not a case that Braille indications are commonly reported only on medicine boxes and elevator keypads. Several software applications for the automatic translation of written text into speech (e.g., Text-To-Speech - TTS) enable reading pieces of documents. However, apart from simple tasks, in many circumstances TTS software is not suitable for understanding very complicated pieces of text requiring to dwell more on specific portions (e.g., mathematical formulas or Greek text). In addition, the experience of reading\writing text is completely different both in terms of engagement, and from an educational perspective. Statistics on the employment rate of blind people show that learning to read and write provides the visually impaired with up to 80% more opportunities of finding a job. Especially in higher educational levels, where the ability to digest very complex text is key, accessibility and availability of Braille plays a fundamental role in reducing drop-out rate of the visually impaired, thus affecting the effectiveness of the constitutional right to get access to education. In this context, the Braille Lab project aims at overcoming these social needs by including affordability in designing and developing assistive tools for visually impaired people. In detail, our awarded project focuses on a technology innovation of the operation principle of existing assistive tools for the visually impaired leaving the Human-Machine Interface unchanged. This can result in a significant reduction of the production costs and consequently of tool selling prices, thus representing an important opportunity for social entrepreneurship. The first two assistive tools designed within the Braille Lab project following the proposed approach aims to provide the possibility to personally print documents and handouts and to read texts written in Braille using refreshable Braille display, respectively. The former, named ‘Braille Cartridge’, represents an alternative solution for printing in Braille and consists in the realization of an electronic-controlled dispenser printing (cartridge) which can be integrated within traditional ink-jet printers, in order to leverage the efficiency and cost of the device mechanical structure which are already being used. The latter, named ‘Braille Cursor’, is an innovative Braille display featuring a substantial technology innovation by means of a unique cursor virtualizing Braille cells, thus limiting the number of active pins needed for Braille characters.Keywords: Human rights, social challenges and technology innovations, visually impaired, affordability, assistive tools
Procedia PDF Downloads 27376 Risks for Cyanobacteria Harmful Algal Blooms in Georgia Piedmont Waterbodies Due to Land Management and Climate Interactions
Authors: Sam Weber, Deepak Mishra, Susan Wilde, Elizabeth Kramer
Abstract:
The frequency and severity of cyanobacteria harmful blooms (CyanoHABs) have been increasing over time, with point and non-point source eutrophication and shifting climate paradigms being blamed as the primary culprits. Excessive nutrients, warm temperatures, quiescent water, and heavy and less regular rainfall create more conducive environments for CyanoHABs. CyanoHABs have the potential to produce a spectrum of toxins that cause gastrointestinal stress, organ failure, and even death in humans and animals. To promote enhanced, proactive CyanoHAB management, risk modeling using geospatial tools can act as predictive mechanisms to supplement current CyanoHAB monitoring, management and mitigation efforts. The risk maps would empower water managers to focus their efforts on high risk water bodies in an attempt to prevent CyanoHABs before they occur, and/or more diligently observe those waterbodies. For this research, exploratory spatial data analysis techniques were used to identify the strongest predicators for CyanoHAB blooms based on remote sensing-derived cyanobacteria cell density values for 771 waterbodies in the Georgia Piedmont and landscape characteristics of their watersheds. In-situ datasets for cyanobacteria cell density, nutrients, temperature, and rainfall patterns are not widely available, so free gridded geospatial datasets were used as proxy variables for assessing CyanoHAB risk. For example, the percent of a watershed that is agriculture was used as a proxy for nutrient loading, and the summer precipitation within a watershed was used as a proxy for water quiescence. Cyanobacteria cell density values were calculated using atmospherically corrected images from the European Space Agency’s Sentinel-2A satellite and multispectral instrument sensor at a 10-meter ground resolution. Seventeen explanatory variables were calculated for each watershed utilizing the multi-petabyte geospatial catalogs available within the Google Earth Engine cloud computing interface. The seventeen variables were then used in a multiple linear regression model, and the strongest predictors of cyanobacteria cell density were selected for the final regression model. The seventeen explanatory variables included land cover composition, winter and summer temperature and precipitation data, topographic derivatives, vegetation index anomalies, and soil characteristics. Watershed maximum summer temperature, percent agriculture, percent forest, percent impervious, and waterbody area emerged as the strongest predictors of cyanobacteria cell density with an adjusted R-squared value of 0.31 and a p-value ~ 0. The final regression equation was used to make a normalized cyanobacteria cell density index, and a Jenks Natural Break classification was used to assign waterbodies designations of low, medium, or high risk. Of the 771 waterbodies, 24.38% were low risk, 37.35% were medium risk, and 38.26% were high risk. This study showed that there are significant relationships between free geospatial datasets representing summer maximum temperatures, nutrient loading associated with land use and land cover, and the area of a waterbody with cyanobacteria cell density. This data analytics approach to CyanoHAB risk assessment corroborated the literature-established environmental triggers for CyanoHABs, and presents a novel approach for CyanoHAB risk mapping in waterbodies across the greater southeastern United States.Keywords: cyanobacteria, land use/land cover, remote sensing, risk mapping
Procedia PDF Downloads 21175 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation
Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong
Abstract:
Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation
Procedia PDF Downloads 19074 Reviving the Past, Enhancing the Future: Preservation of Urban Heritage Connectivity as a Tool for Developing Liveability in Historical Cities in Jordan, Using Salt City as a Case Study
Authors: Sahar Yousef, Chantelle Niblock, Gul Kacmaz
Abstract:
Salt City, in the context of Jordan’s heritage landscape, is a significant case to explore when it comes to the interaction between tangible and intangible qualities of liveable cities. Most city centers, including Jerash, Salt, Irbid, and Amman, are historical locations. Six of these extraordinary sites were designated UNESCO World Heritage Sites. Jordan is widely acknowledged as a developing country characterized by swift urbanization and unrestrained expansion that exacerbate the challenges associated with the preservation of historic urban areas. The aim of this study is to conduct an examination and analysis of the existing condition of heritage connectivity within heritage city centers. This includes outdoor staircases, pedestrian pathways, footpaths, and other public spaces. Case study-style analysis of the urban core of As-Salt is the focus of this investigation. Salt City is widely acknowledged for its substantial tangible and intangible cultural heritage and has been designated as ‘The Place of Tolerance and Urban Hospitality’ by UNESCO since 2021. Liveability in urban heritage, particularly in historic city centers, incorporates several factors that affect our well-being; its enhancement is a critical issue in contemporary society. The dynamic interaction between humans and historical materials, which serves as a vehicle for the expression of their identity and historical narrative, constitutes preservation that transcends simple conservation. This form of engagement enables people to appreciate the diversity of their heritage recognising their previous and planned futures. Heritage preservation is inextricably linked to a larger physical and emotional context; therefore, it is difficult to examine it in isolation. Urban environments, including roads, structures, and other infrastructure, are undergoing unprecedented physical design and construction requirements. Concurrently, heritage reinforces a sense of affiliation with a particular location or space and unifies individuals with their ancestry, thereby defining their identity. However, a considerable body of research has focused on the conservation of heritage buildings in a fragmented manner without considering their integration within a holistic urban context. Insufficient attention is given to the significance of the physical and social roles played by the heritage staircases and baths that serve as connectors between these valued historical buildings. In doing so, the research uses a methodology that is based on consensus. Given that liveability is considered a complex matter with several dimensions. The discussion starts by making initial observations on the physical context and societal norms inside the urban center while simultaneously establishing the definitions of liveability and connectivity and examining the key criteria associated with these concepts. Then, identify the key elements that contribute to liveable connectivity within the framework of urban heritage in Jordanian city centers. Some of the outcomes that will be discussed in the presentation are: (1) There is not enough connectivity between heritage buildings as can be seen, for example, between buildings in Jada and Qala'. (2) Most of the outdoor spaces suffer from physical issues that hinder their use by the public, like in Salalem. (3) Existing activities in the city center are not well attended because of lack of communication between the organisers and the citizens.Keywords: connectivity, Jordan, liveability, salt city, tangible and intangible heritage, urban heritage
Procedia PDF Downloads 7073 Simultech - Innovative Country-Wide Ultrasound Training Center
Authors: Yael Rieder, Yael Gilboa, S. O. Adva, Efrat Halevi, Ronnie Tepper
Abstract:
Background: Operation of ultrasound equipment is a core skill for many clinical specialties. As part of the training program at -Simultech- a simulation center for Ob\Gyn at the Meir Medical Center, Israel, teaching how to operate ultrasound equipment requires dealing with misunderstandings of spatial and 3D orientation, failure of the operator to hold a transducer correctly, and limited ability to evaluate the data on the screen. We have developed a platform intended to endow physicians and sonographers with clinical and operational skills of obstetric ultrasound. Simultech's simulations are focused on medical knowledge, risk management, technology operations and physician-patient communication. The simulations encompass extreme work conditions. Setup: Between eight and ten of the eight hundred and fifty physicians and sonographers of the Clalit health services from seven hospitals and eight community centers across Israel, participate in individual Ob/Gyn training sessions each week. These include Ob/Gyn specialists, experts, interns, and sonographers. Innovative teaching and training methodologies: The six-hour training program includes: (1) An educational computer program that challenges trainees to deal with medical questions based upon ultrasound pictures and films. (2) Sophisticated hands-on simulators that challenge the trainees to practice correct grip of the transducer, elucidate pathology, and practice daily tasks such as biometric measurements and analysis of sonographic data. (3) Participation in a video-taped simulation which focuses on physician-patient communications. In the simulation, the physician is required to diagnose the clinical condition of a hired actress based on the data she provides and by evaluating the assigned ultrasound films accordingly. Giving ‘bad news’ to the patient may put the physician in a stressful situation that must be properly managed. (4) Feedback at the end of each phase is provided by a designated trainer, not a physician, who is specially qualified by Ob\Gyn senior specialists. (5) A group exercise in which the trainer presents a medico-legal case in order to encourage the participants to use their own experience and knowledge to conduct a productive ‘brainstorming’ session. Medical cases are presented and analyzed by the participants together with the trainer's feedback. Findings: (1) The training methods and content that Simultech provides allows trainees to review their medical and communications skills. (2) Simultech training sessions expose physicians to both basic and new, up-to-date cases, refreshing and expanding the trainee's knowledge. (3) Practicing on advanced simulators enables trainees to understand the sonographic space and to implement the basic principles of ultrasound. (4) Communications simulations were found to be beneficial for trainees who were unaware of their interpersonal skills. The trainer feedback, supported by the recorded simulation, allows the trainee to draw conclusions about his performance. Conclusion: Simultech was found to contribute to physicians at all levels of clinical expertise who deal with ultrasound. A break in daily routine together with attendance at a neutral educational center can vastly improve performance and outlook.Keywords: medical training, simulations, ultrasound, Simultech
Procedia PDF Downloads 28072 Analysis of Composite Health Risk Indicators Built at a Regional Scale and Fine Resolution to Detect Hotspot Areas
Authors: Julien Caudeville, Muriel Ismert
Abstract:
Analyzing the relationship between environment and health has become a major preoccupation for public health as evidenced by the emergence of the French national plans for health and environment. These plans have identified the following two priorities: (1) to identify and manage geographic areas, where hotspot exposures are suspected to generate a potential hazard to human health; (2) to reduce exposure inequalities. At a regional scale and fine resolution of exposure outcome prerequisite, environmental monitoring networks are not sufficient to characterize the multidimensionality of the exposure concept. In an attempt to increase representativeness of spatial exposure assessment approaches, risk composite indicators could be built using additional available databases and theoretical framework approaches to combine factor risks. To achieve those objectives, combining data process and transfer modeling with a spatial approach is a fundamental prerequisite that implies the need to first overcome different scientific limitations: to define interest variables and indicators that could be built to associate and describe the global source-effect chain; to link and process data from different sources and different spatial supports; to develop adapted methods in order to improve spatial data representativeness and resolution. A GIS-based modeling platform for quantifying human exposure to chemical substances (PLAINE: environmental inequalities analysis platform) was used to build health risk indicators within the Lorraine region (France). Those indicators combined chemical substances (in soil, air and water) and noise risk factors. Tools have been developed using modeling, spatial analysis and geostatistic methods to build and discretize interest variables from different supports and resolutions on a 1 km2 regular grid within the Lorraine region. By example, surface soil concentrations have been estimated by developing a Kriging method able to integrate surface and point spatial supports. Then, an exposure model developed by INERIS was used to assess the transfer from soil to individual exposure through ingestion pathways. We used distance from polluted soil site to build a proxy for contaminated site. Air indicator combined modeled concentrations and estimated emissions to take in account 30 polluants in the analysis. For water, drinking water concentrations were compared to drinking water standards to build a score spatialized using a distribution unit serve map. The Lden (day-evening-night) indicator was used to map noise around road infrastructures. Aggregation of the different factor risks was made using different methodologies to discuss weighting and aggregation procedures impact on the effectiveness of risk maps to take decisions for safeguarding citizen health. Results permit to identify pollutant sources, determinants of exposure, and potential hotspots areas. A diagnostic tool was developed for stakeholders to visualize and analyze the composite indicators in an operational and accurate manner. The designed support system will be used in many applications and contexts: (1) mapping environmental disparities throughout the Lorraine region; (2) identifying vulnerable population and determinants of exposure to set priorities and target for pollution prevention, regulation and remediation; (3) providing exposure database to quantify relationships between environmental indicators and cancer mortality data provided by French Regional Health Observatories.Keywords: health risk, environment, composite indicator, hotspot areas
Procedia PDF Downloads 24771 Environmental Impacts of Point and Non-Point Source Pollution in Krishnagiri Reservoir: A Case Study in South India
Authors: N. K. Ambujam, V. Sudha
Abstract:
Reservoirs are being contaminated all around the world with point source and Non-Point Source (NPS) pollution. The most common NPS pollutants are sediments and nutrients. Krishnagiri Reservoir (KR) has been chosen for the present case study, which is located in the tropical semi-arid climatic zone of Tamil Nadu, South India. It is the main source of surface water in Krishnagiri district to meet the freshwater demands. The reservoir has lost about 40% of its water holding capacity due to sedimentation over the period of 50 years. Hence, from the research and management perspective, there is a need for a sound knowledge on the spatial and seasonal variations of KR water quality. The present study encompasses the specific objectives as (i) to investigate the longitudinal heterogeneity and seasonal variations of physicochemical parameters, nutrients and biological characteristics of KR water and (ii) to examine the extent of degradation of water quality in KR. 15 sampling points were identified by uniform stratified method and a systematic monthly sampling strategy was selected due to high dynamic nature in its hydrological characteristics. The physicochemical parameters, major ions, nutrients and Chlorophyll a (Chl a) were analysed. Trophic status of KR was classified by using Carlson's Trophic State Index (TSI). All statistical analyses were performed by using Statistical Package for Social Sciences programme, version-16.0. Spatial maps were prepared for Chl a using Arc GIS. Observations in KR pointed out that electrical conductivity and major ions are highly variable factors as it receives inflow from the catchment with different land use activities. The study of major ions in KR exhibited different trends in their values and it could be concluded that as the monsoon progresses the major ions in the water decreases or water quality stabilizes. The inflow point of KR showed comparatively higher concentration of nutrients including nitrate, soluble reactive phosphorus (SRP), total phosphors (TP), total suspended phosphorus (TSP) and total dissolved phosphorus (TDP) during monsoon seasons. This evidently showed the input of significant amount of nutrients from the catchment side through agricultural runoff. High concentration of TDP and TSP at the lacustrine zone of the reservoir during summer season evidently revealed that there was a significant release of phosphorus from the bottom sediments. Carlson’s TSI of KR ranged between 81 and 92 during northeast monsoon and summer seasons. High and permanent Cyanobacterial bloom in KR could be mainly due to the internal loading of phosphorus from the bottom sediments. According to Carlson’s TSI classification Krishnagiri reservoir was ranked in the hyper-eutrophic category. This study provides necessary basic data on the spatio-temporal variations of water quality in KR and also proves the impact of point and NPS pollution from the catchment area. High TSI warrants a greater threat for the recovery of internal P loading and hyper-eutrophic condition of KR. Several expensive internal measures for the reduction of internal loading of P were introduced by many scientists. However, the outcome of the present research suggests for the innovative algae harvesting technique for the removal of sediment nutrients.Keywords: NPS pollution, nutrients, hyper-eutrophication, krishnagiri reservoir
Procedia PDF Downloads 324