Search results for: spatial metrics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2941

Search results for: spatial metrics

511 Lateralisation of Visual Function in Yellow-Eyed Mullet (Aldrichetta forsteri) and Its Role in Schooling Behaviour

Authors: Karen L. Middlemiss, Denham G. Cook, Peter Jaksons, Alistair Jerrett, William Davison

Abstract:

Lateralisation of cognitive function is a common phenomenon found throughout the animal kingdom. Strong biases in functional behaviours have evolved from asymmetrical brain hemispheres which differ in structure and/or cognitive function. In fish, lateralisation is involved in visually mediated behaviours such as schooling, predator avoidance, and foraging, and is considered to have a direct impact on species fitness. Currently, there is very little literature on the role of lateralisation in fish schools. The yellow-eyed mullet (Aldrichetta forsteri), is an estuarine and coastal species found commonly throughout temperate regions of Australia and New Zealand. This study sought to quantify visually mediated behaviours in yellow-eyed mullet to identify the significance of lateralisation, and the factors which influence functional behaviours in schooling fish. Our approach to study design was to conduct a series of tank based experiments investigating; a) individual and population level lateralisation, b) schooling behaviour, and d) optic lobe anatomy. Yellow-eyed mullet showed individual variation in direction and strength of lateralisation in juveniles, and trait specific spatial positioning within the school was evidenced in strongly lateralised fish. In combination with observed differences in schooling behaviour, the possibility of ontogenetic plasticity in both behavioural lateralisation and optic lobe morphology in adults is suggested. These findings highlight the need for research into the genetic and environmental factors (epigenetics) which drive functional behaviours such as schooling, feeding and aggression. Improved knowledge on collective behaviour could have significant benefits to captive rearing programmes through improved culture techniques and will add to the limited body of knowledge on the complex ecophysiological interactions present in our inshore fisheries.

Keywords: cerebral asymmetry, fisheries, schooling, visual bias

Procedia PDF Downloads 209
510 Optical Flow Technique for Supersonic Jet Measurements

Authors: Haoxiang Desmond Lim, Jie Wu, Tze How Daniel New, Shengxian Shi

Abstract:

This paper outlines the development of a novel experimental technique in quantifying supersonic jet flows, in an attempt to avoid seeding particle problems frequently associated with particle-image velocimetry (PIV) techniques at high Mach numbers. Based on optical flow algorithms, the idea behind the technique involves using high speed cameras to capture Schlieren images of the supersonic jet shear layers, before they are subjected to an adapted optical flow algorithm based on the Horn-Schnuck method to determine the associated flow fields. The proposed method is capable of offering full-field unsteady flow information with potentially higher accuracy and resolution than existing point-measurements or PIV techniques. Preliminary study via numerical simulations of a circular de Laval jet nozzle successfully reveals flow and shock structures typically associated with supersonic jet flows, which serve as useful data for subsequent validation of the optical flow based experimental results. For experimental technique, a Z-type Schlieren setup is proposed with supersonic jet operated in cold mode, stagnation pressure of 8.2 bar and exit velocity of Mach 1.5. High-speed single-frame or double-frame cameras are used to capture successive Schlieren images. As implementation of optical flow technique to supersonic flows remains rare, the current focus revolves around methodology validation through synthetic images. The results of validation test offers valuable insight into how the optical flow algorithm can be further improved to improve robustness and accuracy. Details of the methodology employed and challenges faced will be further elaborated in the final conference paper should the abstract be accepted. Despite these challenges however, this novel supersonic flow measurement technique may potentially offer a simpler way to identify and quantify the fine spatial structures within the shock shear layer.

Keywords: Schlieren, optical flow, supersonic jets, shock shear layer

Procedia PDF Downloads 310
509 Trend Analysis of Rainfall: A Climate Change Paradigm

Authors: Shyamli Singh, Ishupinder Kaur, Vinod K. Sharma

Abstract:

Climate Change refers to the change in climate for extended period of time. Climate is changing from the past history of earth but anthropogenic activities accelerate this rate of change and which is now being a global issue. Increase in greenhouse gas emissions is causing global warming and climate change related issues at an alarming rate. Increasing temperature results in climate variability across the globe. Changes in rainfall patterns, intensity and extreme events are some of the impacts of climate change. Rainfall variability refers to the degree to which rainfall patterns varies over a region (spatial) or through time period (temporal). Temporal rainfall variability can be directly or indirectly linked to climate change. Such variability in rainfall increases the vulnerability of communities towards climate change. Increasing urbanization and unplanned developmental activities, the air quality is deteriorating. This paper mainly focuses on the rainfall variability due to increasing level of greenhouse gases. Rainfall data of 65 years (1951-2015) of Safdarjung station of Delhi was collected from Indian Meteorological Department and analyzed using Mann-Kendall test for time-series data analysis. Mann-Kendall test is a statistical tool helps in analysis of trend in the given data sets. The slope of the trend can be measured through Sen’s slope estimator. Data was analyzed monthly, seasonally and yearly across the period of 65 years. The monthly rainfall data for the said period do not follow any increasing or decreasing trend. Monsoon season shows no increasing trend but here was an increasing trend in the pre-monsoon season. Hence, the actual rainfall differs from the normal trend of the rainfall. Through this analysis, it can be projected that there will be an increase in pre-monsoon rainfall than the actual monsoon season. Pre-monsoon rainfall causes cooling effect and results in drier monsoon season. This will increase the vulnerability of communities towards climate change and also effect related developmental activities.

Keywords: greenhouse gases, Mann-Kendall test, rainfall variability, Sen's slope

Procedia PDF Downloads 199
508 Supporting Regulation and Shared Attention to Facilitate the Foundations for Development of Children and Adolescents with Complex Individual Profiles

Authors: Patsy Tan, Dana Baltutis

Abstract:

This presentation demonstrates the effectiveness of music therapy in co-treatment with speech pathology and occupational therapy as an innovative way when working with children and adolescents with complex individual differences to facilitate communication, emotional, motor and social skills development. Each child with special needs and their carer has an individual profile which encompasses their visual-spatial, auditory, language, learning, mental health, family dynamic, sensory-motor, motor planning and sequencing profiles. The most common issues among children with special needs, especially those diagnosed with Autism Spectrum Disorder, are in the areas of regulation, communication, and social-emotional development. The ability of children living with challenges to communicate and use language and understand verbal and non-verbal information, as well as move their bodies to explore and interact with their environments in social situations, depends on the children being regulated both internally and externally and trusting their communication partners and understanding what is happening in the moment. For carers, it is about understanding the tempo, rhythm, pacing, and timing of their own individual profile, as well as the profile of the child they are interacting with, and how these can sync together. In this study, music therapy is used in co-treatment sessions with a speech pathologist and/or an occupational therapist using the DIRFloortime approach to facilitate the regulation, attention, engagement, reciprocity and social-emotional capacities of children presenting with complex individual differences. Documented changes in 10 domains of children’s development over a 12-month period using the Individual Music Therapy Assessment Profile (IMTAP) were observed. Children were assessed biannually, and results show significant improvements in the social-emotional, musicality and receptive language domains indicating that co-treatment with a music therapist using the DIRFloortime framework is highly effective. This presentation will highlight strategies that facilitate regulation, social-emotional and communication development for children and adolescents with complex individual profiles.

Keywords: communication, shared attention, regulation, social emotional

Procedia PDF Downloads 251
507 High School Gain Analytics From National Assessment Program – Literacy and Numeracy and Australian Tertiary Admission Rankin Linkage

Authors: Andrew Laming, John Hattie, Mark Wilson

Abstract:

Nine Queensland Independent high schools provided deidentified student-matched ATAR and NAPLAN data for all 1217 ATAR graduates since 2020 who also sat NAPLAN at the school. Graduating cohorts from the nine schools contained a mean 100 ATAR graduates with previous NAPLAN data from their school. Excluded were vocational students (mean=27) and any ATAR graduates without NAPLAN data (mean=20). Based on Index of Community Socio-Educational Access (ICSEA) prediction, all schools had larger that predicted proportions of their students graduating with ATARs. There were an additional 173 students not releasing their ATARs to their school (14%), requiring this data to be inferred by schools. Gain was established by first converting each student’s strongest NAPLAN domain to a statewide percentile, then subtracting this result from final ATAR. The resulting ‘percentile shift’ was corrected for plausible ATAR participation at each NAPLAN level. Strongest NAPLAN domain had the highest correlation with ATAR (R2=0.58). RESULTS School mean NAPLAN scores fitted ICSEA closely (R2=0.97). Schools achieved a mean cohort gain of two ATAR rankings, but only 66% of students gained. This ranged from 46% of top-NAPLAN decile students gaining, rising to 75% achieving gains outside the top decile. The 54% of top-decile students whose ATAR fell short of prediction lost a mean 4.0 percentiles (or 6.2 percentiles prior to correction for regression to the mean). 71% of students in smaller schools gained, compared to 63% in larger schools. NAPLAN variability in each of the 13 ICSEA1100 cohorts was 17%, with both intra-school and inter-school variation of these values extremely low (0.3% to 1.8%). Mean ATAR change between years in each school was just 1.1 ATAR ranks. This suggests consecutive school cohorts and ICSEA-similar schools share very similar distributions and outcomes over time. Quantile analysis of the NAPLAN/ATAR revealed heteroscedasticity, but splines offered little additional benefit over simple linear regression. The NAPLAN/ATAR R2 was 0.33. DISCUSSION Standardised data like NAPLAN and ATAR offer educators a simple no-cost progression metric to analyse performance in conjunction with their internal test results. Change is expressed in percentiles, or ATAR shift per student, which is layperson intuitive. Findings may also reduce ATAR/vocational stream mismatch, reveal proportions of cohorts meeting or falling short of expectation and demonstrate by how much. Finally, ‘crashed’ ATARs well below expectation are revealed, which schools can reasonably work to minimise. The percentile shift method is neither value-add nor a growth percentile. In the absence of exit NAPLAN testing, this metric is unable to discriminate academic gain from legitimate ATAR-maximizing strategies. But by controlling for ICSEA, ATAR proportion variation and student mobility, it uncovers progression to ATAR metrics which are not currently publicly available. However achieved, ATAR maximisation is a sought-after private good. So long as standardised nationwide data is available, this analysis offers useful analytics for educators and reasonable predictivity when counselling subsequent cohorts about their ATAR prospects.  

Keywords: NAPLAN, ATAR, analytics, measurement, gain, performance, data, percentile, value-added, high school, numeracy, reading comprehension, variability, regression to the mean

Procedia PDF Downloads 62
506 Healthcare-SignNet: Advanced Video Classification for Medical Sign Language Recognition Using CNN and RNN Models

Authors: Chithra A. V., Somoshree Datta, Sandeep Nithyanandan

Abstract:

Sign Language Recognition (SLR) is the process of interpreting and translating sign language into spoken or written language using technological systems. It involves recognizing hand gestures, facial expressions, and body movements that makeup sign language communication. The primary goal of SLR is to facilitate communication between hearing- and speech-impaired communities and those who do not understand sign language. Due to the increased awareness and greater recognition of the rights and needs of the hearing- and speech-impaired community, sign language recognition has gained significant importance over the past 10 years. Technological advancements in the fields of Artificial Intelligence and Machine Learning have made it more practical and feasible to create accurate SLR systems. This paper presents a distinct approach to SLR by framing it as a video classification problem using Deep Learning (DL), whereby a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) has been used. This research targets the integration of sign language recognition into healthcare settings, aiming to improve communication between medical professionals and patients with hearing impairments. The spatial features from each video frame are extracted using a CNN, which captures essential elements such as hand shapes, movements, and facial expressions. These features are then fed into an RNN network that learns the temporal dependencies and patterns inherent in sign language sequences. The INCLUDE dataset has been enhanced with more videos from the healthcare domain and the model is evaluated on the same. Our model achieves 91% accuracy, representing state-of-the-art performance in this domain. The results highlight the effectiveness of treating SLR as a video classification task with the CNN-RNN architecture. This approach not only improves recognition accuracy but also offers a scalable solution for real-time SLR applications, significantly advancing the field of accessible communication technologies.

Keywords: sign language recognition, deep learning, convolution neural network, recurrent neural network

Procedia PDF Downloads 14
505 Modelling Dengue Disease With Climate Variables Using Geospatial Data For Mekong River Delta Region of Vietnam

Authors: Thi Thanh Nga Pham, Damien Philippon, Alexis Drogoul, Thi Thu Thuy Nguyen, Tien Cong Nguyen

Abstract:

Mekong River Delta region of Vietnam is recognized as one of the most vulnerable to climate change due to flooding and seawater rise and therefore an increased burden of climate change-related diseases. Changes in temperature and precipitation are likely to alter the incidence and distribution of vector-borne diseases such as dengue fever. In this region, the peak of the dengue epidemic period is around July to September during the rainy season. It is believed that climate is an important factor for dengue transmission. This study aims to enhance the capacity of dengue prediction by the relationship of dengue incidences with climate and environmental variables for Mekong River Delta of Vietnam during 2005-2015. Mathematical models for vector-host infectious disease, including larva, mosquito, and human being were used to calculate the impacts of climate to the dengue transmission with incorporating geospatial data for model input. Monthly dengue incidence data were collected at provincial level. Precipitation data were extracted from satellite observations of GSMaP (Global Satellite Mapping of Precipitation), land surface temperature and land cover data were from MODIS. The value of seasonal reproduction number was estimated to evaluate the potential, severity and persistence of dengue infection, while the final infected number was derived to check the outbreak of dengue. The result shows that the dengue infection depends on the seasonal variation of climate variables with the peak during the rainy season and predicted dengue incidence follows well with this dynamic for the whole studied region. However, the highest outbreak of 2007 dengue was not captured by the model reflecting nonlinear dependences of transmission on climate. Other possible effects will be discussed to address the limitation of the model. This suggested the need of considering of both climate variables and another variability across temporal and spatial scales.

Keywords: infectious disease, dengue, geospatial data, climate

Procedia PDF Downloads 374
504 Enhancing Early Detection of Coronary Heart Disease Through Cloud-Based AI and Novel Simulation Techniques

Authors: Md. Abu Sufian, Robiqul Islam, Imam Hossain Shajid, Mahesh Hanumanthu, Jarasree Varadarajan, Md. Sipon Miah, Mingbo Niu

Abstract:

Coronary Heart Disease (CHD) remains a principal cause of global morbidity and mortality, characterized by atherosclerosis—the build-up of fatty deposits inside the arteries. The study introduces an innovative methodology that leverages cloud-based platforms like AWS Live Streaming and Artificial Intelligence (AI) to early detect and prevent CHD symptoms in web applications. By employing novel simulation processes and AI algorithms, this research aims to significantly mitigate the health and societal impacts of CHD. Methodology: This study introduces a novel simulation process alongside a multi-phased model development strategy. Initially, health-related data, including heart rate variability, blood pressure, lipid profiles, and ECG readings, were collected through user interactions with web-based applications as well as API Integration. The novel simulation process involved creating synthetic datasets that mimic early-stage CHD symptoms, allowing for the refinement and training of AI algorithms under controlled conditions without compromising patient privacy. AWS Live Streaming was utilized to capture real-time health data, which was then processed and analysed using advanced AI techniques. The novel aspect of our methodology lies in the simulation of CHD symptom progression, which provides a dynamic training environment for our AI models enhancing their predictive accuracy and robustness. Model Development: it developed a machine learning model trained on both real and simulated datasets. Incorporating a variety of algorithms including neural networks and ensemble learning model to identify early signs of CHD. The model's continuous learning mechanism allows it to evolve adapting to new data inputs and improving its predictive performance over time. Results and Findings: The deployment of our model yielded promising results. In the validation phase, it achieved an accuracy of 92% in predicting early CHD symptoms surpassing existing models. The precision and recall metrics stood at 89% and 91% respectively, indicating a high level of reliability in identifying at-risk individuals. These results underscore the effectiveness of combining live data streaming with AI in the early detection of CHD. Societal Implications: The implementation of cloud-based AI for CHD symptom detection represents a significant step forward in preventive healthcare. By facilitating early intervention, this approach has the potential to reduce the incidence of CHD-related complications, decrease healthcare costs, and improve patient outcomes. Moreover, the accessibility and scalability of cloud-based solutions democratize advanced health monitoring, making it available to a broader population. This study illustrates the transformative potential of integrating technology and healthcare, setting a new standard for the early detection and management of chronic diseases.

Keywords: coronary heart disease, cloud-based ai, machine learning, novel simulation techniques, early detection, preventive healthcare

Procedia PDF Downloads 57
503 Phenology and Size in the Social Sweat Bee, Halictus ligatus, in an Urban Environment

Authors: Rachel A. Brant, Grace E. Kenny, Paige A. Muñiz, Gerardo R. Camilo

Abstract:

The social sweat bee, Halictus ligatus, has been documented to alter its phenology as a response to changes in temporal dynamics of resources. Furthermore, H. ligatus exhibits polyethism in natural environments as a consequence of the variation in resources. Yet, we do not know if or how H. ligatus responds to these variations in urban environments. As urban environments become much more widespread, and human population is expected to reach nine billion by 2050, it is crucial to distinguish how resources are allocated by bees in cities. We hypothesize that in urban regions, where floral availability varies with human activity, H. ligatus will exhibit polyethism in order to match the extremely localized spatial variability of resources. We predict that in an urban setting, where resources vary both spatially and temporally, the phenology of H. ligatus will alter in response to these fluctuations. This study was conducted in Saint Louis, Missouri, at fifteen sites each varying in size and management type (community garden, urban farm, prairie restoration). Bees were collected by hand netting from 2013-2016. Results suggest that the largest individuals, mostly gynes, occurred in lower income neighborhood community gardens in May and August. We used a model averaging procedure, based on information theoretical methods, to determine a best model for predicting bee size. Our results suggest that month and locality within the city are the best predictors of bee size. Halictus ligatus was observed to comply with the predictions of polyethism from 2013 to 2015. However, in 2016 there was an almost complete absence of the smallest worker castes. This is a significant deviation from what is expected under polyethism. This could be attributed to shifts in planting decisions, shifts in plant-pollinator matches, or local climatic conditions. Further research is needed to determine if this divergence from polyethism is a new strategy for the social sweat bee as climate continues to alter or a response to human dominated landscapes.

Keywords: polyethism, urban environment, phenology, social sweat bee

Procedia PDF Downloads 213
502 The Integration of Geographical Information Systems and Capacitated Vehicle Routing Problem with Simulated Demand for Humanitarian Logistics in Tsunami-Prone Area: A Case Study of Phuket, Thailand

Authors: Kiatkulchai Jitt-Aer, Graham Wall, Dylan Jones

Abstract:

As a result of the Indian Ocean tsunami in 2004, logistics applied to disaster relief operations has received great attention in the humanitarian sector. As learned from such disaster, preparing and responding to the aspect of delivering essential items from distribution centres to affected locations are of the importance for relief operations as the nature of disasters is uncertain especially in suffering figures, which are normally proportional to quantity of supplies. Thus, this study proposes a spatial decision support system (SDSS) for humanitarian logistics by integrating Geographical Information Systems (GIS) and the capacitated vehicle routing problem (CVRP). The GIS is utilised for acquiring demands simulated from the tsunami flooding model of the affected area in the first stage, and visualising the simulation solutions in the last stage. While CVRP in this study encompasses designing the relief routes of a set of homogeneous vehicles from a relief centre to a set of geographically distributed evacuation points in which their demands are estimated by using both simulation and randomisation techniques. The CVRP is modeled as a multi-objective optimization problem where both total travelling distance and total transport resources used are minimized, while demand-cost efficiency of each route is maximized in order to determine route priority. As the model is a NP-hard combinatorial optimization problem, the Clarke and Wright Saving heuristics is proposed to solve the problem for the near-optimal solutions. The real-case instances in the coastal area of Phuket, Thailand are studied to perform the SDSS that allows a decision maker to visually analyse the simulation scenarios through different decision factors.

Keywords: demand simulation, humanitarian logistics, geographical information systems, relief operations, capacitated vehicle routing problem

Procedia PDF Downloads 242
501 Computational Modelling of Epoxy-Graphene Composite Adhesive towards the Development of Cryosorption Pump

Authors: Ravi Verma

Abstract:

Cryosorption pump is the best solution to achieve clean, vibration free ultra-high vacuum. Furthermore, the operation of cryosorption pump is free from the influence of electric and magnetic fields. Due to these attributes, this pump is used in the space simulation chamber to create the ultra-high vacuum. The cryosorption pump comprises of three parts (a) panel which is cooled with the help of cryogen or cryocooler, (b) an adsorbent which is used to adsorb the gas molecules, (c) an epoxy which holds the adsorbent and the panel together thereby aiding in heat transfer from adsorbent to the panel. The performance of cryosorption pump depends on the temperature of the adsorbent and hence, on the thermal conductivity of the epoxy. Therefore we have made an attempt to increase the thermal conductivity of epoxy adhesive by mixing nano-sized graphene filler particles. The thermal conductivity of epoxy-graphene composite adhesive is measured with the help of indigenously developed experimental setup in the temperature range from 4.5 K to 7 K, which is generally the operating temperature range of cryosorption pump for efficiently pumping of hydrogen and helium gas. In this article, we have presented the experimental results of epoxy-graphene composite adhesive in the temperature range from 4.5 K to 7 K. We have also proposed an analytical heat conduction model to find the thermal conductivity of the composite. In this case, the filler particles, such as graphene, are randomly distributed in a base matrix of epoxy. The developed model considers the complete spatial random distribution of filler particles and this distribution is explained by Binomial distribution. The results obtained by the model have been compared with the experimental results as well as with the other established models. The developed model is able to predict the thermal conductivity in both isotropic regions as well as in anisotropic region over the required temperature range from 4.5 K to 7 K. Due to the non-empirical nature of the proposed model, it will be useful for the prediction of other properties of composite materials involving the filler in a base matrix. The present studies will aid in the understanding of low temperature heat transfer which in turn will be useful towards the development of high performance cryosorption pump.

Keywords: composite adhesive, computational modelling, cryosorption pump, thermal conductivity

Procedia PDF Downloads 85
500 Worldbuilding as Critical Architectural Pedagogy

Authors: Jesse Rafeiro

Abstract:

This paper discusses worldbuilding as a pedagogical approach to the first-year architectural design studio. The studio ran for three consecutive terms between 2016-2018. Taking its departure from the fifty-five city narratives in Italo Calvino’s Invisible Cities, students collectively designed in a “nowhere” space where intersecting and diverging narratives could be played out. Along with Calvino, students navigated between three main exercises and their imposed limits to develop architectural insight at three scales simulating the considerations of architectural practice: detail, building, and city. The first exercise asked each student to design and model a ruin based on randomly assigned incongruent fragments. Each student was given one plan fragment and two section fragments from different Renaissance Treatises. The students were asked to translate these in alternating axonometric projection and model-making explorations. Although the fragments themselves were imposed, students were free to interpret how the drawings fit together by imagining new details and atypical placements. An undulating terrain model was introduced in the second exercise to ground the worldbuilding exercises. Here, students were required to negotiate with one another to design a city of ruins. Free to place their models anywhere on the site, the students were restricted by the negotiation of territories marked by other students and the requirement to provide thresholds, open spaces, and corridors. The third exercise introduced new life into the ruined city through a series of design interventions. Each student was assigned an atypical building program suggesting a place for an activity, human or nonhuman. The atypical nature of the programs challenged the triviality of functional planning through explorations in spatial narratives free from preconceived assumptions. By contesting, playing out, or dreaming responses to realities taught in other coursework, this third exercise actualized learnings that are too often self-contained in the silos of differing course agendas. As such, the studio fostered an initial worldbuilding space within which to sharpen sensibility and criticality for subsequent years of education.

Keywords: architectural pedagogy, critical pedagogy, Italo Calvino, worldbuilding

Procedia PDF Downloads 126
499 Impure Water, a Future Disaster: A Case Study of Lahore Ground Water Quality with GIS Techniques

Authors: Rana Waqar Aslam, Urooj Saeed, Hammad Mehmood, Hameed Ullah, Imtiaz Younas

Abstract:

This research has been conducted to assess the water quality in and around Lahore Metropolitan area on the basis of three different land uses, i.e. residential, commercial, and industrial land uses. For this, 29 sample sites have been selected on the basis of simple random sampling technique. Samples were collected at the source (WASA tube wells). The criteria for selecting sample sites are to have a maximum concentration of population in the selected land uses. The results showed that in the residential land use the proportion of nitrate and turbidity is at their highest level in the areas of Allama Iqbal Town and Samanabad Town. Commercial land use of Gulberg and Data Gunj Bakhsh Town have highest level of proportion of chlorides, calcium, TDS, pH, Mg, total hardness, arsenic and alkalinity. Whereas in industrial type of land use in Ravi and Wahga Town have the proportion of arsenic, Mg, nitrate, pH, and turbidity are at their highest level. The high rate of concentration of these parameters in these areas is basically due to the old and fractured pipelines that allow bacterial as well as physiochemical contaminants to contaminate the portable water at the sources. Furthermore, it is seen in most areas that waste water from domestic, industrial, as well as municipal sources may get easy discharge into open spaces and water bodies, like, cannels, rivers, lakes that seeps and become a part of ground water. In addition, huge dumps located in Lahore are becoming the cause of ground water contamination as when the rain falls, the water gets seep into the ground and impures the ground water quality. On the basis of the derived results with the help of Geo-spatial technology ACRGIS 9.3 Interpolation (IDW), it is recommended that water filtration plants must be installed with specific parameter control. A separate team for proper inspection has to be made for water quality check at the source. Old water pipelines must be replaced with the new pipelines, and safe water depth must be ensured at the source end.

Keywords: GIS, remote sensing, pH, nitrate, disaster, IDW

Procedia PDF Downloads 221
498 Connotation Reform and Problem Response of Rural Social Relations under the Influence of the Earthquake: With a Review of Wenchuan Decade

Authors: Yanqun Li, Hong Geng

Abstract:

The occurrence of Wenchuan earthquake in 2008 has led to severe damage to the rural areas of Chengdu city, such as the rupture of the social network, the stagnation of economic production and the rupture of living space. The post-disaster reconstruction has become a sustainable issue. As an important link to maintain the order of rural social development, social network should be an important content of post-disaster reconstruction. Therefore, this paper takes rural reconstruction communities in earthquake-stricken areas of Chengdu as the research object and adopts sociological research methods such as field survey, observation and interview to try to understand the transformation of rural social relations network under the influence of earthquake and its impact on rural space. It has found that rural societies under the earthquake generally experienced three phases: the break of stable social relations, the transition of temporary non-normal state, and the reorganization of social networks. The connotation of phased rural social relations also changed accordingly: turn to a new division of labor on the social orientation, turn to a capital flow and redistribution in new production mode on the capital orientation, and turn to relative decentralization after concentration on the spatial dimension. Along with such changes, rural areas have emerged some social issues such as the alienation of competition in the new industry division, the low social connection, the significant redistribution of capital, and the lack of public space. Based on a comprehensive review of these issues, this paper proposes the corresponding response mechanism. First of all, a reasonable division of labor should be established within the villages to realize diversified commodity supply. Secondly, the villages should adjust the industrial type to promote the equitable participation of capital allocation groups. Finally, external public spaces should be added to strengthen the field of social interaction within the communities.

Keywords: social relations, social support networks, industrial division, capital allocation, public space

Procedia PDF Downloads 151
497 3D Codes for Unsteady Interaction Problems of Continuous Mechanics in Euler Variables

Authors: M. Abuziarov

Abstract:

The designed complex is intended for the numerical simulation of fast dynamic processes of interaction of heterogeneous environments susceptible to the significant formability. The main challenges in solving such problems are associated with the construction of the numerical meshes. Currently, there are two basic approaches to solve this problem. One is using of Lagrangian or Lagrangian Eulerian grid associated with the boundaries of media and the second is associated with the fixed Eulerian mesh, boundary cells of which cut boundaries of the environment medium and requires the calculation of these cut volumes. Both approaches require the complex grid generators and significant time for preparing the code’s data for simulation. In this codes these problems are solved using two grids, regular fixed and mobile local Euler Lagrange - Eulerian (ALE approach) accompanying the contact and free boundaries, the surfaces of shock waves and phase transitions, and other possible features of solutions, with mutual interpolation of integrated parameters. For modeling of both liquids and gases, and deformable solids the Godunov scheme of increased accuracy is used in Lagrangian - Eulerian variables, the same for the Euler equations and for the Euler- Cauchy, describing the deformation of the solid. The increased accuracy of the scheme is achieved by using 3D spatial time dependent solution of the discontinuity problem (3D space time dependent Riemann's Problem solver). The same solution is used to calculate the interaction at the liquid-solid surface (Fluid Structure Interaction problem). The codes does not require complex 3D mesh generators, only the surfaces of the calculating objects as the STL files created by means of engineering graphics are given by the user, which greatly simplifies the preparing the task and makes it convenient to use directly by the designer at the design stage. The results of the test solutions and applications related to the generation and extension of the detonation and shock waves, loading the constructions are presented.

Keywords: fluid structure interaction, Riemann's solver, Euler variables, 3D codes

Procedia PDF Downloads 430
496 Promoting Creative and Critical Thinking in Mathematics

Authors: Ana Maria Reis D'Azevedo Breda, Catarina Maria Neto da Cruz

Abstract:

The Japanese art of origami provides a rich context for designing exploratory mathematical activities for children and young people. By folding a simple sheet of paper, fascinating and surprising planar and spatial configurations emerge. Equally surprising is the unfolding process, which also produces striking patterns. The procedure of folding, unfolding, and folding again allows the exploration of interesting geometric patterns. When adequately and systematically done, we may deduce some of the mathematical rules ruling origami. As the child/youth folds the sheet of paper repeatedly, he can physically observe how the forms he obtains are transformed and how they relate to the pattern of the corresponding unfolding, creating space for the understanding/discovery of mathematical principles regulating the folding-unfolding process. As part of a 2023 Summer Academy organized by a Portuguese university, a session entitled “Folding, Thinking and Generalizing” took place. Twenty-three students attended the session, all enrolled in the 2nd cycle of Portuguese Basic Education and aged between 10 and 12 years old. The main focus of this session was to foster the development of critical cognitive and socio-emotional skills among these young learners using origami. These skills included creativity, critical analysis, mathematical reasoning, collaboration, and communication. Employing a qualitative, descriptive, and interpretative analysis of data collected during the session through field notes and students’ written productions, our findings reveal that structured origami-based activities not only promote student engagement with mathematical concepts in a playful and interactive but also facilitate the development of socio-emotional skills, which include collaboration and effective communication between participants. This research highlights the value of integrating origami into educational practices, highlighting its role in supporting comprehensive cognitive and emotional learning experiences.

Keywords: skills, origami rules, active learning, hands-on activities

Procedia PDF Downloads 61
495 Disparity in New Born Care Practices Reducing in Uttar Pradesh: Evidences from NFHS and DLHS

Authors: Gudakesh Yadav

Abstract:

Utter Pradesh, which is one of the largest states of India with unequal distribution of resources and different socioeconomic and cultural characteristics, level of different new born health care indicators varies a lot from one district to another district. State shared more than 21 percent of total live births of India; whereas, it accounts for 28 percent of total infant deaths of the country, with the 53 per thousand infant mortality rate. The present paper attempts to examine tempo-spatial changes in new born care practices during NFHS-1 to NFHS-3 and DLHS-2 to DLHS-3 in Uttar Pradesh and different regions. Descriptive statistics, rate-ratios, concentration index, multivariate and decomposition analysis has been used for the study. Findings of the study reveal that new born care practices have improved over the time in the state and across all the regions because of giving more emphasis on venerable groups like poor, rural, less educated mothers and scheduled caste & tribes but still it did not achieve the desired successes. Regional analysis of third rounds of DLHS shows that, coverage of intuitional delivery was the lowest in the central region. Performance of the southern region was the lowest in terms of initiation of breastfeeding, keeping baby warm and dry after the birth. The study calls for proper follow up of new born children to accelerate new born and child health care service and prioritises increasing antenatal check-ups and institutional delivery, which helps to improve level of other new born care services. At the policy level there is need to reach venerable groups like scheduled caste and tribes, poor and uneducated, and new mother especially in rural areas. High focused district should be allocated for better implementation of new born care promotion programme in low performing districts. Partnership with the private sector health professional is necessary to reach the every part of population.

Keywords: decomposition, inequality, initiation of breastfeeding, institutional delivery

Procedia PDF Downloads 234
494 Spatio-temporal Distribution of Surface Water Quality in the Kebir Rhumel Basin, Algeria

Authors: Lazhar Belkhiri, Ammar Tiri, Lotfi Mouni, Fatma Elhadj Lakouas

Abstract:

This research aims to present a surface water quality assessment of hydrochemical parameters in the Kebir Rhumel Basin, Algeria. The water quality index (WQI), Mann–Kendall (MK) test, and hierarchical cluster analysis (HCA) were used in oder to understand the spatio-temporal distribution of the surface water quality in the study area. Eleven hydrochemical parameters were measured monthly at eight stations from January 2016 to December 2020. The dominant cation in the surface water was found to be calcium, followed by sodium, and the dominant anion was sulfate, followed by chloride. In terms of WQI, a significant percentage of surface water samples at stations Ain Smara (AS), Beni Haroune (BH), Grarem (GR), and Sidi Khlifa (SK) exhibited poor water quality, with approximately 89.5%, 90.6%, 78.2%, and 62.7%, respectively, falling into this category. Mann–Kendall trend analysis revealed a significantly increasing trend in WQI values at stations Oued Boumerzoug (ON) and SK, indicating that the temporal variation of WQI in these stations is significant. Hierarchical clustering analysis classified the data into three clusters. The first cluster contained approximately 22% of the total number of months, the second cluster included about 30%, and the third cluster had the highest representation, approximately 48% of the total number of months. Within these clusters, certain stations exhibited higher WQI values. In the first cluster, stations GR and ON had the highest WQI values. In the second cluster, stations Oued Boumerzoug (OB) and SK showed the highest WQI values, while in the last cluster, stations AS, BH, El Milia (EM), and Hammam Grouz (HG) had the highest mean WQI values. Also, approximately 38%, 41%, and 38% of the total water samples in the first, second, and third clusters, respectively, were classified as having poor water quality. The findings of this study can serve as a scientific basis for decision-makers to formulate strategies for surface water quality restoration and management in the region.

Keywords: surface water, water quality index (WQI), Mann Kendall (MK) test, hierarchical cluster analysis (HCA), spatial-temporal distribution, Kebir Rhumel Basin

Procedia PDF Downloads 10
493 Fair Federated Learning in Wireless Communications

Authors: Shayan Mohajer Hamidi

Abstract:

Federated Learning (FL) has emerged as a promising paradigm for training machine learning models on distributed data without the need for centralized data aggregation. In the realm of wireless communications, FL has the potential to leverage the vast amounts of data generated by wireless devices to improve model performance and enable intelligent applications. However, the fairness aspect of FL in wireless communications remains largely unexplored. This abstract presents an idea for fair federated learning in wireless communications, addressing the challenges of imbalanced data distribution, privacy preservation, and resource allocation. Firstly, the proposed approach aims to tackle the issue of imbalanced data distribution in wireless networks. In typical FL scenarios, the distribution of data across wireless devices can be highly skewed, resulting in unfair model updates. To address this, we propose a weighted aggregation strategy that assigns higher importance to devices with fewer samples during the aggregation process. By incorporating fairness-aware weighting mechanisms, the proposed approach ensures that each participating device's contribution is proportional to its data distribution, thereby mitigating the impact of data imbalance on model performance. Secondly, privacy preservation is a critical concern in federated learning, especially in wireless communications where sensitive user data is involved. The proposed approach incorporates privacy-enhancing techniques, such as differential privacy, to protect user privacy during the model training process. By adding carefully calibrated noise to the gradient updates, the proposed approach ensures that the privacy of individual devices is preserved without compromising the overall model accuracy. Moreover, the approach considers the heterogeneity of devices in terms of computational capabilities and energy constraints, allowing devices to adaptively adjust the level of privacy preservation to strike a balance between privacy and utility. Thirdly, efficient resource allocation is crucial for federated learning in wireless communications, as devices operate under limited bandwidth, energy, and computational resources. The proposed approach leverages optimization techniques to allocate resources effectively among the participating devices, considering factors such as data quality, network conditions, and device capabilities. By intelligently distributing the computational load, communication bandwidth, and energy consumption, the proposed approach minimizes resource wastage and ensures a fair and efficient FL process in wireless networks. To evaluate the performance of the proposed fair federated learning approach, extensive simulations and experiments will be conducted. The experiments will involve a diverse set of wireless devices, ranging from smartphones to Internet of Things (IoT) devices, operating in various scenarios with different data distributions and network conditions. The evaluation metrics will include model accuracy, fairness measures, privacy preservation, and resource utilization. The expected outcomes of this research include improved model performance, fair allocation of resources, enhanced privacy preservation, and a better understanding of the challenges and solutions for fair federated learning in wireless communications. The proposed approach has the potential to revolutionize wireless communication systems by enabling intelligent applications while addressing fairness concerns and preserving user privacy.

Keywords: federated learning, wireless communications, fairness, imbalanced data, privacy preservation, resource allocation, differential privacy, optimization

Procedia PDF Downloads 65
492 Understanding Responses of the Bee Community to an Urbanizing Landscape in Bengaluru, South India

Authors: Chethana V. Casiker, Jagadishakumara B., Sunil G. M., Chaithra K., M. Soubadra Devy

Abstract:

A majority of the world’s food crops depends on insects for pollination, among which bees are the most dominant taxon. Bees pollinate vegetables, fruits and oilseeds which are rich in essential micronutrients. Besides being a prerequisite for a nutritionally secure diet, agrarian economies such as India depend heavily on pollination for good yield and quality of the product. As cities all over the world expand rapidly, large tracts of green spaces are being built up. This, along with high usage of agricultural chemicals has reduced floral diversity and shrunk bee habitats. Indeed, pollinator decline is being reported from various parts of the world. Further, the FAO has reported a huge increase in the area of land under cultivation of pollinator-dependent crops. In the light of increasing demand for pollination and disappearing natural habitats, it is critical to understand whether and how urban spaces can support pollinators. To this end, this study investigates the influence of landscape and local habitat quality on bee community dynamics. To capture the dynamics of expanding cityscapes, the study employs a space for time substitution, wherein a transect along the gradient of urbanization substitutes a timeframe of increasing urbanization. This will help understand how pollinators would respond to changes induced by increasing intensity of urbanization in the future. Bengaluru, one of the fastest growing cities of Southern India, is an excellent site to study impacts associated with urbanization. With sites moving away from the Bengaluru’s centre and towards its peripheries, this study captures the changes in bee species diversity and richness along a gradient of urbanization. Bees were sampled under different land use types as well as in different types of vegetation, including plantations, croplands, fallow land, parks, lake embankments, and private gardens. The relationship between bee community metrics and key drivers such as a percentage of built-up area, land use practices, and floral resources was examined. Additionally, data collected using questionnaire interviews were used to understand people’s perceptions towards and level of dependence on pollinators. Our results showed that urban areas are capable of supporting bees. In fact, a greater diversity of bees was recorded in urban sites compared to adjoining rural areas. This suggests that bees are able to seek out patchy resources and survive in small fragments of habitat. Bee abundance and species richness correlated positively with floral abundance and richness, indicating the role of vegetation in providing forage and nesting sites which are crucial to their survival. Bee numbers were seen to decrease with increase in built-up area demonstrating that impervious surfaces could act as deterrents. Findings from this study challenge the popular notion of cities being biodiversity-bare spaces. There is indeed scope for conserving bees in urban landscapes, provided that there are city-scale planning and local initiative. Bee conservation can go hand in hand with efforts such as urban gardening and terrace farming that could help cities urbanize sustainably.

Keywords: bee, landscape ecology, urbanization, urban pollination

Procedia PDF Downloads 164
491 Linking the Built Environment, Activities and Well-Being: Examining the Stories among Older Adults during Ageing-in-Place

Authors: Wenquan Gan, Peiyu Zhao, Xinyu Zhao

Abstract:

Under the background of the rapid development of China’s ageing population, ageing-in-place has become a primary strategy to cope with this problem promoted by the Chinese government. However, most older adults currently living in old residential communities are insufficient to support their ageing-in-place. Therefore, exploring how to retrofit existing communities towards ageing-friendly standards to support older adults is essential for healthy ageing. To better cope with this issue, this study aims to shed light on the inter-relationship among the built environment, daily activities, and well-being of older adults in urban China. Using mixed research methods including GPS tracking, structured observation, and in-depth interview to examine: (a) what specific places or facilities are most commonly used by the elderly in the ageing-in-place process; (b) what specific built environment characteristics attract older adults in these frequently used places; (c) how has the use of these spaces impacted the well-being of older adults. Specifically, structured observation and GPS are used to record and map the older residents’ behaviour and movement in Suzhou, China, a city with a highly aged population and suitable as a research case. Subsequently, a follow-up interview is conducted to explore what impact of activities and the built environment on their well-being. Results showed that for the elderly with good functional ability, the facilities promoted by the Chinese government to support ageing-in-place, such as community nursing homes for the aged, day-care centre, and activity centres for the aged, are rarely used by older adults. Additionally, older adults have their preferred activities and built environment characteristics that contribute to their well-being. Our findings indicate that a complex interrelationship between the built environment and activities can influence the well-being of the elderly. Further investigations are needed to understand how to support healthy ageing-in-place, especially in addition to providing permanent elder-ly-care facilities, but to attend to the design interventions that can enhance these particularly built environment characteristics to facilitate a healthy lifestyle in later life.

Keywords: older adults, built environment, spatial behavior, community activity, healthy ageing

Procedia PDF Downloads 89
490 Book Exchange System with a Hybrid Recommendation Engine

Authors: Nilki Upathissa, Torin Wirasinghe

Abstract:

This solution addresses the challenges faced by traditional bookstores and the limitations of digital media, striking a balance between the tactile experience of printed books and the convenience of modern technology. The book exchange system offers a sustainable alternative, empowering users to access a diverse range of books while promoting community engagement. The user-friendly interfaces incorporated into the book exchange system ensure a seamless and enjoyable experience for users. Intuitive features for book management, search, and messaging facilitate effortless exchanges and interactions between users. By streamlining the process, the system encourages readers to explore new books aligned with their interests, enhancing the overall reading experience. Central to the system's success is the hybrid recommendation engine, which leverages advanced technologies such as Long Short-Term Memory (LSTM) models. By analyzing user input, the engine accurately predicts genre preferences, enabling personalized book recommendations. The hybrid approach integrates multiple technologies, including user interfaces, machine learning models, and recommendation algorithms, to ensure the accuracy and diversity of the recommendations. The evaluation of the book exchange system with the hybrid recommendation engine demonstrated exceptional performance across key metrics. The high accuracy score of 0.97 highlights the system's ability to provide relevant recommendations, enhancing users' chances of discovering books that resonate with their interests. The commendable precision, recall, and F1score scores further validate the system's efficacy in offering appropriate book suggestions. Additionally, the curve classifications substantiate the system's effectiveness in distinguishing positive and negative recommendations. This metric provides confidence in the system's ability to navigate the vast landscape of book choices and deliver recommendations that align with users' preferences. Furthermore, the implementation of this book exchange system with a hybrid recommendation engine has the potential to revolutionize the way readers interact with printed books. By facilitating book exchanges and providing personalized recommendations, the system encourages a sense of community and exploration within the reading community. Moreover, the emphasis on sustainability aligns with the growing global consciousness towards eco-friendly practices. With its robust technical approach and promising evaluation results, this solution paves the way for a more inclusive, accessible, and enjoyable reading experience for book lovers worldwide. In conclusion, the developed book exchange system with a hybrid recommendation engine represents a progressive solution to the challenges faced by traditional bookstores and the limitations of digital media. By promoting sustainability, widening access to printed books, and fostering engagement with reading, this system addresses the evolving needs of book enthusiasts. The integration of user-friendly interfaces, advanced machine learning models, and recommendation algorithms ensure accurate and diverse book recommendations, enriching the reading experience for users.

Keywords: recommendation systems, hybrid recommendation systems, machine learning, data science, long short-term memory, recurrent neural network

Procedia PDF Downloads 83
489 Predictive Analytics for Theory Building

Authors: Ho-Won Jung, Donghun Lee, Hyung-Jin Kim

Abstract:

Predictive analytics (data analysis) uses a subset of measurements (the features, predictor, or independent variable) to predict another measurement (the outcome, target, or dependent variable) on a single person or unit. It applies empirical methods in statistics, operations research, and machine learning to predict the future, or otherwise unknown events or outcome on a single or person or unit, based on patterns in data. Most analyses of metabolic syndrome are not predictive analytics but statistical explanatory studies that build a proposed model (theory building) and then validate metabolic syndrome predictors hypothesized (theory testing). A proposed theoretical model forms with causal hypotheses that specify how and why certain empirical phenomena occur. Predictive analytics and explanatory modeling have their own territories in analysis. However, predictive analytics can perform vital roles in explanatory studies, i.e., scientific activities such as theory building, theory testing, and relevance assessment. In the context, this study is to demonstrate how to use our predictive analytics to support theory building (i.e., hypothesis generation). For the purpose, this study utilized a big data predictive analytics platform TM based on a co-occurrence graph. The co-occurrence graph is depicted with nodes (e.g., items in a basket) and arcs (direct connections between two nodes), where items in a basket are fully connected. A cluster is a collection of fully connected items, where the specific group of items has co-occurred in several rows in a data set. Clusters can be ranked using importance metrics, such as node size (number of items), frequency, surprise (observed frequency vs. expected), among others. The size of a graph can be represented by the numbers of nodes and arcs. Since the size of a co-occurrence graph does not depend directly on the number of observations (transactions), huge amounts of transactions can be represented and processed efficiently. For a demonstration, a total of 13,254 metabolic syndrome training data is plugged into the analytics platform to generate rules (potential hypotheses). Each observation includes 31 predictors, for example, associated with sociodemographic, habits, and activities. Some are intentionally included to get predictive analytics insights on variable selection such as cancer examination, house type, and vaccination. The platform automatically generates plausible hypotheses (rules) without statistical modeling. Then the rules are validated with an external testing dataset including 4,090 observations. Results as a kind of inductive reasoning show potential hypotheses extracted as a set of association rules. Most statistical models generate just one estimated equation. On the other hand, a set of rules (many estimated equations from a statistical perspective) in this study may imply heterogeneity in a population (i.e., different subpopulations with unique features are aggregated). Next step of theory development, i.e., theory testing, statistically tests whether a proposed theoretical model is a plausible explanation of a phenomenon interested in. If hypotheses generated are tested statistically with several thousand observations, most of the variables will become significant as the p-values approach zero. Thus, theory validation needs statistical methods utilizing a part of observations such as bootstrap resampling with an appropriate sample size.

Keywords: explanatory modeling, metabolic syndrome, predictive analytics, theory building

Procedia PDF Downloads 271
488 Vertical Urbanization Over Public Structures: The Example of Mostar Junction in Belgrade, Serbia

Authors: Sladjana Popovic

Abstract:

The concept of vertical space urbanization, defined in English as "air rights development," can be considered a mechanism for the development of public spaces in urban areas of high density. A chronological overview of the transformation of space within the vertical projection of the existing traffic infrastructure that penetrates through the central areas of a city is given in this paper through the analysis of two illustrative case studies: more advanced and recent - "Plot 13" in Boston, and less well-known European example of structures erected above highways throughout Italy - the "Pavesi auto grill" chain. The backbone of this analysis is the examination of the possibility of yielding air rights within the vertical projection of public structures in the two examples by considering the factors that would enable its potential application in capitals in Southeastern Europe. The cession of air rights in the Southeastern Europe region, as a phenomenon, has not been a recognized practice in urban planning. In a formal sense, legal and physical feasibility can be seen to some extent in local models of structures built above protected historical heritage (i.e., archaeological sites); however, the mechanisms of the legal process of assigning the right to use and develop air rights above public structures is not a recognized concept. The goal of the analysis is to shed light on the influence of institutional participants in the implementation of innovative solutions for vertical urbanization, as well as strategic planning mechanisms in public-private partnership models that would enable the implementation of the concept in the region. The main question is whether the manipulation of the vertical projection of space could provide for innovative urban solutions that overcome the deficit and excessive use of the available construction land, particularly above the dominant public spaces and traffic infrastructure that penetrate central parts of a city. Conclusions reflect upon vertical urbanization that can bridge the spatial separation of the city, reduce noise pollution and contribute to more efficient urban planning along main transportation corridors.

Keywords: air rights development, innovative urbanism, public-private partnership, transport infrastructure, vertical urbanization

Procedia PDF Downloads 67
487 Assessment of Environmental Quality of an Urban Setting

Authors: Namrata Khatri

Abstract:

The rapid growth of cities is transforming the urban environment and posing significant challenges for environmental quality. This study examines the urban environment of Belagavi in Karnataka, India, using geostatistical methods to assess the spatial pattern and land use distribution of the city and to evaluate the quality of the urban environment. The study is driven by the necessity to assess the environmental impact of urbanisation. Satellite data was utilised to derive information on land use and land cover. The investigation revealed that land use had changed significantly over time, with a drop in plant cover and an increase in built-up areas. High-resolution satellite data was also utilised to map the city's open areas and gardens. GIS-based research was used to assess public green space accessibility and to identify regions with inadequate waste management practises. The findings revealed that garbage collection and disposal techniques in specific areas of the city needed to be improved. Moreover, the study evaluated the city's thermal environment using Landsat 8 land surface temperature (LST) data. The investigation found that built-up regions had higher LST values than green areas, pointing to the city's urban heat island (UHI) impact. The study's conclusions have far-reaching ramifications for urban planners and politicians in Belgaum and other similar cities. The findings may be utilised to create sustainable urban planning strategies that address the environmental effect of urbanisation while also improving the quality of life for city dwellers. Satellite data and high-resolution satellite pictures were gathered for the study, and remote sensing and GIS tools were utilised to process and analyse the data. Ground truthing surveys were also carried out to confirm the accuracy of the remote sensing and GIS-based data. Overall, this study provides a complete assessment of Belgaum's environmental quality and emphasizes the potential of remote sensing and geographic information systems (GIS) approaches in environmental assessment and management.

Keywords: environmental quality, UEQ, remote sensing, GIS

Procedia PDF Downloads 73
486 Landslide Susceptibility Mapping Using Soft Computing in Amhara Saint

Authors: Semachew M. Kassa, Africa M Geremew, Tezera F. Azmatch, Nandyala Darga Kumar

Abstract:

Frequency ratio (FR) and analytical hierarchy process (AHP) methods are developed based on past landslide failure points to identify the landslide susceptibility mapping because landslides can seriously harm both the environment and society. However, it is still difficult to select the most efficient method and correctly identify the main driving factors for particular regions. In this study, we used fourteen landslide conditioning factors (LCFs) and five soft computing algorithms, including Random Forest (RF), Support Vector Machine (SVM), Logistic Regression (LR), Artificial Neural Network (ANN), and Naïve Bayes (NB), to predict the landslide susceptibility at 12.5 m spatial scale. The performance of the RF (F1-score: 0.88, AUC: 0.94), ANN (F1-score: 0.85, AUC: 0.92), and SVM (F1-score: 0.82, AUC: 0.86) methods was significantly better than the LR (F1-score: 0.75, AUC: 0.76) and NB (F1-score: 0.73, AUC: 0.75) method, according to the classification results based on inventory landslide points. The findings also showed that around 35% of the study region was made up of places with high and very high landslide risk (susceptibility greater than 0.5). The very high-risk locations were primarily found in the western and southeastern regions, and all five models showed good agreement and similar geographic distribution patterns in landslide susceptibility. The towns with the highest landslide risk include Amhara Saint Town's western part, the Northern part, and St. Gebreal Church villages, with mean susceptibility values greater than 0.5. However, rainfall, distance to road, and slope were typically among the top leading factors for most villages. The primary contributing factors to landslide vulnerability were slightly varied for the five models. Decision-makers and policy planners can use the information from our study to make informed decisions and establish policies. It also suggests that various places should take different safeguards to reduce or prevent serious damage from landslide events.

Keywords: artificial neural network, logistic regression, landslide susceptibility, naïve Bayes, random forest, support vector machine

Procedia PDF Downloads 68
485 Transient Freshwater-Saltwater Transition-Zone Dynamics in Heterogeneous Coastal Aquifers

Authors: Antoifi Abdoulhalik, Ashraf Ahmed

Abstract:

The ever growing threat of saltwater intrusion has prompted the need to further advance the understanding of underlying processes related to SWI for effective water resource management. While research efforts have mainly been focused on steady state analysis, studies on the transience of saltwater intrusion mechanism remain very scarce and studies considering transient SWI in heterogeneous medium are, as per our knowledge, simply inexistent. This study provides for the first time a quantitative analysis of the effect of both inland and coastal water level changes on the transition zone under transient conditions in layered coastal aquifer. In all, two sets of four experiments were completed, including a homogeneous case, and four layered cases: case LH and case HL presented were two bi-layered scenarios where a low K layer was set at the top and the bottom, respectively; case HLH and case LHL presented two stratified aquifers with High K–Low K–High K and Low K–High K– Low K pattern, respectively. Experimental automated image analysis technique was used here to quantify the main SWI parameters under high spatial and temporal resolution. The findings of this study provide an invaluable insight on the underlying processes responsible of transition zone dynamics in coastal aquifers. The results show that in all the investigated cases, the width of the transition zone remains almost unchanged throughout the saltwater intrusion process regardless of where the boundary change occurs. However, the results demonstrate that the width of the transition zone considerably increases during the retreat, with largest amplitude observed in cases LH and LHL, where a low K was set at the top of the system. In all the scenarios, the amplitude of widening was slightly smaller when the retreat was prompted by instantaneous drop of the saltwater level than when caused by inland freshwater rise, despite equivalent absolute head change magnitude. The magnitude of head change significantly caused larger widening during the saltwater wedge retreat, while having no impact during the intrusion phase.

Keywords: freshwater-saltwater transition-zone dynamics, heterogeneous coastal aquifers, laboratory experiments, transience seawater intrusion

Procedia PDF Downloads 232
484 Intermittent Effect of Coupled Thermal and Acoustic Sources on Combustion: A Spatial Perspective

Authors: Pallavi Gajjar, Vinayak Malhotra

Abstract:

Rockets have been known to have played a predominant role in spacecraft propulsion. The quintessential aspect of combustion-related requirements of a rocket engine is the minimization of the surrounding risks/hazards. Over time, it has become imperative to understand the combustion rate variation in presence of external energy source(s). Rocket propulsion represents a special domain of chemical propulsion assisted by high speed flows in presence of acoustics and thermal source(s). Jet noise leads to a significant loss of resources and every year a huge amount of financial aid is spent to prevent it. External heat source(s) induce high possibility of fire risk/hazards which can sufficiently endanger the operation of a space vehicle. Appreciable work had been done with justifiable simplification and emphasis on the linear variation of external energy source(s), which yields good physical insight but does not cater to accurate predictions. Present work experimentally attempts to understand the correlation between inter-energy conversions with the non-linear placement of external energy source(s). The work is motivated by the need to have better fire safety and enhanced combustion. The specific objectives of the work are a) To interpret the related energy transfer for combustion in presence of alternate external energy source(s) viz., thermal and acoustic, b) To fundamentally understand the role of key controlling parameters viz., separation distance, the number of the source(s), selected configurations and their non-linear variation to resemble real-life cases. An experimental setup was prepared using incense sticks as potential fuel and paraffin wax candles as the external energy source(s). The acoustics was generated using frequency generator, and source(s) were placed at selected locations. Non-equidistant parametric experimentation was carried out, and the effects were noted on regression rate changes. The results are expected to be very helpful in offering a new perspective into futuristic rocket designs and safety.

Keywords: combustion, acoustic energy, external energy sources, regression rate

Procedia PDF Downloads 137
483 Negotiating Strangeness: Narratives of Forced Return Migration and the Construction of Identities

Authors: Cheryl-Ann Sarita Boodram

Abstract:

Historically, the movement of people has been the subject of socio-political and economic regulatory policies which congeal to regulate human mobility and establish geopolitical and spatial identities and borderlands. As migratory practices evolved, so too has the problematization associated with movement, migration and citizenship. The emerging trends have led to active development of immigration technology governing human mobility and the naming of migratory practices. One such named phenomenon is ‘deportation’ or the forced removal of individuals from their adopted country. Deportation, has gained much attention within the human mobility landscape in the past twenty years following the September 2001 terrorist attack on the World Trade Centre in New York. In a reactionary move, several metropolitan countries, including Canada and the United Kingdom enacted or reviewed immigration laws which further enabled the removal of foreign born criminals to the land of their birth in the global south. Existing studies fall short of understanding the multiple textures of the forced returned migration experiences and the social injustices resulting from deportation displacement. This study brings together indigenous research methodologies through the use of participatory action research and social work with returned migrants in Trinidad and Tobago to uncover the experiences of displacement of deported nationals. The study explores the experiences of negotiating life as a ‘stranger’ and how return has influenced the construction of identities of returned nationals. Findings from this study reveal that deportation has led to inequalities and facilitated ‘othering’ of this group within their own country of birth. The study further highlighted that deportation leads to circuits of dispossession, and perpetuates inequalities. This study provides original insights into the way returned migrants negotiate, map and embody ‘strangeness’ and manage their return to a soil they consider unfamiliar and alien.

Keywords: stranger, alien geographies, displacement, deportation, negotiating strangeness, identity, otherness, alien landscapes

Procedia PDF Downloads 501
482 Medicompills Architecture: A Mathematical Precise Tool to Reduce the Risk of Diagnosis Errors on Precise Medicine

Authors: Adriana Haulica

Abstract:

Powered by Machine Learning, Precise medicine is tailored by now to use genetic and molecular profiling, with the aim of optimizing the therapeutic benefits for cohorts of patients. As the majority of Machine Language algorithms come from heuristics, the outputs have contextual validity. This is not very restrictive in the sense that medicine itself is not an exact science. Meanwhile, the progress made in Molecular Biology, Bioinformatics, Computational Biology, and Precise Medicine, correlated with the huge amount of human biology data and the increase in computational power, opens new healthcare challenges. A more accurate diagnosis is needed along with real-time treatments by processing as much as possible from the available information. The purpose of this paper is to present a deeper vision for the future of Artificial Intelligence in Precise medicine. In fact, actual Machine Learning algorithms use standard mathematical knowledge, mostly Euclidian metrics and standard computation rules. The loss of information arising from the classical methods prevents obtaining 100% evidence on the diagnosis process. To overcome these problems, we introduce MEDICOMPILLS, a new architectural concept tool of information processing in Precise medicine that delivers diagnosis and therapy advice. This tool processes poly-field digital resources: global knowledge related to biomedicine in a direct or indirect manner but also technical databases, Natural Language Processing algorithms, and strong class optimization functions. As the name suggests, the heart of this tool is a compiler. The approach is completely new, tailored for omics and clinical data. Firstly, the intrinsic biological intuition is different from the well-known “a needle in a haystack” approach usually used when Machine Learning algorithms have to process differential genomic or molecular data to find biomarkers. Also, even if the input is seized from various types of data, the working engine inside the MEDICOMPILLS does not search for patterns as an integrative tool. This approach deciphers the biological meaning of input data up to the metabolic and physiologic mechanisms, based on a compiler with grammars issued from bio-algebra-inspired mathematics. It translates input data into bio-semantic units with the help of contextual information iteratively until Bio-Logical operations can be performed on the base of the “common denominator “rule. The rigorousness of MEDICOMPILLS comes from the structure of the contextual information on functions, built to be analogous to mathematical “proofs”. The major impact of this architecture is expressed by the high accuracy of the diagnosis. Detected as a multiple conditions diagnostic, constituted by some main diseases along with unhealthy biological states, this format is highly suitable for therapy proposal and disease prevention. The use of MEDICOMPILLS architecture is highly beneficial for the healthcare industry. The expectation is to generate a strategic trend in Precise medicine, making medicine more like an exact science and reducing the considerable risk of errors in diagnostics and therapies. The tool can be used by pharmaceutical laboratories for the discovery of new cures. It will also contribute to better design of clinical trials and speed them up.

Keywords: bio-semantic units, multiple conditions diagnosis, NLP, omics

Procedia PDF Downloads 63