Search results for: distributed network
552 Analytical Model of Locomotion of a Thin-Film Piezoelectric 2D Soft Robot Including Gravity Effects
Authors: Zhiwu Zheng, Prakhar Kumar, Sigurd Wagner, Naveen Verma, James C. Sturm
Abstract:
Soft robots have drawn great interest recently due to a rich range of possible shapes and motions they can take on to address new applications, compared to traditional rigid robots. Large-area electronics (LAE) provides a unique platform for creating soft robots by leveraging thin-film technology to enable the integration of a large number of actuators, sensors, and control circuits on flexible sheets. However, the rich shapes and motions possible, especially when interacting with complex environments, pose significant challenges to forming well-generalized and robust models necessary for robot design and control. In this work, we describe an analytical model for predicting the shape and locomotion of a flexible (steel-foil-based) piezoelectric-actuated 2D robot based on Euler-Bernoulli beam theory. It is nominally (unpowered) lying flat on the ground, and when powered, its shape is controlled by an array of piezoelectric thin-film actuators. Key features of the models are its ability to incorporate the significant effects of gravity on the shape and to precisely predict the spatial distribution of friction against the contacting surfaces, necessary for determining inchworm-type motion. We verified the model by developing a distributed discrete element representation of a continuous piezoelectric actuator and by comparing its analytical predictions to discrete-element robot simulations using PyBullet. Without gravity, predicting the shape of a sheet with a linear array of piezoelectric actuators at arbitrary voltages is straightforward. However, gravity significantly distorts the shape of the sheet, causing some segments to flatten against the ground. Our work includes the following contributions: (i) A self-consistent approach was developed to exactly determine which parts of the soft robot are lifted off the ground, and the exact shape of these sections, for an arbitrary array of piezoelectric voltages and configurations. (ii) Inchworm-type motion relies on controlling the relative friction with the ground surface in different sections of the robot. By adding torque-balance to our model and analyzing shear forces, the model can then determine the exact spatial distribution of the vertical force that the ground is exerting on the soft robot. Through this, the spatial distribution of friction forces between ground and robot can be determined. (iii) By combining this spatial friction distribution with the shape of the soft robot, in the function of time as piezoelectric actuator voltages are changed, the inchworm-type locomotion of the robot can be determined. As a practical example, we calculated the performance of a 5-actuator system on a 50-µm thick steel foil. Piezoelectric properties of commercially available thin-film piezoelectric actuators were assumed. The model predicted inchworm motion of up to 200 µm per step. For independent verification, we also modelled the system using PyBullet, a discrete-element robot simulator. To model a continuous thin-film piezoelectric actuator, we broke each actuator into multiple segments, each of which consisted of two rigid arms with appropriate mass connected with a 'motor' whose torque was set by the applied actuator voltage. Excellent agreement between our analytical model and the discrete-element simulator was shown for both for the full deformation shape and motion of the robot.Keywords: analytical modeling, piezoelectric actuators, soft robot locomotion, thin-film technology
Procedia PDF Downloads 180551 Recent Findings of Late Bronze Age Mining and Archaeometallurgy Activities in the Mountain Region of Colchis (Southern Lechkhumi, Georgia)
Authors: Rusudan Chagelishvili, Nino Sulava, Tamar Beridze, Nana Rezesidze, Nikoloz Tatuashvili
Abstract:
The South Caucasus is one of the most important centers of prehistoric metallurgy, known for its Colchian bronze culture. Modern Lechkhumi – historical Mountainous Colchis where the existence of prehistoric metallurgy is confirmed by the discovery of many artifacts is a part of this area. Studies focused on prehistoric smelting sites, related artefacts, and ore deposits have been conducted during last ten years in Lechkhumi. More than 20 prehistoric smelting sites and artefacts associated with metallurgical activities (ore roasting furnaces, slags, crucible, and tuyères fragments) have been identified so far. Within the framework of integrated studies was established that these sites were operating in 13-9 centuries B.C. and used for copper smelting. Palynological studies of slags revealed that chestnut (Castanea sativa) and hornbeam (Carpinus sp.) wood were used as smelting fuel. Geological exploration-analytical studies revealed that copper ore mining, processing, and smelting sites were distributed close to each other. Despite recent complex data, the signs of prehistoric mines (trenches) haven’t been found in this part of the study area so far. Since 2018 the archaeological-geological exploration has been focused on the southern part of Lechkhumi and covered the areas of villages Okureshi and Opitara. Several copper smelting sites (Okureshi 1 and 2, Opitara 1), as well as a Colchian Bronze culture settlement, have been identified here. Three mine workings have been found in the narrow gorge of the river Rtkhmelebisgele in the vicinities of the village Opitara. In order to establish a link between the Opitara-Okureshi archaeometallurgical sites, Late Bronze Age settlements, and mines, various scientific analytical methods -mineralized rock and slags petrography and atomic absorption spectrophotometry (AAS) analysis have been applied. The careful examination of Opitara mine workings revealed that there is a striking difference between the mine #1 on the right bank of the river and mines #2 and #3 on the left bank. The first one has all characteristic features of the Soviet period mine working (e. g. high portal with angular ribs and roof showing signs of blasting). In contrast, mines #2 and #3, which are located very close to each other, have round-shaped portals/entrances, low roofs, and fairly smooth ribs and are filled with thick layers of river sediments and collapsed weathered rock mass. A thorough review of the publications related to prehistoric mine workings revealed some striking similarities between mines #2 and #3 with their worldwide analogues. Apparently, the ore extraction from these mines was conducted by fire-setting applying primitive tools. It was also established that mines are cut in Jurassic mineralized volcanic rocks. Ore minerals (chalcopyrite, pyrite, galena) are related to calcite and quartz veins. The results obtained through the petrochemical and petrography studies of mineralized rock samples from Opitara mines and prehistoric slags are in complete correlation with each other, establishing the direct link between copper mining and smelting within the study area. Acknowledgment: This work was supported by the Shota Rustaveli National Science Foundation of Georgia (grant # FR-19-13022).Keywords: archaeometallurgy, Mountainous Colchis, mining, ore minerals
Procedia PDF Downloads 180550 Highlighting Strategies Implemented by Migrant Parents to Support Their Child's Educational and Academic Success in the Host Society
Authors: Josee Charette
Abstract:
The academic and educational success of migrant students is a current issue in education, especially in western societies such in the province of Quebec, in Canada. For people who immigrate with school-age children, the success of the family’s migratory project is often measured by the benefits drawn by children from the educational institutions of their host society. In order to support the academic achievement of their children, migrant parents try to develop practices that derive from their representations of school and related challenges inspired by the socio-cultural context of their country of origin. These findings lead us to the following question: How does strategies implemented by migrant parents to manage the representational distance between school of their country of origin and school of their host society support or not the academic and educational success of their child? In the context of a qualitative exploratory approach, we have made interviews in the French , English and Spanish languages with 32 newly immigrated parents and 10 of their children. Parents were invited to complete a network of free associations about «School in Quebec» as a premise for the interview. The objective of this paper is to present strategies implemented by migrant parents to manage the distance between their representations of schools in their country of origin and in the host society, and to explore the influence of this management on their child’s academic and educational trajectories. Data analysis led us to develop various types of strategies, such as continuity, adaptation, resources mobilization, compensation and "return to basics" strategies. These strategies seem to be part of a continuum from oppositional-conflict scenario, in which parental strategies act as a risk factor, to conciliator-integrator scenario, in which parental strategies act as a protective factor for migrant students’ academic and educational success. In conclusion, we believe that our research helps in highlighting strategies implemented by migrant parents to support their child’s academic and educational success in the host society and also helps in providing a more efficient support to migrant parents and contributes to develop a wider portrait of migrant students’ academic achievement.Keywords: academic and educational achievement of immigrant students, family’s migratory project, immigrants parental strategies, representational distance between school of origin and school of host society
Procedia PDF Downloads 445549 Performance Improvement of Long-Reach Optical Access Systems Using Hybrid Optical Amplifiers
Authors: Shreyas Srinivas Rangan, Jurgis Porins
Abstract:
The internet traffic has increased exponentially due to the high demand for data rates by the users, and the constantly increasing metro networks and access networks are focused on improving the maximum transmit distance of the long-reach optical networks. One of the common methods to improve the maximum transmit distance of the long-reach optical networks at the component level is to use broadband optical amplifiers. The Erbium Doped Fiber Amplifier (EDFA) provides high amplification with low noise figure but due to the characteristics of EDFA, its operation is limited to C-band and L-band. In contrast, the Raman amplifier exhibits a wide amplification spectrum, and negative noise figure values can be achieved. To obtain such results, high powered pumping sources are required. Operating Raman amplifiers with such high-powered optical sources may cause fire hazards and it may damage the optical system. In this paper, we implement a hybrid optical amplifier configuration. EDFA and Raman amplifiers are used in this hybrid setup to combine the advantages of both EDFA and Raman amplifiers to improve the reach of the system. Using this setup, we analyze the maximum transmit distance of the network by obtaining a correlation diagram between the length of the single-mode fiber (SMF) and the Bit Error Rate (BER). This hybrid amplifier configuration is implemented in a Wavelength Division Multiplexing (WDM) system with a BER of 10⁻⁹ by using NRZ modulation format, and the gain uniformity noise ratio (signal-to-noise ratio (SNR)), the efficiency of the pumping source, and the optical signal gain efficiency of the amplifier are studied experimentally in a mathematical modelling environment. Numerical simulations were implemented in RSoft OptSim simulation software based on the nonlinear Schrödinger equation using the Split-Step method, the Fourier transform, and the Monte Carlo method for estimating BER.Keywords: Raman amplifier, erbium doped fibre amplifier, bit error rate, hybrid optical amplifiers
Procedia PDF Downloads 70548 The Role of Twitter Bots in Political Discussion on 2019 European Elections
Authors: Thomai Voulgari, Vasilis Vasilopoulos, Antonis Skamnakis
Abstract:
The aim of this study is to investigate the effect of the European election campaigns (May 23-26, 2019) on Twitter achieving with artificial intelligence tools such as troll factories and automated inauthentic accounts. Our research focuses on the last European Parliamentary elections that took place between 23 and 26 May 2019 specifically in Italy, Greece, Germany and France. It is difficult to estimate how many Twitter users are actually bots (Echeverría, 2017). Detection for fake accounts is becoming even more complicated as AI bots are made more advanced. A political bot can be programmed to post comments on a Twitter account for a political candidate, target journalists with manipulated content or engage with politicians and artificially increase their impact and popularity. We analyze variables related to 1) the scope of activity of automated bots accounts and 2) degree of coherence and 3) degree of interaction taking into account different factors, such as the type of content of Twitter messages and their intentions, as well as the spreading to the general public. For this purpose, we collected large volumes of Twitter accounts of party leaders and MEP candidates between 10th of May and 26th of July based on content analysis of tweets based on hashtags while using an innovative network analysis tool known as MediaWatch.io (https://mediawatch.io/). According to our findings, one of the highest percentage (64.6%) of automated “bot” accounts during 2019 European election campaigns was in Greece. In general terms, political bots aim to proliferation of misinformation on social media. Targeting voters is a way that it can be achieved contribute to social media manipulation. We found that political parties and individual politicians create and promote purposeful content on Twitter using algorithmic tools. Based on this analysis, online political advertising play an important role to the process of spreading misinformation during elections campaigns. Overall, inauthentic accounts and social media algorithms are being used to manipulate political behavior and public opinion.Keywords: artificial intelligence tools, human-bot interactions, political manipulation, social networking, troll factories
Procedia PDF Downloads 138547 Experiential Language Learning as a Tool for Effective Global Leadership
Authors: Christiane Dumont
Abstract:
This paper proposes to revisit foreign-language learning as a tool to increase motivation through advocacy and develop effective natural communication skills, which are critical leadership qualities. To this end, collaborative initiatives undertaken by advanced university students of French with local and international community partners will be reviewed. Close attention will be paid to the acquisition of intercultural skills, the reflective process, as well as the challenges and outcomes. Two international development projects conducted in Haiti will be highlighted, i.e., collaboration with a network of providers in the Haitian cultural heritage preservation and tourism sector (2014-15) and development of investigation and teacher training tools for a primary/secondary school in the Port-au-Prince area (current). The choice of community-service learning as a framework to teach French-as-a-second-language stemmed from the need to raise awareness against stereotypes and prejudice, which hinder the development of effective intercultural skills. This type of experiential education also proved very effective in identifying and preventing miscommunication caused by the lack of face-to-face interaction in our increasingly technology-mediated world. Learners experienced first-hand, the challenges and advantages of face-to-face communication, which, in turn, enhanced their motivation for developing effective intercultural skills. Vygotsky's and Kolb's theories, current research on service learning (Dwight, Eyler), action/project-based pedagogy (Beckett), and reflective learning (TSC Farrell), will provide useful background to analyze the benefits and challenges of community-service learning. The ultimate goal of this paper is to find out what makes experiential learning truly unique and transformative for both the learners and the community they wish to serve. It will demonstrate how enhanced motivation, community engagement, and clear, concise, and respectful communication impact and empower learners. The underlying hope is to help students in high-profile, and leading-edge industries become effective global leaders.Keywords: experiential learning, intercultural communication, reflective learning, effective leadership, learner motivation
Procedia PDF Downloads 105546 Dynamic Ambulance Deployment to Reduce Ambulance Response Times Using Geographic Information Systems
Authors: Masoud Swalehe, Semra Günay
Abstract:
Developed countries are losing many lives to non-communicable diseases as compared to their developing counterparts. The effects of these diseases are mostly sudden and manifest at a very short time prior to death or a dangerous attack and this has consolidated the significance of emergency medical system (EMS) as one of the vital areas of healthcare service delivery. The primary objective of this research is to reduce ambulance response times (RT) of Eskişehir province EMS since a number of studies have established a relationship between ambulance response times and survival chances of patients especially out of hospital cardiac arrest (OHCA) victims. It has been found out that patients who receive out of hospital medical attention in few (4) minutes after cardiac arrest because of low ambulance response times stand higher chances of survival than their counterparts who take longer times (more than 12 minutes) to receive out of hospital medical care because of higher ambulance response times. The study will make use of geographic information systems (GIS) technology to dynamically reallocate ambulance resources according to demand and time so as to reduce ambulance response times. Geospatial-time distribution of ambulance calls (demand) will be used as a basis for optimal ambulance deployment using system status management (SSM) strategy to achieve much demand coverage with the same number of ambulance resources to cause response time reduction. Drive-time polygons will be used to come up with time specific facility coverage areas and suggesting additional facility candidate sites where ambulance resources can be moved to serve higher demands making use of network analysis techniques. Emergency Ambulance calls’ data from 1st January 2014 to 31st December 2014 obtained from Eskişehir province health directorate will be used in this study. This study will focus on the reduction of ambulance response times which is a key Emergency Medical Services performance indicator.Keywords: emergency medical services, system status management, ambulance response times, geographic information system, geospatial-time distribution, out of hospital cardiac arrest
Procedia PDF Downloads 300545 Reducing the Stigma of Homelessness through Community Engagement and Reciprocity
Authors: Jessica Federman
Abstract:
The current research offers a longitudinal and qualitative study design to examine how reciprocity improves relations between the homeless and various stakeholders within a community. The study examines a homeless shelter that sought to establish a facility within a community of Los Angeles, that was initially met with strong resistance and opposition from a variety of organizations due to deeply entrenched views about the negative impact of having homeless individuals within the community. The project tested an intervention model that targets the reduction of stigmatization of homeless individuals and promotes synergistic exchanges between conflicted organizational entities in communities. Years later, the data show that there has been a remarkable reversal in the perception of the agency by the very forces that initially prevented it from being established. This reversal was achieved by a few key strategic decisions. Community engagement was the first step toward changing people’s minds and demonstrating how the homeless shelter was helping to alleviate the problem of homelessness instead of contributing to it. Central to the non-profit’s success was the agency’s pioneering formulation of a treatment model known as, Reciprocal Community Engagement Model (RCEM). The model works by reintegrating the homeless back into society through relationship building within a network of programs that foster positive human connections. This approach aims to draw the homeless out of the debilitating isolation of their situation, reintegrate them through purposeful roles in the community while simultaneously providing a reciprocal benefit to the community at large. Through multilevel, simultaneous social interaction, RCEM has a direct impact not only on the homeless shelter’s clients but also for the community as well. The agency’s approach of RCEM led to their homeless clients getting out of the shelter and getting to work in the community directly alongside other community volunteers and for the benefit of other city and community organizations. This led to several opportunities for community members and residents to interact in meaningful ways. Through each successive exposure, the resident and community members’ distrust in one another was gradually eased and a mutually supportive relationship restored. In this process, the community member becomes the locus of change as much as the residents of the shelter. Measurements of community trust and resilience increased while negative perceptions of homeless people decreased.Keywords: stigma, homelessness, reciprocity, identity
Procedia PDF Downloads 182544 Energy Dissipation Characteristics of an Elastomer under Dynamic Condition: A Comprehensive Assessment Using High and Low Frequency Analyser
Authors: K. Anas, M. Selvakumar, Samson David, R. R. Babu, S. Chattopadhyay
Abstract:
The dynamic deformation of a visco elastic material can cause heat generation. This heat generation is aspect energy dissipation. The present work investigates the contribution of various factors like; elastomer structure, cross link type and density, filler networking, reinforcement potential and temperature at energy dissipation mechanism. The influences of these elements are investigated using very high frequency analyzer (VHF ) and dynamical mechanical analysis(DMA).VHF follows transmissibility and vibration isolation principle whereas DMA works on dynamical mechanical deformation principle. VHF analysis of different types of elastomers reveals that elastomer can act as a transmitter or damper of energy depending on the applied frequency ratio (ω/ωn). Dynamic modulus (G') of low damping rubbers like natural rubber does not varies rapidly with frequency but vice-versa for high damping rubber like butyl rubber (IIR). VHF analysis also depicts that polysulfidic linkages has high damping ratio (ζ) than mono sulfidic linkages due to its dissipative nature. At comparable cross link density, mono sulfidic linkages shows higher glass transition temperature (Tg) than poly sulfidic linkages. The intensity and location of loss modulus (G'') peak of different types of carbon black filled natural rubber compounds suggests that segmental relaxation at glass transition temperature (Tg) is seldom affected by filler particles, but the filler networks can influence the cross link density by absorbing the curatives. The filler network breaking and reformation during a dynamic strain is a thermally activated process. Thus, stronger aggregates are highly dissipative in nature. Measurements indicate that at lower temperature regimes polymeric chain friction is highly dissipative in nature.Keywords: damping ratio, natural frequency, crosslinking density, segmental motion, surface activity, dissipative, polymeric chain friction
Procedia PDF Downloads 295543 Incidence of Lymphoma and Gonorrhea Infection: A Retrospective Study
Authors: Diya Kohli, Amalia Ardeljan, Lexi Frankel, Jose Garcia, Lokesh Manjani, Omar Rashid
Abstract:
Gonorrhea is the second most common sexually transmitted disease (STDs) in the United States of America. Gonorrhea affects the urethra, rectum, or throat and the cervix in females. Lymphoma is a cancer of the immune network called the lymphatic system that includes the lymph nodes/glands, spleen, thymus gland, and bone marrow. Lymphoma can affect many organs in the body. When a lymphocyte develops a genetic mutation, it signals other cells into rapid proliferation that causes many mutated lymphocytes. Multiple studies have explored the incidence of cancer in people infected with STDs such as Gonorrhea. For instance, the studies conducted by Wang Y-C and Co., as well as Caini, S and Co. established a direct co-relationship between Gonorrhea infection and incidence of prostate cancer. We hypothesize that Gonorrhea infection also increases the incidence of Lymphoma in patients. This research study aimed to evaluate the correlation between Gonorrhea infection and the incidence of Lymphoma. The data for the research was provided by a Health Insurance Portability and Accountability Act (HIPAA) compliant national database. This database was utilized to evaluate patients infected with Gonorrhea versus the ones who were not infected to establish a correlation with the prevalence of Lymphoma using ICD-10 and ICD-9 codes. Access to the database was granted by the Holy Cross Health, Fort Lauderdale for academic research. Standard statistical methods were applied throughout. Between January 2010 and December 2019, the query was analyzed and resulted in 254 and 808 patients in both the infected and control group, respectively. The two groups were matched by Age Range and CCI score. The incidence of Lymphoma was 0.998% (254 patients out of 25455) in the Gonorrhea group (patients infected with Gonorrhea that was Lymphoma Positive) compared to 3.174% and 808 patients in the control group (Patients negative for Gonorrhea but with Lymphoma). This was statistically significant by a p-value < 2.210-16 with an OR= 0.431 (95% CI 0.381-0.487). The patients were then matched by antibiotic treatment to avoid treatment bias. The incidence of Lymphoma was 1.215% (82 patients out of 6,748) in the Gonorrhea group compared to 2.949% (199 patients out of 6748) in the control group. This was statistically significant by a p-value <5.410-10 with an OR= 0.468 (95% CI 0.367-0.596). The study shows a statistically significant correlation between Gonorrhea and a reduced incidence of Lymphoma. Further evaluation is recommended to assess the potential of Gonorrhea in reducing Lymphoma.Keywords: gonorrhea, lymphoma, STDs, cancer, ICD
Procedia PDF Downloads 195542 On the Survival of Individuals with Type 2 Diabetes Mellitus in the United Kingdom: A Retrospective Case-Control Study
Authors: Njabulo Ncube, Elena Kulinskaya, Nicholas Steel, Dmitry Pshezhetskiy
Abstract:
Life expectancy in the United Kingdom (UK) has been near constant since 2010, particularly for the individuals of 65 years and older. This trend has been also noted in several other countries. This slowdown in the increase of life expectancy was concurrent with the increase in the number of deaths caused by non-communicable diseases. Of particular concern is the world-wide exponential increase in the number of diabetes related deaths. Previous studies have reported increased mortality hazards among diabetics compared to non-diabetics, and on the differing effects of antidiabetic drugs on mortality hazards. This study aimed to estimate the all-cause mortality hazards and related life expectancies among type 2 diabetes (T2DM) patients in the UK using the time-variant Gompertz-Cox model with frailty. The study also aimed to understand the major causes of the change in life expectancy growth in the last decade. A total of 221 182 (30.8% T2DM, 57.6% Males) individuals aged 50 years and above, born between 1930 and 1960, inclusive, and diagnosed between 2000 and 2016, were selected from The Health Improvement Network (THIN) database of the UK primary care data and followed up to 31 December 2016. About 13.4% of participants died during the follow-up period. The overall all-cause mortality hazard ratio of T2DM compared to non-diabetic controls was 1.467 (1.381-1.558) and 1.38 (1.307-1.457) when diagnosed between 50 to 59 years and 60 to 74 years, respectively. The estimated life expectancies among T2DM individuals without further comorbidities diagnosed at the age of 60 years were 2.43 (1930-1939 birth cohort), 2.53 (1940-1949 birth cohort) and 3.28 (1950-1960 birth cohort) years less than those of non-diabetic controls. However, the 1950-1960 birth cohort had a steeper hazard function compared to the 1940-1949 birth cohort for both T2DM and non-diabetic individuals. In conclusion, mortality hazards for people with T2DM continue to be higher than for non-diabetics. The steeper mortality hazard slope for the 1950-1960 birth cohort might indicate the sub-population contributing to a slowdown in the growth of the life expectancy.Keywords: T2DM, Gompetz-Cox model with frailty, all-cause mortality, life expectancy
Procedia PDF Downloads 119541 Suitable Site Selection of Small Dams Using Geo-Spatial Technique: A Case Study of Dadu Tehsil, Sindh
Authors: Zahid Khalil, Saad Ul Haque, Asif Khan
Abstract:
Decision making about identifying suitable sites for any project by considering different parameters is difficult. Using GIS and Multi-Criteria Analysis (MCA) can make it easy for those projects. This technology has proved to be an efficient and adequate in acquiring the desired information. In this study, GIS and MCA were employed to identify the suitable sites for small dams in Dadu Tehsil, Sindh. The GIS software is used to create all the spatial parameters for the analysis. The parameters that derived are slope, drainage density, rainfall, land use / land cover, soil groups, Curve Number (CN) and runoff index with a spatial resolution of 30m. The data used for deriving above layers include 30-meter resolution SRTM DEM, Landsat 8 imagery, and rainfall from National Centre of Environment Prediction (NCEP) and soil data from World Harmonized Soil Data (WHSD). Land use/Land cover map is derived from Landsat 8 using supervised classification. Slope, drainage network and watershed are delineated by terrain processing of DEM. The Soil Conservation Services (SCS) method is implemented to estimate the surface runoff from the rainfall. Prior to this, SCS-CN grid is developed by integrating the soil and land use/land cover raster. These layers with some technical and ecological constraints are assigned weights on the basis of suitability criteria. The pairwise comparison method, also known as Analytical Hierarchy Process (AHP) is taken into account as MCA for assigning weights on each decision element. All the parameters and group of parameters are integrated using weighted overlay in GIS environment to produce suitable sites for the Dams. The resultant layer is then classified into four classes namely, best suitable, suitable, moderate and less suitable. This study reveals a contribution to decision-making about suitable sites analysis for small dams using geospatial data with minimal amount of ground data. This suitability maps can be helpful for water resource management organizations in determination of feasible rainwater harvesting structures (RWH).Keywords: Remote sensing, GIS, AHP, RWH
Procedia PDF Downloads 389540 Molecular Pathogenesis of NASH through the Dysregulation of Metabolic Organ Network in the NASH-HCC Model Mouse Treated with Streptozotocin-High Fat Diet
Authors: Bui Phuong Linh, Yuki Sakakibara, Ryuto Tanaka, Elizabeth H. Pigney, Taishi Hashiguchi
Abstract:
NASH is an increasingly prevalent chronic liver disease that can progress to hepatocellular carcinoma and now is attracting interest worldwide. The STAM™ model is a clinically-correlated murine NASH model which shows the same pathological progression as NASH patients and has been widely used for pharmacological and basic research. The multiple parallel hits hypothesis suggests abnormalities in adipocytokines, intestinal microflora, and endotoxins are intertwined and could contribute to the development of NASH. In fact, NASH patients often exhibit gut dysbiosis and dysfunction in adipose tissue and metabolism. However, the analysis of the STAM™ model has only focused on the liver. To clarify whether the STAM™ model can also mimic multiple pathways of NASH progression, we analyzed the organ crosstalk interactions between the liver and the gut and the phenotype of adipose tissue in the STAM™ model. NASH was induced in male mice by a single subcutaneous injection of 200 µg streptozotocin 2 days after birth and feeding with high-fat diet after 4 weeks of age. The mice were sacrificed at NASH stage. Colon samples were snap-frozen in liquid nitrogen and stored at -80˚C for tight junction-related protein analysis. Adipose tissue was prepared into paraffin blocks for HE staining. Blood adiponectin was analyzed to confirm changes in the adipocytokine profile. Tight junction-related proteins in the intestine showed that expression of ZO-1 decreased with the progression of the disease. Increased expression of endotoxin in the blood and decreased expression of Adiponectin were also observed. HE staining revealed hypertrophy of adipocytes. Decreased expression of ZO-1 in the intestine of STAM™ mice suggests the occurrence of leaky gut, and abnormalities in adipocytokine secretion were also observed. Together with the liver, phenotypes in these organs are highly similar to human NASH patients and might be involved in the pathogenesis of NASH.Keywords: Non-alcoholic steatohepatitis, hepatocellular carcinoma, fibrosis, organ crosstalk, leaky gut
Procedia PDF Downloads 159539 A Copula-Based Approach for the Assessment of Severity of Illness and Probability of Mortality: An Exploratory Study Applied to Intensive Care Patients
Authors: Ainura Tursunalieva, Irene Hudson
Abstract:
Continuous improvement of both the quality and safety of health care is an important goal in Australia and internationally. The intensive care unit (ICU) receives patients with a wide variety of and severity of illnesses. Accurately identifying patients at risk of developing complications or dying is crucial to increasing healthcare efficiency. Thus, it is essential for clinicians and researchers to have a robust framework capable of evaluating the risk profile of a patient. ICU scoring systems provide such a framework. The Acute Physiology and Chronic Health Evaluation III and the Simplified Acute Physiology Score II are ICU scoring systems frequently used for assessing the severity of acute illness. These scoring systems collect multiple risk factors for each patient including physiological measurements then render the assessment outcomes of individual risk factors into a single numerical value. A higher score is related to a more severe patient condition. Furthermore, the Mortality Probability Model II uses logistic regression based on independent risk factors to predict a patient’s probability of mortality. An important overlooked limitation of SAPS II and MPM II is that they do not, to date, include interaction terms between a patient’s vital signs. This is a prominent oversight as it is likely there is an interplay among vital signs. The co-existence of certain conditions may pose a greater health risk than when these conditions exist independently. One barrier to including such interaction terms in predictive models is the dimensionality issue as it becomes difficult to use variable selection. We propose an innovative scoring system which takes into account a dependence structure among patient’s vital signs, such as systolic and diastolic blood pressures, heart rate, pulse interval, and peripheral oxygen saturation. Copulas will capture the dependence among normally distributed and skewed variables as some of the vital sign distributions are skewed. The estimated dependence parameter will then be incorporated into the traditional scoring systems to adjust the points allocated for the individual vital sign measurements. The same dependence parameter will also be used to create an alternative copula-based model for predicting a patient’s probability of mortality. The new copula-based approach will accommodate not only a patient’s trajectories of vital signs but also the joint dependence probabilities among the vital signs. We hypothesise that this approach will produce more stable assessments and lead to more time efficient and accurate predictions. We will use two data sets: (1) 250 ICU patients admitted once to the Chui Regional Hospital (Kyrgyzstan) and (2) 37 ICU patients’ agitation-sedation profiles collected by the Hunter Medical Research Institute (Australia). Both the traditional scoring approach and our copula-based approach will be evaluated using the Brier score to indicate overall model performance, the concordance (or c) statistic to indicate the discriminative ability (or area under the receiver operating characteristic (ROC) curve), and goodness-of-fit statistics for calibration. We will also report discrimination and calibration values and establish visualization of the copulas and high dimensional regions of risk interrelating two or three vital signs in so-called higher dimensional ROCs.Keywords: copula, intensive unit scoring system, ROC curves, vital sign dependence
Procedia PDF Downloads 152538 Neuroevolution Based on Adaptive Ensembles of Biologically Inspired Optimization Algorithms Applied for Modeling a Chemical Engineering Process
Authors: Sabina-Adriana Floria, Marius Gavrilescu, Florin Leon, Silvia Curteanu, Costel Anton
Abstract:
Neuroevolution is a subfield of artificial intelligence used to solve various problems in different application areas. Specifically, neuroevolution is a technique that applies biologically inspired methods to generate neural network architectures and optimize their parameters automatically. In this paper, we use different biologically inspired optimization algorithms in an ensemble strategy with the aim of training multilayer perceptron neural networks, resulting in regression models used to simulate the industrial chemical process of obtaining bricks from silicone-based materials. Installations in the raw ceramics industry, i.e., bricks, are characterized by significant energy consumption and large quantities of emissions. In addition, the initial conditions that were taken into account during the design and commissioning of the installation can change over time, which leads to the need to add new mixes to adjust the operating conditions for the desired purpose, e.g., material properties and energy saving. The present approach follows the study by simulation of a process of obtaining bricks from silicone-based materials, i.e., the modeling and optimization of the process. Optimization aims to determine the working conditions that minimize the emissions represented by nitrogen monoxide. We first use a search procedure to find the best values for the parameters of various biologically inspired optimization algorithms. Then, we propose an adaptive ensemble strategy that uses only a subset of the best algorithms identified in the search stage. The adaptive ensemble strategy combines the results of selected algorithms and automatically assigns more processing capacity to the more efficient algorithms. Their efficiency may also vary at different stages of the optimization process. In a given ensemble iteration, the most efficient algorithms aim to maintain good convergence, while the less efficient algorithms can improve population diversity. The proposed adaptive ensemble strategy outperforms the individual optimizers and the non-adaptive ensemble strategy in convergence speed, and the obtained results provide lower error values.Keywords: optimization, biologically inspired algorithm, neuroevolution, ensembles, bricks, emission minimization
Procedia PDF Downloads 116537 Customer Experiences and Perspectives on Mobile Money Service Fraud: A Case Study of the University of Education, Winneba
Authors: Mavis Ofosuah Asante, Abena Abokoma Asemanyi, Belinda Osei-mensah, Stephen Osei Akyiaw
Abstract:
The study examined mobile money service fraud experiences and perspectives on control practices at University of Education, Winneba. The objectives of the study included to examine the forms of MoMo fraud strategies experienced by customers of MoMo on UEW Campus, to examine and classify the main perpetrators of the MoMo fraud among UEW students as well as the framework for fraud detection put together by the Telco’s and consumers on UEW Campus. The study adopted the case study research design. The purposive sampling technique was used to select the UEW Campus. Using the convenience sampling technique, five respondents were sampled for the study. The outcome of the in-depth interviews conducted revealed Mobile money fraud was committed in various forms, such as anonymous calls and text messages from scammers, fraudsters calling to deceive subscribers that they are to deliver goods from abroad or from a close relative under false pretexts. Finally, fraudsters sending false cash-out messages to merchants for authorization of which the physical cash is issued by the merchant to the fraudster without the equivalent e-cash. Mobile money fraud has been perpetuated in diverse forms such as mobile money network systems fraud, false promotion fraud, and reversal of erroneous transactions, fortuitous scams, and mobile money agents' fraud. Finally, the frameworks that have been used to detect mobile money fraud include the display of national identifies cards for the transaction, digital identification systems, the use of firewall to protect mobile money accounts, effective information technology architecture for mobile money services, reporting of mobile money fraud to telecoms and the sanctioning of mobile money fraudsters. The study suggested there should be public education and awareness creation on the activities of mobile money fraudsters in Ghana by telecommunication companies in conjunction with the National Communications Authority and the Bank of Ghana. The study, therefore, concluded that the menace of mobile money fraud threatens the integrity of the mobile money financial services.Keywords: mobile money, fraud, telecommunication, merchant
Procedia PDF Downloads 78536 The Estimation Method of Stress Distribution for Beam Structures Using the Terrestrial Laser Scanning
Authors: Sang Wook Park, Jun Su Park, Byung Kwan Oh, Yousok Kim, Hyo Seon Park
Abstract:
This study suggests the estimation method of stress distribution for the beam structures based on TLS (Terrestrial Laser Scanning). The main components of method are the creation of the lattices of raw data from TLS to satisfy the suitable condition and application of CSSI (Cubic Smoothing Spline Interpolation) for estimating stress distribution. Estimation of stress distribution for the structural member or the whole structure is one of the important factors for safety evaluation of the structure. Existing sensors which include ESG (Electric strain gauge) and LVDT (Linear Variable Differential Transformer) can be categorized as contact type sensor which should be installed on the structural members and also there are various limitations such as the need of separate space where the network cables are installed and the difficulty of access for sensor installation in real buildings. To overcome these problems inherent in the contact type sensors, TLS system of LiDAR (light detection and ranging), which can measure the displacement of a target in a long range without the influence of surrounding environment and also get the whole shape of the structure, has been applied to the field of structural health monitoring. The important characteristic of TLS measuring is a formation of point clouds which has many points including the local coordinate. Point clouds is not linear distribution but dispersed shape. Thus, to analyze point clouds, the interpolation is needed vitally. Through formation of averaged lattices and CSSI for the raw data, the method which can estimate the displacement of simple beam was developed. Also, the developed method can be extended to calculate the strain and finally applicable to estimate a stress distribution of a structural member. To verify the validity of the method, the loading test on a simple beam was conducted and TLS measured it. Through a comparison of the estimated stress and reference stress, the validity of the method is confirmed.Keywords: structural healthcare monitoring, terrestrial laser scanning, estimation of stress distribution, coordinate transformation, cubic smoothing spline interpolation
Procedia PDF Downloads 433535 Women Perception of Spatial Safety Relating to Working in Historic Cairo’s Retail Street Markets
Authors: Toka M. Abufarag
Abstract:
This research primarily studies the correlation between the existence of different spatial factors in relation to the perception of females towards safely participating in the labor force within selected areas of economic bustle in Historic Cairo. This research measures the following independent variables: (1) perception regarding spatial safety on the street as controlled by street network, (2) vegetation as a facilitator and inhibitor of feeling safe in public places, and (3) outdoor lighting; in relation to the following dependent variable: the perception of females towards safely participating in the labor force in Historic Cairo. The objective of this research lies within adding to the design guidelines of urban design and planning in terms of design recommendations, making them more inclusive, especially those dealing with conserving and enhancing the built environment of old and historic cities. It is hypothesized that a balanced male-to-female ratio in terms of street activity, increased visibility of street in terms of its volume, a decrease in street obstacles, creation of open sighted vegetation, and increased visibility due to proper lighting will show up as positive response relating to the female perception of safety. The site chosen as an area to host this exercise of data collection is Al-Ataba. The site is within the borders of Historic Cairo and was chosen for two reasons: firstly, it provides a major source of economic bustle in Historic Cairo; and secondly, it hosts retail economic activities. This is a cross-sectional study. The data collected will consist of three parts: (1) observations by the researcher regarding the percentage of female participation, as well as perception of females on site, (2) interviews with women working on-site regarding the percentage of female participation, as well as their perception on participating, and (3) an anonymous online survey that studies the perception of a random sample of women towards the site as a place to exist in. The survey will aid in producing design recommendations on how to design an open 'souk' that suits women’s perception of a safe space.Keywords: urban design, women empowerment, safety perception, street markets, historic Cairo
Procedia PDF Downloads 125534 Determination of the Structural Parameters of Calcium Phosphate for Biomedical Use
Authors: María Magdalena Méndez-González, Miguel García Rocha, Carlos Manuel Yermo De la Cruz
Abstract:
Calcium phosphate (Ca5(PO4)3(X)) is widely used in orthopedic applications and is widely used as powder and granules. However, their presence in bone is in the form of nanometric needles 60 nm in length with a non-stoichiometric phase of apatite contains CO3-2, Na+, OH-, F-, and other ions in a matrix of collagen fibers. The crystal size, morphology control and interaction with cells are essential for the development of nanotechnology. The structural results of calcium phosphate, synthesized by chemical precipitation with crystal size of 22.85 nm are presented in this paper. The calcium phosphate powders were analyzed by X-ray diffraction, energy dispersive spectroscopy (EDS), infrared spectroscopy and FT-IR transmission electron microscopy. Network parameters, atomic positions, the indexing of the planes and the calculation of FWHM (full width at half maximum) were obtained. The crystal size was also calculated using the Scherer equation d (hkl) = cλ/βcosѲ. Where c is a constant related to the shape of the crystal, the wavelength of the radiation used for a copper anode is 1.54060Å, Ѳ is the Bragg diffraction angle, and β is the width average peak height of greater intensity. Diffraction pattern corresponding to the calcium phosphate called hydroxyapatite phase of a hexagonal crystal system was obtained. It belongs to the space group P63m with lattice parameters a = 9.4394 Å and c = 6.8861 Å. The most intense peak is obtained 2Ѳ = 31.55 (FWHM = 0.4798), with a preferred orientation in 121. The intensity difference between the experimental data and the calculated values is attributable to the temperature at which the sintering was performed. The intensity of the highest peak is at angle 2Ѳ = 32.11. The structure of calcium phosphate obtained was a hexagonal configuration. The intensity changes in the peaks of the diffraction pattern, in the lattice parameters at the corners, indicating the possible presence of a dopant. That each calcium atom is surrounded by a tetrahedron of oxygen and hydrogen was observed by infrared spectra. The unit cell pattern corresponds to hydroxyapatite and transmission electron microscopic crystal morphology corresponding to the hexagonal phase with a preferential growth along the c-plane was obtained.Keywords: structure, nanoparticles, calcium phosphate, metallurgical and materials engineering
Procedia PDF Downloads 504533 Geospatial Multi-Criteria Evaluation to Predict Landslide Hazard Potential in the Catchment of Lake Naivasha, Kenya
Authors: Abdel Rahman Khider Hassan
Abstract:
This paper describes a multi-criteria geospatial model for prediction of landslide hazard zonation (LHZ) for Lake Naivasha catchment (Kenya), based on spatial analysis of integrated datasets of location intrinsic parameters (slope stability factors) and external landslides triggering factors (natural and man-made factors). The intrinsic dataset included: lithology, geometry of slope (slope inclination, aspect, elevation, and curvature) and land use/land cover. The landslides triggering factors included: rainfall as the climatic factor, in addition to the destructive effects reflected by proximity of roads and drainage network to areas that are susceptible to landslides. No published study on landslides has been obtained for this area. Thus, digital datasets of the above spatial parameters were conveniently acquired, stored, manipulated and analyzed in a Geographical Information System (GIS) using a multi-criteria grid overlay technique (in ArcGIS 10.2.2 environment). Deduction of landslide hazard zonation is done by applying weights based on relative contribution of each parameter to the slope instability, and finally, the weighted parameters grids were overlaid together to generate a map of the potential landslide hazard zonation (LHZ) for the lake catchment. From the total surface of 3200 km² of the lake catchment, most of the region (78.7 %; 2518.4 km²) is susceptible to moderate landslide hazards, whilst about 13% (416 km²) is occurring under high hazards. Only 1.0% (32 km²) of the catchment is displaying very high landslide hazards, and the remaining area (7.3 %; 233.6 km²) displays low probability of landslide hazards. This result confirms the importance of steep slope angles, lithology, vegetation land cover and slope orientation (aspect) as the major determining factors of slope failures. The information provided by the produced map of landslide hazard zonation (LHZ) could lay the basis for decision making as well as mitigation and applications in avoiding potential losses caused by landslides in the Lake Naivasha catchment in the Kenya Highlands.Keywords: decision making, geospatial, landslide, multi-criteria, Naivasha
Procedia PDF Downloads 206532 Overcoming Challenges of Teaching English as a Foreign Language in Technical Classrooms: A Case Study at TVTC College of Technology
Authors: Sreekanth Reddy Ballarapu
Abstract:
The perception of the whole process of teaching and learning is undergoing a drastic and radical change. More and more student-centered, pragmatic, and flexible approaches are gradually replacing teacher-centered lecturing and structural-syllabus instruction. The issue of teaching English as a Foreign language is no exception in this regard. The traditional Present-Practice-Produce (P-P-P) method of teaching English is overtaken by Task-Based Teaching which is a subsidiary branch of Communicative Language Teaching. At this juncture this article strongly tries to convey that - Task-based learning, has an advantage over other traditional methods of teaching. All teachers of English must try to customize their texts into productive tasks, apply them, and evaluate the students as well as themselves. Task Based Learning is a double edged tool which can enhance the performance of both the teacher and the taught. The sample for this case study is a class of 35 students from Semester III - Network branch at TVTC College of Technology, Adhum - Kingdom of Saudi Arabia. The students are high school passed out and aged between 19-21years.For the present study the prescribed textbook Technical English 1 by David Bonamy was used and a number of language tasks were chalked out during the pre- task stage and the learners were made to participate voluntarily and actively. The Action Research methodology was adopted within the dual framework of Communicative Language Teaching and Task-Based Learning. The different tools such as questionnaires, feedback and interviews were used to collect data. This study provides information about various techniques of Communicative Language Teaching and Task Based Learning and focuses primarily on the advantages of using a Task Based Learning approach. This article presents in detail the objectives of the study, the planning and implementation of the action research, the challenges encountered during the execution of the plan, and the pedagogical outcome of this project. These research findings serve two purposes: first, it evaluates the effectiveness of Task Based Learning and, second, it empowers the teacher's professionalism in designing and implementing the tasks. In the end, the possibility of scope for further research is presented in brief.Keywords: action research, communicative language teaching, task based learning, perception
Procedia PDF Downloads 238531 Sustainable Integrated Waste Management System
Authors: Lidia Lombardi
Abstract:
Waste management in Europe and North America is evolving towards sustainable materials management, intended as a systemic approach to using and reusing materials more productively over their entire life cycles. Various waste management strategies are prioritized and ranked from the most to the least environmentally preferred, placing emphasis on reducing, reusing, and recycling as key to sustainable materials management. However, non-recyclable materials must also be appropriately addressed, and waste-to-energy (WtE) offers a solution to manage them, especially when a WtE plant is integrated within a complex system of waste and wastewater treatment plants and potential users of the output flows. To evaluate the environmental effects of such system integration, Life Cycle Assessment (LCA) is a helpful and powerful tool. LCA has been largely applied to the waste management sector, dating back to the late 1990s, producing a large number of theoretical studies and applications to the real world as support to waste management planning. However, LCA still has a fundamental role in helping the development of waste management systems supporting decisions. Thus, LCA was applied to evaluate the environmental performances of a Municipal Solid Waste (MSW) management system, with improved separate material collection and recycling and an integrated network of treatment plants including WtE, anaerobic digestion (AD) and also wastewater treatment plant (WWTP), for a reference study case area. The proposed system was compared to the actual situation, characterized by poor recycling, large landfilling and absence of WtE. The LCA results showed that the increased recycling significantly increases the environmental performances, but there is still room for improvement through the introduction of energy recovery (especially by WtE) and through its use within the system, for instance, by feeding the heat to the AD, to sludge recovery processes and supporting the water reuse practice. WtE offers a solution to manage non-recyclable MSW and allows saving important resources (such as landfill volumes and non-renewable energy), reducing the contribution to global warming, and providing an essential contribution to fulfill the goals of really sustainable waste management.Keywords: anaerobic digestion, life cycle assessment, waste-to-energy, municipal solid waste
Procedia PDF Downloads 60530 Building Social Capital for Social Inclusion: The Use of Social Networks in Government
Authors: Suha Alawadhi, Malak Alrasheed
Abstract:
In the recent past, public participation in governments has been declined to a great extent, as citizens have been isolated from community life and their ability to articulate demands for good government has been noticeably decreased. However, the Internet has introduced new forms of interaction that could enhance different types of relationships, including government-public relationship. In fact, technology-enabled government has become a catalyst for enabling social inclusion. This exploratory study seeks to investigate public perceptions in Kuwait regarding the use of social media networks in government where social capital is built to achieve social inclusion. Social capital has been defined as social networks and connections amongst individuals, that are based on shared trust, ideas and norms, enable participants of a network to act effectively to pursue a shared objective. The quantitative method was used to generate empirical evidence. A questionnaire was designed to address the research objective and reflect the identified constructs: social capital dimensions (bridging, bonding and maintaining social capital), social inclusion, and social equality. In this pilot study, data was collected from a random sample of 61 subjects. The results indicate that all participants have a positive attitude towards the dimensions of social capital (bridging, bonding and maintaining), social inclusion and social equality constructs. Tests of identified constructs against demographic characteristics indicate that there are significant differences between male and female as they perceived bonding and maintaining social capital, social inclusion and social equality whereas no difference was identified in their perceptions of bridging social capital. Also, those who are aged 26-30 perceived bonding and maintaining social capital, social inclusion and social equality negatively compared to those aged 20-25, 31-35, and 40-above whose perceptions were positive. With regard to education, the results also show that those holding high school, university degree and diploma perceived maintaining social capital positively higher than with those who hold graduate degrees. Moreover, a regression model is proposed to study the effect of bridging, bonding, and maintaining social capital on social inclusion via social equality as a mediator. This exploratory study is necessary for testing the validity and reliability of the questionnaire which will be used in the main study that aims to investigate the perceptions of individuals towards building social capital to achieve social inclusion.Keywords: government, social capital, social inclusion, social networks
Procedia PDF Downloads 326529 Large-Scale Simulations of Turbulence Using Discontinuous Spectral Element Method
Authors: A. Peyvan, D. Li, J. Komperda, F. Mashayek
Abstract:
Turbulence can be observed in a variety fluid motions in nature and industrial applications. Recent investment in high-speed aircraft and propulsion systems has revitalized fundamental research on turbulent flows. In these systems, capturing chaotic fluid structures with different length and time scales is accomplished through the Direct Numerical Simulation (DNS) approach since it accurately simulates flows down to smallest dissipative scales, i.e., Kolmogorov’s scales. The discontinuous spectral element method (DSEM) is a high-order technique that uses spectral functions for approximating the solution. The DSEM code has been developed by our research group over the course of more than two decades. Recently, the code has been improved to run large cases in the order of billions of solution points. Running big simulations requires a considerable amount of RAM. Therefore, the DSEM code must be highly parallelized and able to start on multiple computational nodes on an HPC cluster with distributed memory. However, some pre-processing procedures, such as determining global element information, creating a global face list, and assigning global partitioning and element connection information of the domain for communication, must be done sequentially with a single processing core. A separate code has been written to perform the pre-processing procedures on a local machine. It stores the minimum amount of information that is required for the DSEM code to start in parallel, extracted from the mesh file, into text files (pre-files). It packs integer type information with a Stream Binary format in pre-files that are portable between machines. The files are generated to ensure fast read performance on different file-systems, such as Lustre and General Parallel File System (GPFS). A new subroutine has been added to the DSEM code to read the startup files using parallel MPI I/O, for Lustre, in a way that each MPI rank acquires its information from the file in parallel. In case of GPFS, in each computational node, a single MPI rank reads data from the file, which is specifically generated for the computational node, and send them to other ranks on the node using point to point non-blocking MPI communication. This way, communication takes place locally on each node and signals do not cross the switches of the cluster. The read subroutine has been tested on Argonne National Laboratory’s Mira (GPFS), National Center for Supercomputing Application’s Blue Waters (Lustre), San Diego Supercomputer Center’s Comet (Lustre), and UIC’s Extreme (Lustre). The tests showed that one file per node is suited for GPFS and parallel MPI I/O is the best choice for Lustre file system. The DSEM code relies on heavily optimized linear algebra operation such as matrix-matrix and matrix-vector products for calculation of the solution in every time-step. For this, the code can either make use of its matrix math library, BLAS, Intel MKL, or ATLAS. This fact and the discontinuous nature of the method makes the DSEM code run efficiently in parallel. The results of weak scaling tests performed on Blue Waters showed a scalable and efficient performance of the code in parallel computing.Keywords: computational fluid dynamics, direct numerical simulation, spectral element, turbulent flow
Procedia PDF Downloads 133528 Transport Hubs as Loci of Multi-Layer Ecosystems of Innovation: Case Study of Airports
Authors: Carolyn Hatch, Laurent Simon
Abstract:
Urban mobility and the transportation industry are undergoing a transformation, shifting from an auto production-consumption model that has dominated since the early 20th century towards new forms of personal and shared multi-modality [1]. This is shaped by key forces such as climate change, which has induced a shift in production and consumption patterns and efforts to decarbonize and improve transport services through, for instance, the integration of vehicle automation, electrification and mobility sharing [2]. Advanced innovation practices and platforms for experimentation and validation of new mobility products and services that are increasingly complex and multi-stakeholder-oriented are shaping this new world of mobility. Transportation hubs – such as airports - are emblematic of these disruptive forces playing out in the mobility industry. Airports are emerging as the core of innovation ecosystems on and around contemporary mobility issues, and increasingly recognized as complex public/private nodes operating in many societal dimensions [3,4]. These include urban development, sustainability transitions, digital experimentation, customer experience, infrastructure development and data exploitation (for instance, airports generate massive and often untapped data flows, with significant potential for use, commercialization and social benefit). Yet airport innovation practices have not been well documented in the innovation literature. This paper addresses this gap by proposing a model of airport innovation that aims to equip airport stakeholders to respond to these new and complex innovation needs in practice. The methodology involves: 1 – a literature review bringing together key research and theory on airport innovation management, open innovation and innovation ecosystems in order to evaluate airport practices through an innovation lens; 2 – an international benchmarking of leading airports and their innovation practices, including such examples as Aéroports de Paris, Schipol in Amsterdam, Changi in Singapore, and others; and 3 – semi-structured interviews with airport managers on key aspects of organizational practice, facilitated through a close partnership with the Airport Council International (ACI), a major stakeholder in this research project. Preliminary results find that the most successful airports are those that have shifted to a multi-stakeholder, platform ecosystem model of innovation. The recent entrance of new actors in airports (Google, Amazon, Accor, Vinci, Airbnb and others) have forced the opening of organizational boundaries to share and exchange knowledge with a broader set of ecosystem players. This has also led to new forms of governance and intermediation by airport actors to connect complex, highly distributed knowledge, along with new kinds of inter-organizational collaboration, co-creation and collective ideation processes. Leading airports in the case study have demonstrated a unique capacity to force traditionally siloed activities to “think together”, “explore together” and “act together”, to share data, contribute expertise and pioneer new governance approaches and collaborative practices. In so doing, they have successfully integrated these many disruptive change pathways and forced their implementation and coordination towards innovative mobility outcomes, with positive societal, environmental and economic impacts. This research has implications for: 1 - innovation theory, 2 - urban and transport policy, and 3 - organizational practice - within the mobility industry and across the economy.Keywords: airport management, ecosystem, innovation, mobility, platform, transport hubs
Procedia PDF Downloads 181527 Charcoal Traditional Production in Portugal: Contribution to the Quantification of Air Pollutant Emissions
Authors: Cátia Gonçalves, Teresa Nunes, Inês Pina, Ana Vicente, C. Alves, Felix Charvet, Daniel Neves, A. Matos
Abstract:
The production of charcoal relies on rudimentary technologies using traditional brick kilns. Charcoal is produced under pyrolysis conditions: breaking down the chemical structure of biomass under high temperature in the absence of air. The amount of the pyrolysis products (charcoal, pyroligneous extract, and flue gas) depends on various parameters, including temperature, time, pressure, kiln design, and wood characteristics like the moisture content. This activity is recognized for its inefficiency and high pollution levels, but it is poorly characterized. This activity is widely distributed and is a vital economic activity in certain regions of Portugal, playing a relevant role in the management of woody residues. The location of the units establishes the biomass used for charcoal production. The Portalegre district, in the Alto Alentejo region (Portugal), is a good example, essentially with rural characteristics, with a predominant farming, agricultural, and forestry profile, and with a significant charcoal production activity. In this district, a recent inventory identifies almost 50 charcoal production units, equivalent to more than 450 kilns, of which 80% appear to be in operation. A field campaign was designed with the objective of determining the composition of the emissions released during a charcoal production cycle. A total of 30 samples of particulate matter and 20 gas samples in Tedlar bags were collected. Particulate and gas samplings were performed in parallel, 2 in the morning and 2 in the afternoon, alternating the inlet heads (PM₁₀ and PM₂.₅), in the particulate sampler. The gas and particulate samples were collected in the plume as close as the emission chimney point. The biomass (dry basis) used in the carbonization process was a mixture of cork oak (77 wt.%), holm oak (7 wt.%), stumps (11 wt.%), and charred wood (5 wt.%) from previous carbonization processes. A cylindrical batch kiln (80 m³) with 4.5 m diameter and 5 m of height was used in this study. The composition of the gases was determined by gas chromatography, while the particulate samples (PM₁₀, PM₂.₅) were subjected to different analytical techniques (thermo-optical transmission technique, ion chromatography, HPAE-PAD, and GC-MS after solvent extraction) after prior gravimetric determination, to study their organic and inorganic constituents. The charcoal production cycle presents widely varying operating conditions, which will be reflected in the composition of gases and particles produced and emitted throughout the process. The concentration of PM₁₀ and PM₂.₅ in the plume was calculated, ranging between 0.003 and 0.293 g m⁻³, and 0.004 and 0.292 g m⁻³, respectively. Total carbon, inorganic ions, and sugars account, in average, for PM10 and PM₂.₅, 65 % and 56 %, 2.8 % and 2.3 %, 1.27 %, and 1.21 %, respectively. The organic fraction studied until now includes more than 30 aliphatic compounds and 20 PAHs. The emission factors of particulate matter to produce charcoal in the traditional kiln were 33 g/kg (wooddb) and 27 g/kg (wooddb) for PM₁₀ and PM₂.₅, respectively. With the data obtained in this study, it is possible to fill the lack of information about the environmental impact of the traditional charcoal production in Portugal. Acknowledgment: Authors thanks to FCT – Portuguese Science Foundation, I.P. and to Ministry of Science, Technology and Higher Education of Portugal for financial support within the scope of the project CHARCLEAN (PCIF/GVB/0179/2017) and CESAM (UIDP/50017/2020 + UIDB/50017/2020).Keywords: brick kilns, charcoal, emission factors, PAHs, total carbon
Procedia PDF Downloads 142526 Effect of a Chatbot-Assisted Adoption of Self-Regulated Spaced Practice on Students' Vocabulary Acquisition and Cognitive Load
Authors: Ngoc-Nguyen Nguyen, Hsiu-Ling Chen, Thanh-Truc Lai Huynh
Abstract:
In foreign language learning, vocabulary acquisition has consistently posed challenges to learners, especially for those at lower levels. Conventional approaches often fail to promote vocabulary learning and ensure engaging experiences alike. The emergence of mobile learning, particularly the integration of chatbot systems, has offered alternative ways to facilitate this practice. Chatbots have proven effective in educational contexts by offering interactive learning experiences in a constructivist manner. These tools have caught attention in the field of mobile-assisted language learning (MALL) in recent years. This research is conducted in an English for Specific Purposes (ESP) course at the A2 level of the CEFR, designed for non-English majors. Participants are first-year Vietnamese students aged 18 to 20 at a university. This quasi-experimental study follows a pretest-posttest control group design over five weeks, with two classes randomly assigned as the experimental and control groups. The experimental group engages in chatbot-assisted spaced practice with SRL components, while the control group uses the same spaced practice without SRL. The two classes are taught by the same lecturer. Data are collected through pre- and post-tests, cognitive load surveys, and semi-structured interviews. The combination of self-regulated learning (SRL) and distributed practice, grounded in the spacing effect, forms the basis of the present study. SRL elements, which concern goal setting and strategy planning, are integrated into the system. The spaced practice method, similar to those used in widely recognized learning platforms like Duolingo and Anki flashcards, spreads out learning over multiple sessions. This study’s design features quizzes progressively increasing in difficulty. These quizzes are aimed at targeting both the Recognition-Recall and Comprehension-Use dimensions for a comprehensive acquisition of vocabulary. The mobile-based chatbot system is built using Golang, an open-source programming language developed by Google. It follows a structured flow that guides learners through a series of 4 quizzes in each week of teacher-led learning. The quizzes start with less cognitively demanding tasks, such as multiple-choice questions, before moving on to more complex exercises. The integration of SRL elements allows students to self-evaluate the difficulty level of vocabulary items, predict scores achieved, and choose appropriate strategy. This research is part one of a two-part project. The initial findings will determine the development of an upgraded chatbot system in part two, where adaptive features in response to the integration of SRL components will be introduced. The research objectives are to assess the effectiveness of the chatbot-assisted approach, based on the combination of spaced practice and SRL, in improving vocabulary acquisition and managing cognitive load, as well as to understand students' perceptions of this learning tool. The insights from this study will contribute to the growing body of research on mobile-assisted language learning and offer practical implications for integrating chatbot systems with spaced practice into educational settings to enhance vocabulary learning.Keywords: mobile learning, mobile-assisted language learning, MALL, chatbots, vocabulary learning, spaced practice, spacing effect, self-regulated learning, SRL, self-regulation, EFL, cognitive load
Procedia PDF Downloads 19525 Technique for Online Condition Monitoring of Surge Arresters
Authors: Anil S. Khopkar, Kartik S. Pandya
Abstract:
Overvoltage in power systems is a phenomenon that cannot be avoided. However, it can be controlled to a certain extent. Power system equipment is to be protected against overvoltage to avoid system failure. Metal Oxide Surge Arresters (MOSA) are connected to the system for the protection of the power system against overvoltages. The MOSA will behave as an insulator under normal working conditions, where it offers a conductive path under voltage conditions. MOSA consists of zinc oxide elements (ZnO Blocks), which have non-linear V-I characteristics. ZnO blocks are connected in series and fitted in ceramic or polymer housing. This degrades due to the aging effect under continuous operation. Degradation of zinc oxide elements increases the leakage current flowing from the surge arresters. This Increased leakage current results in the increased temperature of the surge arrester, which further decreases the resistance of zinc oxide elements. As a result, leakage current increases, which again increases the temperature of a MOSA. This creates thermal runaway conditions for MOSA. Once it reaches the thermal runaway condition, it cannot return to normal working conditions. This condition is a primary cause of premature failure of surge arresters, as MOSA constitutes a core protective device for electrical power systems against transients. It contributes significantly to the reliable operation of the power system network. Hence, the condition monitoring of surge arresters should be done at periodic intervals. Online and Offline condition monitoring techniques are available for surge arresters. Offline condition monitoring techniques are not very popular as they require removing surge arresters from the system, which requires system shutdown. Hence, online condition monitoring techniques are very popular. This paper presents the evaluation technique for the surge arrester condition based on the leakage current analysis. Maximum amplitude of total leakage current (IT), Maximum amplitude of fundamental resistive leakage current (IR) and maximum amplitude of third harmonic resistive leakage current (I3rd) have been analyzed as indicators for surge arrester condition monitoring.Keywords: metal oxide surge arrester (MOSA), over voltage, total leakage current, resistive leakage current
Procedia PDF Downloads 66524 Data and Model-based Metamodels for Prediction of Performance of Extended Hollo-Bolt Connections
Authors: M. Cabrera, W. Tizani, J. Ninic, F. Wang
Abstract:
Open section beam to concrete-filled tubular column structures has been increasingly utilized in construction over the past few decades due to their enhanced structural performance, as well as economic and architectural advantages. However, the use of this configuration in construction is limited due to the difficulties in connecting the structural members as there is no access to the inner part of the tube to install standard bolts. Blind-bolted systems are a relatively new approach to overcome this limitation as they only require access to one side of the tubular section to tighten the bolt. The performance of these connections in concrete-filled steel tubular sections remains uncharacterized due to the complex interactions between concrete, bolt, and steel section. Over the last years, research in structural performance has moved to a more sophisticated and efficient approach consisting of machine learning algorithms to generate metamodels. This method reduces the need for developing complex, and computationally expensive finite element models, optimizing the search for desirable design variables. Metamodels generated by a data fusion approach use numerical and experimental results by combining multiple models to capture the dependency between the simulation design variables and connection performance, learning the relations between different design parameters and predicting a given output. Fully characterizing this connection will transform high-rise and multistorey construction by means of the introduction of design guidance for moment-resisting blind-bolted connections, which is currently unavailable. This paper presents a review of the steps taken to develop metamodels generated by means of artificial neural network algorithms which predict the connection stress and stiffness based on the design parameters when using Extended Hollo-Bolt blind bolts. It also provides consideration of the failure modes and mechanisms that contribute to the deformability as well as the feasibility of achieving blind-bolted rigid connections when using the blind fastener.Keywords: blind-bolted connections, concrete-filled tubular structures, finite element analysis, metamodeling
Procedia PDF Downloads 158523 Crack Growth Life Prediction of a Fighter Aircraft Wing Splice Joint Under Spectrum Loading Using Random Forest Regression and Artificial Neural Networks with Hyperparameter Optimization
Authors: Zafer Yüce, Paşa Yayla, Alev Taşkın
Abstract:
There are heaps of analytical methods to estimate the crack growth life of a component. Soft computing methods have an increasing trend in predicting fatigue life. Their ability to build complex relationships and capability to handle huge amounts of data are motivating researchers and industry professionals to employ them for challenging problems. This study focuses on soft computing methods, especially random forest regressors and artificial neural networks with hyperparameter optimization algorithms such as grid search and random grid search, to estimate the crack growth life of an aircraft wing splice joint under variable amplitude loading. TensorFlow and Scikit-learn libraries of Python are used to build the machine learning models for this study. The material considered in this work is 7050-T7451 aluminum, which is commonly preferred as a structural element in the aerospace industry, and regarding the crack type; corner crack is used. A finite element model is built for the joint to calculate fastener loads and stresses on the structure. Since finite element model results are validated with analytical calculations, findings of the finite element model are fed to AFGROW software to calculate analytical crack growth lives. Based on Fighter Aircraft Loading Standard for Fatigue (FALSTAFF), 90 unique fatigue loading spectra are developed for various load levels, and then, these spectrums are utilized as inputs to the artificial neural network and random forest regression models for predicting crack growth life. Finally, the crack growth life predictions of the machine learning models are compared with analytical calculations. According to the findings, a good correlation is observed between analytical and predicted crack growth lives.Keywords: aircraft, fatigue, joint, life, optimization, prediction.
Procedia PDF Downloads 175