Search results for: Artificial Neural Networks (ANN)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4911

Search results for: Artificial Neural Networks (ANN)

321 Using Computer Vision and Machine Learning to Improve Facility Design for Healthcare Facility Worker Safety

Authors: Hengameh Hosseini

Abstract:

Design of large healthcare facilities – such as hospitals, multi-service line clinics, and nursing facilities - that can accommodate patients with wide-ranging disabilities is a challenging endeavor and one that is poorly understood among healthcare facility managers, administrators, and executives. An even less-understood extension of this problem is the implications of weakly or insufficiently accommodative design of facilities for healthcare workers in physically-intensive jobs who may also suffer from a range of disabilities and who are therefore at increased risk of workplace accident and injury. Combine this reality with the vast range of facility types, ages, and designs, and the problem of universal accommodation becomes even more daunting and complex. In this study, we focus on the implication of facility design for healthcare workers suffering with low vision who also have physically active jobs. The points of difficulty are myriad and could span health service infrastructure, the equipment used in health facilities, and transport to and from appointments and other services can all pose a barrier to health care if they are inaccessible, less accessible, or even simply less comfortable for people with various disabilities. We conduct a series of surveys and interviews with employees and administrators of 7 facilities of a range of sizes and ownership models in the Northeastern United States and combine that corpus with in-facility observations and data collection to identify five major points of failure common to all the facilities that we concluded could pose safety threats to employees with vision impairments, ranging from very minor to severe. We determine that lack of design empathy is a major commonality among facility management and ownership. We subsequently propose three methods for remedying this lack of empathy-informed design, to remedy the dangers posed to employees: the use of an existing open-sourced Augmented Reality application to simulate the low-vision experience for designers and managers; the use of a machine learning model we develop to automatically infer facility shortcomings from large datasets of recorded patient and employee reviews and feedback; and the use of a computer vision model fine tuned on images of each facility to infer and predict facility features, locations, and workflows, that could again pose meaningful dangers to visually impaired employees of each facility. After conducting a series of real-world comparative experiments with each of these approaches, we conclude that each of these are viable solutions under particular sets of conditions, and finally characterize the range of facility types, workforce composition profiles, and work conditions under which each of these methods would be most apt and successful.

Keywords: artificial intelligence, healthcare workers, facility design, disability, visually impaired, workplace safety

Procedia PDF Downloads 83
320 Trophic Variations in Uptake and Assimilation of Cadmium, Manganese and Zinc: An Estuarine Food-Chain Radiotracer Experiment

Authors: K. O’Mara, T. Cresswell

Abstract:

Nearly half of the world’s population live near the coast, and as a result, estuaries and coastal bays in populated or industrialized areas often receive metal pollution. Heavy metals have a chemical affinity for sediment particles and can be stored in estuarine sediments and become biologically available under changing conditions. Organisms inhabiting estuaries can be exposed to metals from a variety of sources including metals dissolved in water, bound to sediment or within contaminated prey. Metal uptake and assimilation responses can vary even between species that are biologically similar, making pollution effects difficult to predict. A multi-trophic level experiment representing a common Eastern Australian estuarine food chain was used to study the sources for Cd, Mn and Zn uptake and assimilation in organisms occupying several trophic levels. Sand cockles (Katelysia scalarina), school prawns (Metapenaeus macleayi) and sand whiting (Sillago ciliata) were exposed to radiolabelled seawater, suspended sediment and food. Three pulse-chase trials on filter-feeding sand cockles were performed using radiolabelled phytoplankton (Tetraselmis sp.), benthic microalgae (Entomoneis sp.) and suspended sediment. Benthic microalgae had lower metal uptake than phytoplankton during labelling but higher cockle assimilation efficiencies (Cd = 51%, Mn = 42%, Zn = 63 %) than both phytoplankton (Cd = 21%, Mn = 32%, Zn = 33%) and suspended sediment (except Mn; (Cd = 38%, Mn = 42%, Zn = 53%)). Sand cockles were also sensitive to uptake of Cd, Mn and Zn dissolved in seawater. Uptake of these metals from the dissolved phase was negligible in prawns and fish, with prawns only accumulating metals during moulting, which were then lost with subsequent moulting in the depuration phase. Diet appears to be the main source of metal assimilation in school prawns, with 65%, 54% and 58% assimilation efficiencies from Cd, Mn and Zn respectively. Whiting fed contaminated prawns were able to exclude the majority of the metal activity through egestion, with only 10%, 23% and 11% assimilation efficiencies from Cd, Mn and Zn respectively. The findings of this study support previous studies that find diet to be the dominant accumulation source for higher level trophic organisms. These results show that assimilation efficiencies can vary depending on the source of exposure; sand cockles assimilated more Cd, Mn, and Zn from the benthic diatom than phytoplankton and assimilation was higher in sand whiting fed prawns compared to artificial pellets. The sensitivity of sand cockles to metal uptake and assimilation from a variety of sources poses concerns for metal availability to predators ingesting the clam tissue, including humans. The high tolerance of sand whiting to these metals is reflected in their widespread presence in Eastern Australian estuaries, including contaminated estuaries such as Botany Bay and Port Jackson.

Keywords: cadmium, food chain, metal, manganese, trophic, zinc

Procedia PDF Downloads 182
319 Recommendations to Improve Classification of Grade Crossings in Urban Areas of Mexico

Authors: Javier Alfonso Bonilla-Chávez, Angélica Lozano

Abstract:

In North America, more than 2,000 people annually die in accidents related to railroad tracks. In 2020, collisions at grade crossings were the main cause of deaths related to railway accidents in Mexico. Railway networks have constant interaction with motor transport users, cyclists, and pedestrians, mainly in grade crossings, where is the greatest vulnerability and risk of accidents. Usually, accidents at grade crossings are directly related to risky behavior and non-compliance with regulations by motorists, cyclists, and pedestrians, especially in developing countries. Around the world, countries classify these crossings in different ways. In Mexico, according to their dangerousness (high, medium, or low), types A, B and C have been established, recommending for each one different type of auditive and visual signaling and gates, as well as horizontal and vertical signaling. This classification is based in a weighting, but regrettably, it is not explained how the weight values were obtained. A review of the variables and the current approach for the grade crossing classification is required, since it is inadequate for some crossings. In contrast, North America (USA and Canada) and European countries consider a broader classification so that attention to each crossing is addressed more precisely and equipment costs are adjusted. Lack of a proper classification, could lead to cost overruns in the equipment and a deficient operation. To exemplify the lack of a good classification, six crossings are studied, three located in the rural area of Mexico and three in Mexico City. These cases show the need of: improving the current regulations, improving the existing infrastructure, and implementing technological systems, including informative signals with nomenclature of the involved crossing and direct telephone line for reporting emergencies. This implementation is unaffordable for most municipal governments. Also, an inventory of the most dangerous grade crossings in urban and rural areas must be obtained. Then, an approach for improving the classification of grade crossings is suggested. This approach must be based on criteria design, characteristics of adjacent roads or intersections which can influence traffic flow through the crossing, accidents related to motorized and non-motorized vehicles, land use and land management, type of area, and services and economic activities in the zone where the grade crossings is located. An expanded classification of grade crossing in Mexico could reduce accidents and improve the efficiency of the railroad.

Keywords: accidents, grade crossing, railroad, traffic safety

Procedia PDF Downloads 93
318 Foreseen the Future: Human Factors Integration in European Horizon Projects

Authors: José Manuel Palma, Paula Pereira, Margarida Tomás

Abstract:

Foreseen the future: Human factors integration in European Horizon Projects The development of new technology as artificial intelligence, smart sensing, robotics, cobotics or intelligent machinery must integrate human factors to address the need to optimize systems and processes, thereby contributing to the creation of a safe and accident-free work environment. Human Factors Integration (HFI) consistently pose a challenge for organizations when applied to daily operations. AGILEHAND and FORTIS projects are grounded in the development of cutting-edge technology - industry 4.0 and 5.0. AGILEHAND aims to create advanced technologies for autonomously sort, handle, and package soft and deformable products, whereas FORTIS focuses on developing a comprehensive Human-Robot Interaction (HRI) solution. Both projects employ different approaches to explore HFI. AGILEHAND is mainly empirical, involving a comparison between the current and future work conditions reality, coupled with an understanding of best practices and the enhancement of safety aspects, primarily through management. FORTIS applies HFI throughout the project, developing a human-centric approach that includes understanding human behavior, perceiving activities, and facilitating contextual human-robot information exchange. it intervention is holistic, merging technology with the physical and social contexts, based on a total safety culture model. In AGILEHAND we will identify safety emergent risks, challenges, their causes and how to overcome them by resorting to interviews, questionnaires, literature review and case studies. Findings and results will be presented in “Strategies for Workers’ Skills Development, Health and Safety, Communication and Engagement” Handbook. The FORTIS project will implement continuous monitoring and guidance of activities, with a critical focus on early detection and elimination (or mitigation) of risks associated with the new technology, as well as guidance to adhere correctly with European Union safety and privacy regulations, ensuring HFI, thereby contributing to an optimized safe work environment. To achieve this, we will embed safety by design, and apply questionnaires, perform site visits, provide risk assessments, and closely track progress while suggesting and recommending best practices. The outcomes of these measures will be compiled in the project deliverable titled “Human Safety and Privacy Measures”. These projects received funding from European Union’s Horizon 2020/Horizon Europe research and innovation program under grant agreement No101092043 (AGILEHAND) and No 101135707 (FORTIS).

Keywords: human factors integration, automation, digitalization, human robot interaction, industry 4.0 and 5.0

Procedia PDF Downloads 38
317 A Double Ended AC Series Arc Fault Location Algorithm Based on Currents Estimation and a Fault Map Trace Generation

Authors: Edwin Calderon-Mendoza, Patrick Schweitzer, Serge Weber

Abstract:

Series arc faults appear frequently and unpredictably in low voltage distribution systems. Many methods have been developed to detect this type of faults and commercial protection systems such AFCI (arc fault circuit interrupter) have been used successfully in electrical networks to prevent damage and catastrophic incidents like fires. However, these devices do not allow series arc faults to be located on the line in operating mode. This paper presents a location algorithm for series arc fault in a low-voltage indoor power line in an AC 230 V-50Hz home network. The method is validated through simulations using the MATLAB software. The fault location method uses electrical parameters (resistance, inductance, capacitance, and conductance) of a 49 m indoor power line. The mathematical model of a series arc fault is based on the analysis of the V-I characteristics of the arc and consists basically of two antiparallel diodes and DC voltage sources. In a first step, the arc fault model is inserted at some different positions across the line which is modeled using lumped parameters. At both ends of the line, currents and voltages are recorded for each arc fault generation at different distances. In the second step, a fault map trace is created by using signature coefficients obtained from Kirchhoff equations which allow a virtual decoupling of the line’s mutual capacitance. Each signature coefficient obtained from the subtraction of estimated currents is calculated taking into account the Discrete Fast Fourier Transform of currents and voltages and also the fault distance value. These parameters are then substituted into Kirchhoff equations. In a third step, the same procedure described previously to calculate signature coefficients is employed but this time by considering hypothetical fault distances where the fault can appear. In this step the fault distance is unknown. The iterative calculus from Kirchhoff equations considering stepped variations of the fault distance entails the obtaining of a curve with a linear trend. Finally, the fault distance location is estimated at the intersection of two curves obtained in steps 2 and 3. The series arc fault model is validated by comparing current registered from simulation with real recorded currents. The model of the complete circuit is obtained for a 49m line with a resistive load. Also, 11 different arc fault positions are considered for the map trace generation. By carrying out the complete simulation, the performance of the method and the perspectives of the work will be presented.

Keywords: indoor power line, fault location, fault map trace, series arc fault

Procedia PDF Downloads 119
316 Urban River As Living Infrastructure: Tidal Flooding And Sea Level Rise In A Working Waterway In Hampton Roads, Virginia

Authors: William Luke Hamel

Abstract:

Existing conceptions of urban flooding caused by tidal fluctuations and sea-level rise have been inadequately conceptualized by metrics of resilience and methods of flow modeling. While a great deal of research has been devoted to the effects of urbanization on pluvial flooding, the kind of tidal flooding experienced by locations like Hampton Roads, Virginia, has not been adequately conceptualized as being a result of human factors such as urbanization and gray infrastructure. Resilience from sea level rise and its associated flooding has been pioneered in the region with the 2015 Norfolk Resilience Plan from 100 Resilient Cities as well as the 2016 Norfolk Vision 2100 plan, which envisions different patterns of land use for the city. Urban resilience still conceptualizes the city as having the ability to maintain an equilibrium in the face of disruptions. This economic and social equilibrium relies on the Elizabeth River, narrowly conceptualized. Intentionally or accidentally, the river was made to be a piece of infrastructure. Its development was meant to serve the docks, shipyards, naval yards, and port infrastructure that gives the region so much of its economic life. Inasmuch as it functions to permit the movement of cargo; the raising and lowering of ships to be repaired, commissioned, or decommissioned; or the provisioning of military vessels, the river as infrastructure is functioning properly. The idea that the infrastructure is malfunctioning when high tides and sea-level rise create flooding is predicated on the idea that the infrastructure is truly a human creation and can be controlled. The natural flooding cycles of an urban river, combined with the action of climate change and sea-level rise, are only abnormal so much as they encroach on the development that first encroached on the river. The urban political ecology of water provides the ability to view the river as an infrastructural extension of urban networks while also calling for its emancipation from stationarity and human control. Understanding the river and city as a hydrosocial territory or as a socio-natural system liberates both actors from the duality of the natural and the social while repositioning river flooding as a normal part of coexistence on a floodplain. This paper argues for the adoption of an urban political ecology lens in the analysis and governance of urban rivers like the Elizabeth River as a departure from the equilibrium-seeking and stability metrics of urban resilience.

Keywords: urban flooding, political ecology, Elizabeth river, Hampton roads

Procedia PDF Downloads 146
315 Knowledge Management in Public Sector Employees: A Case Study of Training Participants at National Institute of Management, Pakistan

Authors: Muhammad Arif Khan, Haroon Idrees, Imran Aziz, Sidra Mushtaq

Abstract:

The purpose of this study is to investigate the current level of knowledge mapping skills of the public sector employees in Pakistan. National Institute of Management is one of the premiere public sector training organization for mid-career public sector employees in Pakistan. This study is conducted on participants of fourteen weeks long training course called Mid-Career Management Course (MCMC) which is mandatory for public sector employees in order to ascertain how to enhance their knowledge mapping skills. Methodology: Researcher used both qualitative and quantitative approach to conduct this study. Primary data about current level of participants’ understanding of knowledge mapping was collected through structured questionnaire. Later on, Participant Observation method was used where researchers acted as part of the group to gathered data from the trainees during their performance in training activities and tasks. Findings: Respondents of the study were examined for skills and abilities to organizing ideas, helping groups to develop conceptual framework, identifying critical knowledge areas of an organization, study large networks and identifying the knowledge flow using nodes and vertices, visualizing information, represent organizational structure etc. Overall, the responses varied in different skills depending on the performance and presentations. However, generally all participants have demonstrated average level of using both the IT and Non-IT K-mapping tools and techniques during simulation exercises, analysis paper de-briefing, case study reports, post visit presentation, course review, current issue presentation, syndicate meetings, and daily synopsis. Research Limitations: This study is conducted on a small-scale population of 67 public sector employees nominated by federal government to undergo 14 weeks extensive training program called MCMC (Mid-Career Management Course) at National Institute of Management, Peshawar, Pakistan. Results, however, reflects only a specific class of public sector employees i.e. working in grade 18 and having more than 5 years of work. Practical Implications: Research findings are useful for trainers, training agencies, government functionaries, and organizations working for capacity building of public sector employees.

Keywords: knowledge management, km in public sector, knowledge management and professional development, knowledge management in training, knowledge mapping

Procedia PDF Downloads 236
314 Riverine Urban Heritage: A Basis for Green Infrastructure

Authors: Ioanna H. Lioliou, Despoina D. Zavraka

Abstract:

The radical reformation that Greek urban space, has undergone over the last century, due to the socio-historical developments, technological development and political–geographic factors, has left its imprint on the urban landscape. While the big cities struggle to regain urban landscape balance, small towns are considered to offer high quality lifescapes, ensuring sustainable development potential. However, their unplanned urbanization process led to the loss of significant areas of nature, lack of essential infrastructure, chaotic built environment, incompatible land uses and urban cohesiveness. Natural environment reference points, such as springs, streams, rivers, forests, suburban greenbelts, and etc.; seems to be detached from urban space, while the public, open and green spaces, unequally distributed in the built environment, they are no longer able to offer a complete experience of nature in the city. This study focuses on Greek mainland, a small town Elassona, and aims to restore spatial coherence between the city’s homonymous river and its urban space surroundings. The existence of a linear aquatic ecosystem, is considered a precious greenway, also referred as blueway, able to initiate natural penetrations and ecosystems empowering. The integration of disconnected natural ecosystems forms the basis of a strategic intervention scheme, where the river becomes the urban integration tool / feature, constituting the main urban corridor and an indispensible part of a wider green network that connects open and green spaces, ensuring the function of all the established networks (transportation, commercial, social) of the town. The proposed intervention, introduces a green network highlighting the old stone bridge at the ‘entrance’ of the river in the town and expanding throughout the town with strategic uses and activities, providing accessibility for all the users. The methodology used, is based on the collection of design tools used in related urban river-design interventions around the world. The reinstallation/reactivation of the balance between natural and urban landscape, besides the environmental benefits, contributes decisively to the illustration/projection of urban green identity and re-enhancement of the quality of lifescape qualities and social interaction.

Keywords: green network, rehabilitation scheme, urban landscape, urban streams

Procedia PDF Downloads 262
313 Seismotectonics and Seismology the North of Algeria

Authors: Djeddi Mabrouk

Abstract:

The slow coming together between the Afro-Eurasia plates seems to be the main cause of the active deformation in the whole of North Africa which in consequence come true in Algeria with a large zone of deformation in an enough large limited band, southern through Saharan atlas and northern through tell atlas. Maghrebin and Atlassian Chain along North Africa are the consequence of this convergence. In junction zone, we have noticed a compressive regime NW-SE with a creases-faults structure and structured overthrust. From a geological point of view the north part of Algeria is younger then Saharan platform, it’s changing so unstable and constantly in movement, it’s characterized by creases openly reversed, overthrusts and reversed faults, and undergo perpetually complex movement vertically and horizontally. On structural level the north of Algeria it's a part of erogenous alpine peri-Mediterranean and essentially the tertiary age It’s spread from east to the west of Algeria over 1200 km.This oogenesis is extended from east to west on broadband of 100 km.The alpine chain is shaped by 3 domains: tell atlas in north, high plateaus in mid and Saharan atlas in the south In extreme south we find the Saharan platform which is made of Precambrian bedrock recovered by Paleozoic practically not deformed. The Algerian north and the Saharan platform are separated by an important accident along of 2000km from Agadir (Morocco) to Gabes (Tunisian). The seismic activity is localized essentially in a coastal band in the north of Algeria shaped by tell atlas, high plateaus, Saharan atlas. Earthquakes are limited in the first 20km of the earth's crust; they are caused by movements along faults of inverted orientation NE-SW or sliding tectonic plates. The center region characterizes Strong Earthquake Activity who locates mainly in the basin of Mitidja (age Neogene).The southern periphery (Atlas Blidéen) constitutes the June, more Important seism genic sources in the city of Algiers and east (Boumerdes region). The North East Region is also part of the tellian area, but it is characterized by a different strain in other parts of northern Algeria. The deformation is slow and low to moderate seismic activity. Seismic activity is related to the tectonic-slip earthquake. The most pronounced is that of 27 October 1985 (Constantine) of seismic moment magnitude Mw = 5.9. North-West region is quite active and also artificial seismic hypocenters which do not exceed 20km. The deep seismicity is concentrated mainly a narrow strip along the edge of Quaternary and Neogene basins Intra Mountains along the coast. The most violent earthquakes in this region are the earthquake of Oran in 1790 and earthquakes Orléansville (El Asnam in 1954 and 1980).

Keywords: alpine chain, seismicity north Algeria, earthquakes in Algeria, geophysics, Earth

Procedia PDF Downloads 384
312 Assessment of Impact of Urbanization in Drainage Urban Systems, Cali-Colombia

Authors: A. Caicedo Padilla, J. Zambrano Nájera

Abstract:

Cali, the capital of Valle del Cauca and the second city of Colombia, is located in the Cauca River Valley between the Western and Central Cordillera that is South West of the country. The topography of the city is mainly flat, but it is possibly to find mountains in the west. The city has increased urbanization during XX century, especially since 1958 when started a rapid growth due to migration of people from other parts of the region. Much of that population has settled in eastern of Cali, an area originally intended for cane cultivation and a zone of flood from Cauca River and its tributaries. Due to the unplanned migration, settling was inadequate and produced changes in natural dynamics of the basins, which has resulted in increases in runoff volumes, peak flows and flow velocities, that in turn increases flood risk. Sewerage networks capacity were not enough for this higher runoff volume, because in first term they were not adequately designed and built, causing its failure. This in turn generates increasingly recurrent floods generating considerable effects on the economy and development of normal activities in Cali. Thus, it becomes very important to know hydrological behavior of Urban Watersheds. This research aims to determine the impact of urbanization on hydrology of watersheds with very low slopes. The project aims to identify changes in natural drainage patterns caused by the changes made on landscape. From the identification of such modifications it will be defined the most critical areas due to recurring flood events in the city of Cali. Critical areas are defined as areas where the sewerage system does not work properly as surface runoff increases considerable with storm events, and floods are recurrent. The assessment will be done from the analysis of Geographic Information Systems (GIS) theme layers from CVC Environmental Institution of Regional Control in Valle del Cauca, hydrological data and disaster database developed by OSSO Corporation. Rainfall data from a network and historical stream flow data will be used for analysis of historical behavior and change of precipitation and hydrological response according to homogeneous zones characterized by EMCALI S.A. public utility enterprise of Cali in 1999.

Keywords: drainage systems, land cover changes, urban hydrology, urban planning

Procedia PDF Downloads 242
311 A Web and Cloud-Based Measurement System Analysis Tool for the Automotive Industry

Authors: C. A. Barros, Ana P. Barroso

Abstract:

Any industrial company needs to determine the amount of variation that exists within its measurement process and guarantee the reliability of their data, studying the performance of their measurement system, in terms of linearity, bias, repeatability and reproducibility and stability. This issue is critical for automotive industry suppliers, who are required to be certified by the 16949:2016 standard (replaces the ISO/TS 16949) of International Automotive Task Force, defining the requirements of a quality management system for companies in the automotive industry. Measurement System Analysis (MSA) is one of the mandatory tools. Frequently, the measurement system in companies is not connected to the equipment and do not incorporate the methods proposed by the Automotive Industry Action Group (AIAG). To address these constraints, an R&D project is in progress, whose objective is to develop a web and cloud-based MSA tool. This MSA tool incorporates Industry 4.0 concepts, such as, Internet of Things (IoT) protocols to assure the connection with the measuring equipment, cloud computing, artificial intelligence, statistical tools, and advanced mathematical algorithms. This paper presents the preliminary findings of the project. The web and cloud-based MSA tool is innovative because it implements all statistical tests proposed in the MSA-4 reference manual from AIAG as well as other emerging methods and techniques. As it is integrated with the measuring devices, it reduces the manual input of data and therefore the errors. The tool ensures traceability of all performed tests and can be used in quality laboratories and in the production lines. Besides, it monitors MSAs over time, allowing both the analysis of deviations from the variation of the measurements performed and the management of measurement equipment and calibrations. To develop the MSA tool a ten-step approach was implemented. Firstly, it was performed a benchmarking analysis of the current competitors and commercial solutions linked to MSA, concerning Industry 4.0 paradigm. Next, an analysis of the size of the target market for the MSA tool was done. Afterwards, data flow and traceability requirements were analysed in order to implement an IoT data network that interconnects with the equipment, preferably via wireless. The MSA web solution was designed under UI/UX principles and an API in python language was developed to perform the algorithms and the statistical analysis. Continuous validation of the tool by companies is being performed to assure real time management of the ‘big data’. The main results of this R&D project are: MSA Tool, web and cloud-based; Python API; New Algorithms to the market; and Style Guide of UI/UX of the tool. The MSA tool proposed adds value to the state of the art as it ensures an effective response to the new challenges of measurement systems, which are increasingly critical in production processes. Although the automotive industry has triggered the development of this innovative MSA tool, other industries would also benefit from it. Currently, companies from molds and plastics, chemical and food industry are already validating it.

Keywords: automotive Industry, industry 4.0, Internet of Things, IATF 16949:2016, measurement system analysis

Procedia PDF Downloads 200
310 Machine Learning and Internet of Thing for Smart-Hydrology of the Mantaro River Basin

Authors: Julio Jesus Salazar, Julio Jesus De Lama

Abstract:

the fundamental objective of hydrological studies applied to the engineering field is to determine the statistically consistent volumes or water flows that, in each case, allow us to size or design a series of elements or structures to effectively manage and develop a river basin. To determine these values, there are several ways of working within the framework of traditional hydrology: (1) Study each of the factors that influence the hydrological cycle, (2) Study the historical behavior of the hydrology of the area, (3) Study the historical behavior of hydrologically similar zones, and (4) Other studies (rain simulators or experimental basins). Of course, this range of studies in a certain basin is very varied and complex and presents the difficulty of collecting the data in real time. In this complex space, the study of variables can only be overcome by collecting and transmitting data to decision centers through the Internet of things and artificial intelligence. Thus, this research work implemented the learning project of the sub-basin of the Shullcas river in the Andean basin of the Mantaro river in Peru. The sensor firmware to collect and communicate hydrological parameter data was programmed and tested in similar basins of the European Union. The Machine Learning applications was programmed to choose the algorithms that direct the best solution to the determination of the rainfall-runoff relationship captured in the different polygons of the sub-basin. Tests were carried out in the mountains of Europe, and in the sub-basins of the Shullcas river (Huancayo) and the Yauli river (Jauja) with heights close to 5000 m.a.s.l., giving the following conclusions: to guarantee a correct communication, the distance between devices should not pass the 15 km. It is advisable to minimize the energy consumption of the devices and avoid collisions between packages, the distances oscillate between 5 and 10 km, in this way the transmission power can be reduced and a higher bitrate can be used. In case the communication elements of the devices of the network (internet of things) installed in the basin do not have good visibility between them, the distance should be reduced to the range of 1-3 km. The energy efficiency of the Atmel microcontrollers present in Arduino is not adequate to meet the requirements of system autonomy. To increase the autonomy of the system, it is recommended to use low consumption systems, such as the Ashton Raggatt McDougall or ARM Cortex L (Ultra Low Power) microcontrollers or even the Cortex M; and high-performance direct current (DC) to direct current (DC) converters. The Machine Learning System has initiated the learning of the Shullcas system to generate the best hydrology of the sub-basin. This will improve as machine learning and the data entered in the big data coincide every second. This will provide services to each of the applications of the complex system to return the best data of determined flows.

Keywords: hydrology, internet of things, machine learning, river basin

Procedia PDF Downloads 143
309 The Potential Role of Some Nutrients and Drugs in Providing Protection from Neurotoxicity Induced by Aluminium in Rats

Authors: Azza A. Ali, Abeer I. Abd El-Fattah, Shaimaa S. Hussein, Hanan A. Abd El-Samea, Karema Abu-Elfotuh

Abstract:

Background: Aluminium (Al) represents an environmental risk factor. Exposure to high levels of Al causes neurotoxic effects and different diseases. Vinpocetine is widely used to improve cognitive functions, it possesses memory-protective and memory-enhancing properties and has the ability to increase cerebral blood flow and glucose uptake. Cocoa bean represents a rich source of iron as well as a potent antioxidant. It can protect from the impact of free radicals, reduces stress as well as depression and promotes better memory and concentration. Wheatgrass is primarily used as a concentrated source of nutrients. It contains vitamins, minerals, carbohydrates, amino acids and possesses antioxidant and anti-inflammatory activities. Coenzyme Q10 (CoQ10) is an intracellular antioxidant and mitochondrial membrane stabilizer. It is effective in improving cognitive disorders and has been used as anti-aging. Zinc is a structural element of many proteins and signaling messenger that is released by neural activity at many central excitatory synapses. Objective: To study the role of some nutrients and drugs as Vinpocetine, Cocoa, Wheatgrass, CoQ10 and Zinc against neurotoxicity induced by Al in rats as well as to compare between their potency in providing protection. Methods: Seven groups of rats were used and received daily for three weeks AlCl3 (70 mg/kg, IP) for Al-toxicity model groups except for the control group which received saline. All groups of Al-toxicity model except one group (non-treated) were co-administered orally together with AlCl3 the following treatments; Vinpocetine (20mg/kg), Cocoa powder (24mg/kg), Wheat grass (100mg/kg), CoQ10 (200mg/kg) or Zinc (32mg/kg). Biochemical changes in the rat brain as acetyl cholinesterase (ACHE), Aβ, brain derived neurotrophic factor (BDNF), inflammatory mediators (TNF-α, IL-1β), oxidative parameters (MDA, SOD, TAC) were estimated for all groups besides histopathological examinations in different brain regions. Results: Neurotoxicity and neurodegenerations in the rat brain after three weeks of Al exposure were indicated by the significant increase in Aβ, ACHE, MDA, TNF-α, IL-1β, DNA fragmentation together with the significant decrease in SOD, TAC, BDNF and confirmed by the histopathological changes in the brain. On the other hand, co-administration of each of Vinpocetine, Cocoa, Wheatgrass, CoQ10 or Zinc together with AlCl3 provided protection against hazards of neurotoxicity and neurodegenerations induced by Al, their protection were indicated by the decrease in Aβ, ACHE, MDA, TNF-α, IL-1β, DNA fragmentation together with the increase in SOD, TAC, BDNF and confirmed by the histopathological examinations of different brain regions. Vinpocetine and Cocoa showed the most pronounced protection while Zinc provided the least protective effects than the other used nutrients and drugs. Conclusion: Different degrees of protection from neurotoxicity and neuronal degenerations induced by Al could be achieved through the co-administration of some nutrients and drugs during its exposure. Vinpocetine and Cocoa provided the most protection than Wheat grass, CoQ10 or Zinc which showed the least protective effects.

Keywords: aluminum, neurotoxicity, vinpocetine, cocoa, wheat grass, coenzyme Q10, Zinc, rats

Procedia PDF Downloads 229
308 A Framework of Virtualized Software Controller for Smart Manufacturing

Authors: Pin Xiu Chen, Shang Liang Chen

Abstract:

A virtualized software controller is developed in this research to replace traditional hardware control units. This virtualized software controller transfers motion interpolation calculations from the motion control units of end devices to edge computing platforms, thereby reducing the end devices' computational load and hardware requirements and making maintenance and updates easier. The study also applies the concept of microservices, dividing the control system into several small functional modules and then deploy into a cloud data server. This reduces the interdependency among modules and enhances the overall system's flexibility and scalability. Finally, with containerization technology, the system can be deployed and started in a matter of seconds, which is more efficient than traditional virtual machine deployment methods. Furthermore, this virtualized software controller communicates with end control devices via wireless networks, making the placement of production equipment or the redesign of processes more flexible and no longer limited by physical wiring. To handle the large data flow and maintain low-latency transmission, this study integrates 5G technology, fully utilizing its high speed, wide bandwidth, and low latency features to achieve rapid and stable remote machine control. An experimental setup is designed to verify the feasibility and test the performance of this framework. This study designs a smart manufacturing site with a 5G communication architecture, serving as a field for experimental data collection and performance testing. The smart manufacturing site includes one robotic arm, three Computer Numerical Control machine tools, several Input/Output ports, and an edge computing architecture. All machinery information is uploaded to edge computing servers and cloud servers via 5G communication and the Internet of Things framework. After analysis and computation, this information is converted into motion control commands, which are transmitted back to the relevant machinery for motion control through 5G communication. The communication time intervals at each stage are calculated using the C++ chrono library to measure the time difference for each command transmission. The relevant test results will be organized and displayed in the full-text.

Keywords: 5G, MEC, microservices, virtualized software controller, smart manufacturing

Procedia PDF Downloads 56
307 Preparation of IPNs and Effect of Swift Heavy Ions Irradiation on their Physico-Chemical Properties

Authors: B. S Kaith, K. Sharma, V. Kumar, S. Kalia

Abstract:

Superabsorbent are three-dimensional networks of linear or branched polymeric chains which can uptake large volume of biological fluids. The ability is due to the presence of functional groups like –NH2, -COOH and –OH. Such cross-linked products based on natural materials, such as cellulose, starch, dextran, gum and chitosan, because of their easy availability, low production cost, non-toxicity and biodegradability have attracted the attention of Scientists and Technologists all over the world. Since natural polymers have better biocompatibility and are non-toxic than most synthetic one, therefore, such materials can be applied in the preparation of controlled drug delivery devices, biosensors, tissue engineering, contact lenses, soil conditioning, removal of heavy metal ions and dyes. Gums are natural potential antioxidants and are used as food additives. They have excellent properties like high solubility, pH stability, non-toxicity and gelling characteristics. Till date lot of methods have been applied for the synthesis and modifications of cross-linked materials with improved properties suitable for different applications. It is well known that ion beam irradiation can play a crucial role to synthesize, modify, crosslink or degrade polymeric materials. High energetic heavy ions irradiation on polymer film induces significant changes like chain scission, cross-linking, structural changes, amorphization and degradation in bulk. Various researchers reported the effects of low and heavy ion irradiation on the properties of polymeric materials and observed significant improvement in optical, electrical, chemical, thermal and dielectric properties. Moreover, modifications induced in the materials mainly depend on the structure, the ion beam parameters like energy, linear energy transfer, fluence, mass, charge and the nature of the target material. Ion-beam irradiation is a useful technique for improving the surface properties of biodegradable polymers without missing the bulk properties. Therefore, a considerable interest has been grown to study the effects of SHIs irradiation on the properties of synthesized semi-IPNs and IPNs. The present work deals with the preparation of semi-IPNs and IPNs and impact of SHI like O7+ and Ni9+ irradiation on optical, chemical, structural, morphological and thermal properties along with impact on different applications. The results have been discussed on the basis of Linear Energy Transfer (LET) of the ions.

Keywords: adsorbent, gel, IPNs, semi-IPNs

Procedia PDF Downloads 356
306 Efficient Estimation of Maximum Theoretical Productivity from Batch Cultures via Dynamic Optimization of Flux Balance Models

Authors: Peter C. St. John, Michael F. Crowley, Yannick J. Bomble

Abstract:

Production of chemicals from engineered organisms in a batch culture typically involves a trade-off between productivity, yield, and titer. However, strategies for strain design typically involve designing mutations to achieve the highest yield possible while maintaining growth viability. Such approaches tend to follow the principle of designing static networks with minimum metabolic functionality to achieve desired yields. While these methods are computationally tractable, optimum productivity is likely achieved by a dynamic strategy, in which intracellular fluxes change their distribution over time. One can use multi-stage fermentations to increase either productivity or yield. Such strategies would range from simple manipulations (aerobic growth phase, anaerobic production phase), to more complex genetic toggle switches. Additionally, some computational methods can also be developed to aid in optimizing two-stage fermentation systems. One can assume an initial control strategy (i.e., a single reaction target) in maximizing productivity - but it is unclear how close this productivity would come to a global optimum. The calculation of maximum theoretical yield in metabolic engineering can help guide strain and pathway selection for static strain design efforts. Here, we present a method for the calculation of a maximum theoretical productivity of a batch culture system. This method follows the traditional assumptions of dynamic flux balance analysis: that internal metabolite fluxes are governed by a pseudo-steady state and external metabolite fluxes are represented by dynamic system including Michealis-Menten or hill-type regulation. The productivity optimization is achieved via dynamic programming, and accounts explicitly for an arbitrary number of fermentation stages and flux variable changes. We have applied our method to succinate production in two common microbial hosts: E. coli and A. succinogenes. The method can be further extended to calculate the complete productivity versus yield Pareto surface. Our results demonstrate that nearly optimal yields and productivities can indeed be achieved with only two discrete flux stages.

Keywords: A. succinogenes, E. coli, metabolic engineering, metabolite fluxes, multi-stage fermentations, succinate

Procedia PDF Downloads 196
305 Geographical Location and the Global Airline Industry: A Delphi Study into the Future of Home Base Requirements

Authors: Darren J. Ellis

Abstract:

This paper investigates the key industry-level consequences and future prospects for the global airline industry of the requirement for airlines to have a home base. This industry context results in geographical location playing a central role in determining how and where international airlines can operate, and the extent to which their international networks can develop. Data from a five stage mixed-methods Delphi study into the global airline industry’s likely future trajectory conducted in 2013 and 2014 are utilized to better understand the likelihood and consequences of home base requirements changing in future. Expert views and forecasts were collected to gauge core industry trends over a ten year timeframe. Attempts to change or bypass this industry requirement have not been successful to date outside of the European single air market. Europe remains the only prominent exception to the general rule in this regard. Most of the industry is founded on air space sovereignty, the nationality rule, and the bilateral system of traffic rights. Europe’s exceptionalism has seen it evolve into a single air market with characteristics similar to a nation-state, rather than to become a force for wider industry change and regional multilateralism. Europe has indeed become a key actor in global aviation, but Europe seems to now be part of the industry’s status quo, not a vehicle for substantially wider multilateralism around the world. The findings from this research indicate that the bilateral system is not viewed by most study experts as disappearing or substantially weakening in the foreseeable future. However, regional multilateralism was also viewed as progressively taking hold in the industry in future, demonstrating that for most industry experts the two are not seen as mutually exclusive but rather as being able to co-exist with each other. This reality ensures that geographical location will continue to play an important role in the global airline industry in future and that, home base requirements will not disappear any time soon either. Even moves in some aviation jurisdictions to dilute nationality requirements for airlines, and instead replace ownership and control restrictions with principal place of business tests, do not ultimately free airlines from their home base. Likewise, an expansion of what constitutes home base to include a regional grouping of countries – again, a currently uncommon reality in global aviation – does not fundamentally weaken the continued relevance of geographical location to the global industry’s future growth and development realities and prospects.

Keywords: airline industry, air space sovereignty, geographical location, home base

Procedia PDF Downloads 117
304 The Quantum Theory of Music and Languages

Authors: Mballa Abanda Serge, Henda Gnakate Biba, Romaric Guemno Kuate, Akono Rufine Nicole, Petfiang Sidonie, Bella Sidonie

Abstract:

The main hypotheses proposed around the definition of the syllable and of music, of the common origin of music and language, should lead the reader to reflect on the cross-cutting questions raised by the debate on the notion of universals in linguistics and musicology. These are objects of controversy, and there lies its interest: the debate raises questions that are at the heart of theories on language. It is an inventive, original and innovative research thesis. The main hypotheses proposed around the definition of the syllable and of music, of the common origin of music and language, should lead the reader to reflect on the cross-cutting questions raised by the debate on the notion of universals in linguistics and musicology. These are objects of controversy, and there lies its interest: the debate raises questions that are at the heart of theories on language. It is an inventive, original and innovative research thesis. A contribution to the theoretical, musicological, ethno musicological and linguistic conceptualization of languages, giving rise to the practice of interlocution between the social and cognitive sciences, the activities of artistic creation and the question of modeling in the human sciences: mathematics, computer science, translation automation and artificial intelligence. When you apply this theory to any text of a folksong of a world-tone language, you do not only piece together the exact melody, rhythm, and harmonies of that song as if you knew it in advance but also the exact speaking of this language. The author believes that the issue of the disappearance of tonal languages and their preservation has been structurally resolved, as well as one of the greatest cultural equations related to the composition and creation of tonal, polytonal and random music. The experimentation confirming the theorization, It designed a semi-digital, semi-analog application which translates the tonal languages of Africa (about 2,100 languages) into blues, jazz, world music, polyphonic music, tonal and anatonal music and deterministic and random music). To test this application, I use a music reading and writing software that allows me to collect the data extracted from my mother tongue, which is already modeled in the musical staves saved in the ethnographic (semiotic) dictionary for automatic translation ( volume 2 of the book). Translation is done (from writing to writing, from writing to speech and from writing to music). Mode of operation: you type a text on your computer, a structured song (chorus-verse), and you command the machine a melody of blues, jazz and world music or variety etc. The software runs, giving you the option to choose harmonies, and then you select your melody.

Keywords: music, entanglement, langauge, science

Procedia PDF Downloads 62
303 Methodological Deficiencies in Knowledge Representation Conceptual Theories of Artificial Intelligence

Authors: Nasser Salah Eldin Mohammed Salih Shebka

Abstract:

Current problematic issues in AI fields are mainly due to those of knowledge representation conceptual theories, which in turn reflected on the entire scope of cognitive sciences. Knowledge representation methods and tools are driven from theoretical concepts regarding human scientific perception of the conception, nature, and process of knowledge acquisition, knowledge engineering and knowledge generation. And although, these theoretical conceptions were themselves driven from the study of the human knowledge representation process and related theories; some essential factors were overlooked or underestimated, thus causing critical methodological deficiencies in the conceptual theories of human knowledge and knowledge representation conceptions. The evaluation criteria of human cumulative knowledge from the perspectives of nature and theoretical aspects of knowledge representation conceptions are affected greatly by the very materialistic nature of cognitive sciences. This nature caused what we define as methodological deficiencies in the nature of theoretical aspects of knowledge representation concepts in AI. These methodological deficiencies are not confined to applications of knowledge representation theories throughout AI fields, but also exceeds to cover the scientific nature of cognitive sciences. The methodological deficiencies we investigated in our work are: - The Segregation between cognitive abilities in knowledge driven models.- Insufficiency of the two-value logic used to represent knowledge particularly on machine language level in relation to the problematic issues of semantics and meaning theories. - Deficient consideration of the parameters of (existence) and (time) in the structure of knowledge. The latter requires that we present a more detailed introduction of the manner in which the meanings of Existence and Time are to be considered in the structure of knowledge. This doesn’t imply that it’s easy to apply in structures of knowledge representation systems, but outlining a deficiency caused by the absence of such essential parameters, can be considered as an attempt to redefine knowledge representation conceptual approaches, or if proven impossible; constructs a perspective on the possibility of simulating human cognition on machines. Furthermore, a redirection of the aforementioned expressions is required in order to formulate the exact meaning under discussion. This redirection of meaning alters the role of Existence and time factors to the Frame Work Environment of knowledge structure; and therefore; knowledge representation conceptual theories. Findings of our work indicate the necessity to differentiate between two comparative concepts when addressing the relation between existence and time parameters, and between that of the structure of human knowledge. The topics presented throughout the paper can also be viewed as an evaluation criterion to determine AI’s capability to achieve its ultimate objectives. Ultimately, we argue some of the implications of our findings that suggests that; although scientific progress may have not reached its peak, or that human scientific evolution has reached a point where it’s not possible to discover evolutionary facts about the human Brain and detailed descriptions of how it represents knowledge, but it simply implies that; unless these methodological deficiencies are properly addressed; the future of AI’s qualitative progress remains questionable.

Keywords: cognitive sciences, knowledge representation, ontological reasoning, temporal logic

Procedia PDF Downloads 96
302 Development of an Interface between BIM-model and an AI-based Control System for Building Facades with Integrated PV Technology

Authors: Moser Stephan, Lukasser Gerald, Weitlaner Robert

Abstract:

Urban structures will be used more intensively in the future through redensification or new planned districts with high building densities. Especially, to achieve positive energy balances like requested for Positive Energy Districts (PED) the single use of roofs is not sufficient for dense urban areas. However, the increasing share of window significantly reduces the facade area available for use in PV generation. Through the use of PV technology at other building components, such as external venetian blinds, onsite generation can be maximized and standard functionalities of this product can be positively extended. While offering advantages in terms of infrastructure, sustainability in the use of resources and efficiency, these systems require an increased optimization in planning and control strategies of buildings. External venetian blinds with PV technology require an intelligent control concept to meet the required demands such as maximum power generation, glare prevention, high daylight autonomy, avoidance of summer overheating but also use of passive solar gains in wintertime. Today, geometric representation of outdoor spaces and at the building level, three-dimensional geometric information is available for planning with Building Information Modeling (BIM). In a research project, a web application which is called HELLA DECART was developed to provide this data structure to extract the data required for the simulation from the BIM models and to make it usable for the calculations and coupled simulations. The investigated object is uploaded as an IFC file to this web application and includes the object as well as the neighboring buildings and possible remote shading. This tool uses a ray tracing method to determine possible glare from solar reflections of a neighboring building as well as near and far shadows per window on the object. Subsequently, an annual estimate of the sunlight per window is calculated by taking weather data into account. This optimized daylight assessment per window provides the ability to calculate an estimation of the potential power generation at the integrated PV on the venetian blind but also for the daylight and solar entry. As a next step, these results of the calculations as well as all necessary parameters for the thermal simulation can be provided. The overall aim of this workflow is to advance the coordination between the BIM model and coupled building simulation with the resulting shading and daylighting system with the artificial lighting system and maximum power generation in a control system. In the research project Powershade, an AI based control concept for PV integrated façade elements with coupled simulation results is investigated. The developed automated workflow concept in this paper is tested by using an office living lab at the HELLA company.

Keywords: BIPV, building simulation, optimized control strategy, planning tool

Procedia PDF Downloads 91
301 Issues of Accounting of Lease and Revenue according to International Financial Reporting Standards

Authors: Nadezhda Kvatashidze, Elena Kharabadze

Abstract:

It is broadly known that lease is a flexible means of funding enterprises. Lease reduces the risk related to access and possession of assets, as well as obtainment of funding. Therefore, it is important to refine lease accounting. The lease accounting regulations under the applicable standard (International Accounting Standards 17) make concealment of liabilities possible. As a result, the information users get inaccurate and incomprehensive information and have to resort to an additional assessment of the off-balance sheet lease liabilities. In order to address the problem, the International Financial Reporting Standards Board decided to change the approach to lease accounting. With the deficiencies of the applicable standard taken into account, the new standard (IFRS 16 ‘Leases’) aims at supplying appropriate and fair lease-related information to the users. Save certain exclusions; the lessee is obliged to recognize all the lease agreements in its financial report. The approach was determined by the fact that under the lease agreement, rights and obligations arise by way of assets and liabilities. Immediately upon conclusion of the lease agreement, the lessee takes an asset into its disposal and assumes the obligation to effect the lease-related payments in order to meet the recognition criteria defined by the Conceptual Framework for Financial Reporting. The payments are to be entered into the financial report. The new lease accounting standard secures supply of quality and comparable information to the financial information users. The International Accounting Standards Board and the US Financial Accounting Standards Board jointly developed IFRS 15: ‘Revenue from Contracts with Customers’. The standard allows the establishment of detailed revenue recognition practical criteria such as identification of the performance obligations in the contract, determination of the transaction price and its components, especially price variable considerations and other important components, as well as passage of control over the asset to the customer. IFRS 15: ‘Revenue from Contracts with Customers’ is very similar to the relevant US standards and includes requirements more specific and consistent than those of the standards in place. The new standard is going to change the recognition terms and techniques in the industries, such as construction, telecommunications (mobile and cable networks), licensing (media, science, franchising), real property, software etc.

Keywords: assessment of the lease assets and liabilities, contractual liability, division of contract, identification of contracts, contract price, lease identification, lease liabilities, off-balance sheet, transaction value

Procedia PDF Downloads 300
300 Prevention of Preterm Birth and Management of Uterine Contractions with Traditional Korean Medicine: Integrative Approach

Authors: Eun-Seop Kim, Eun-Ha Jang, Rana R. Kim, Sae-Byul Jang

Abstract:

Objective: Preterm labor is the most common antecedent of preterm birth(PTB), which is characterized by regular uterine contraction before 37 weeks of pregnancy and cervical change. In acute preterm labor, tocolytics are administered as the first-line medication to suppress uterine contractions but rarely delay pregnancy to 37 weeks of gestation. On the other hand, according to the Korean Traditional Medicine, PTB is caused by the deficiency of Qi and unnecessary energy in the body of the mother. The aim of this study was to demonstrate the benefit of Traditional Korean Medicine as an adjuvant therapy in management of early uterine contractions and the prevention of PTB. Methods: It is a case report of a 38-year-old woman (0-0-6-0) hospitalized for irregular uterine contractions and cervical change at 33+3/7 weeks of gestation. Past history includes chemical pregnancies achieved by Artificial Rroductive Technology(ART), one stillbirth (at 7 weeks) and a laparoscopic surgery for endometriosis. After seven trials of IVF and articificial insemination, she had succeeded in conception via in-vitro fertilization (IVF) with help of Traditional Korean Medicine (TKM) treatments. Due to irregular uterine contractions and cervical changes, 2 TKM were prescribed: Gami-Dangguisan, and Antae-eum, known to nourish blood and clear away heat. 120ml of Gami-Dangguisan was given twice a day monring and evening along with same amount of Antae-eum once a day from 31 August 2013 to 28 November 2013. Tocolytics (Ritodrine) was administered as a first aid for maintenance of pregnancy. Information regarding progress until the delivery was collected during the patient’s visit. Results: On admission, the cervix of 15mm in length and cervical os with 0.5cm-dilated were observed via ultrasonography. 50% cervical effacement was also detected in physical examination. Tocolysis had been temporarily maintained. As a supportive therapy, TKM herbal preparations(gami-dangguisan and Antae-eum) were concomitantly given. As of 34+2/7 weeks of gestation, however intermittent uterine contractions appeared (5-12min) on cardiotocography and vaginal bleeding was also smeared at 34+3/7 weeks. However, enhanced tocolytics and continuous administration of herbal medicine sustained the pregnancy to term. At 37+2/7 weeks, no sign of labor with restored cervical length was confirmed. The woman gave a term birth to a healthy infant via vaginal delivery at 39+3/7 gestational weeks. Conclusions: This is the first successful case report about a preter labor patient administered with conventional tocolytic agents as well as TKM herbal decoctions, delaying delivery to term. This case deserves attention considering it is rare to maintain gestation to term only with tocolytic intervention. Our report implies the potential of herbal medicine as an adjuvant therapy for preterm labor treatment. Further studies are needed to assess the safety and efficacy of TKM herbal medicine as a therapeutic alternative for curing preterm birth.

Keywords: preterm labor, traditional Korean medicine, herbal medicine, integrative treatment, complementary and alternative medicine

Procedia PDF Downloads 352
299 Is Brain Death Reversal Possible in Near Future: Intrathecal Sodium Nitroprusside (SNP) Superfusion in Brain Death Patients=The 10,000 Fold Effect

Authors: Vinod Kumar Tewari, Mazhar Husain, Hari Kishan Das Gupta

Abstract:

Background: Primary or secondary brain death is also accompanied with vasospasm of the perforators other than tissue disruption & further exaggerates the anoxic damage, in the form of neuropraxia. In normal conditions the excitatory impulse propagates as anterograde neurotransmission (ANT) and at the level of synapse, glutamate activates NMDA receptors on postsynaptic membrane. Nitric oxide (NO) is produced by Nitric oxide Synthetase (NOS) in postsynaptic dendride or cell body and travels backwards across a chemical synapse to bind to the axon terminal of a presynaptic neuron for regulation of ANT this process is called as the retrograde neurotransmission (RNT). Thus the primary function of NO is RNT and the purpose of RNT is regulation of chemical neurotransmission at synapse. For this reason, RNT allows neural circuits to create feedback loops. The haem is the ligand binding site of NO receptor (sGC) at presynaptic membrane. The affinity of haem exhibits > 10,000-fold excess for NO than Oxygen (THE 10,000 FOLD EFFECT). In pathological conditions ANT, normal synaptic activity including RNT is absent. NO donors like sodium nitroprusside (SNP) releases NO by activating NOS at the level of postsynaptic area. NO now travels backwards across a chemical synapse to bind to the haem of NO receptor at axon terminal of a presynaptic neuron as in normal condition. NO now acts as impulse generator (at presynaptic membrane) thus bypasses the normal ANT. Also the arteriolar perforators are having Nitric Oxide Synthetase (NOS) at the adventitial side (outer border) on which sodium nitroprusside (SNP) acts; causing release of Nitric Oxide (NO) which vasodilates the perforators causing gush of blood in brain’s tissue and reversal of brain death. Objective: In brain death cases we only think for various transplantations but this study being a pilot study reverses some criteria of brain death by vasodilating the arteriolar perforators. To study the effect of intrathecal sodium nitroprusside (IT SNP) in cases of brain death in which: 1. Retrograde transmission = assessed by the hyperacute timings of reversal 2. The arteriolar perforator vasodilatation caused by NO and the maintenance of reversal of brain death reversal. Methods: 35 year old male, who became brain death after head injury and has not shown any signs of improvement after every maneuver for 6 hours, a single superfusion done by SNP via transoptic canal route for quadrigeminal cistern and cisternal puncture for IV ventricular with SNP done. Results: He showed spontaneous respiration (7 bouts) with TCD studies showing start of pulsations of various branches of common carotid arteries. Conclusions: In future we can give this SNP via transoptic canal route and in IV ventricle before declaring the body to be utilized for transplantations or dead or in broader way we can say that in near future it is possible to revert back from brain death or we have to modify our criterion.

Keywords: brain death, intrathecal sodium nitroprusside, TCD studies, perforators, vasodilatations, retrograde transmission, 10, 000 fold effect

Procedia PDF Downloads 380
298 Unleashing the Potential of Green Finance in Architecture: A Promising Path for Balkan Countries

Authors: Luan Vardari, Dena Arapi Vardari

Abstract:

The Balkan countries, known for their diverse landscapes and cultural heritage, face the dual challenge of promoting economic growth while addressing pressing environmental concerns. In recent years, the concept of green finance has emerged as a powerful tool to achieve sustainable development and mitigate the environmental impact of various sectors, including architecture. This extended abstract explores the untapped potential of green finance in architecture within the Balkan region and highlights its role in driving sustainable construction practices and fostering a greener future. The abstract begins by defining green finance and emphasizing its relevance in the context of the architectural sector in Balkan countries. It underlines the benefits of green finance, such as economic growth, environmental conservation, and social well-being. Integrating green finance into architectural projects is important as a means to achieve sustainable development goals while promoting financial viability. Also, delves into the current state of green building practices in the Balkan countries and identifies the need for financial support to further drive adoption. It explores the existing regulatory frameworks and policies that promote sustainable architecture and discusses how green finance can complement these initiatives. Unique challenges faced by Balkan countries are highlighted, along with the potential opportunities that green finance presents in overcoming these challenges. We highlight successful sustainable architectural projects in the region to showcase the practical application of green finance in the Balkans. These projects exemplify the effective utilization of green finance mechanisms, resulting in tangible economic and environmental impacts, including job creation, energy efficiency, and reduced carbon emissions. The abstract concludes by identifying replicable models and lessons learned from these projects that can serve as a blueprint for future sustainable architecture initiatives in the Balkans. The importance of collaboration and knowledge sharing among stakeholders is emphasized. Engaging architects, financial institutions, governments, and local communities is crucial to promoting green finance in architecture. The abstract suggests the establishment of knowledge exchange platforms and regional/international networks to foster collaboration and facilitate the sharing of expertise among Balkan countries.

Keywords: sustainable finance, renewable energy, Balkan region, investment opportunities, green infrastructure, ESG criteria, architecture

Procedia PDF Downloads 48
297 The Effect of Soil-Structure Interaction on the Post-Earthquake Fire Performance of Structures

Authors: A. T. Al-Isawi, P. E. F. Collins

Abstract:

The behaviour of structures exposed to fire after an earthquake is not a new area of engineering research, but there remain a number of areas where further work is required. Such areas relate to the way in which seismic excitation is applied to a structure, taking into account the effect of soil-structure interaction (SSI) and the method of analysis, in addition to identifying the excitation load properties. The selection of earthquake data input for use in nonlinear analysis and the method of analysis are still challenging issues. Thus, realistic artificial ground motion input data must be developed to certify that site properties parameters adequately describe the effects of the nonlinear inelastic behaviour of the system and that the characteristics of these parameters are coherent with the characteristics of the target parameters. Conversely, ignoring the significance of some attributes, such as frequency content, soil site properties and earthquake parameters may lead to misleading results, due to the misinterpretation of required input data and the incorrect synthesise of analysis hypothesis. This paper presents a study of the post-earthquake fire (PEF) performance of a multi-storey steel-framed building resting on soft clay, taking into account the effects of the nonlinear inelastic behaviour of the structure and soil, and the soil-structure interaction (SSI). Structures subjected to an earthquake may experience various levels of damage; the geometrical damage, which indicates the change in the initial structure’s geometry due to the residual deformation as a result of plastic behaviour, and the mechanical damage which identifies the degradation of the mechanical properties of the structural elements involved in the plastic range of deformation. Consequently, the structure presumably experiences partial structural damage but is then exposed to fire under its new residual material properties, which may result in building failure caused by a decrease in fire resistance. This scenario would be more complicated if SSI was also considered. Indeed, most earthquake design codes ignore the probability of PEF as well as the effect that SSI has on the behaviour of structures, in order to simplify the analysis procedure. Therefore, the design of structures based on existing codes which neglect the importance of PEF and SSI can create a significant risk of structural failure. In order to examine the criteria for the behaviour of a structure under PEF conditions, a two-dimensional nonlinear elasto-plastic model is developed using ABAQUS software; the effects of SSI are included. Both geometrical and mechanical damages have been taken into account after the earthquake analysis step. For comparison, an identical model is also created, which does not include the effects of soil-structure interaction. It is shown that damage to structural elements is underestimated if SSI is not included in the analysis, and the maximum percentage reduction in fire resistance is detected in the case when SSI is included in the scenario. The results are validated using the literature.

Keywords: Abaqus Software, Finite Element Analysis, post-earthquake fire, seismic analysis, soil-structure interaction

Procedia PDF Downloads 109
296 Staying When Everybody Else Is Leaving: Coping with High Out-Migration in Rural Areas of Serbia

Authors: Anne Allmrodt

Abstract:

Regions of South-East Europe are characterised by high out-migration for decades. The reasons for leaving range from the hope of a better work situation to a better health care system and beyond. In Serbia, this high out-migration hits the rural areas in particular so that the population number is in the red repeatedly. It might not be hard to guess that this negative population growth has the potential to create different challenges for those who stay in rural areas. So how are they coping with the – statistically proven – high out-migration? Having this in mind, the study is investigating the people‘s individual awareness of the social phenomenon high out-migration and their daily life strategies in rural areas. Furthermore, the study seeks to find out the people’s resilient skills in that context. Is the condition of high out-migration conducive for resilience? The methodology combines a quantitative and a qualitative approach (mixed methods). For the quantitative part, a standardised questionnaire has been developed, including a multiple choice section and a choice experiment. The questionnaire was handed out to people living in rural areas of Serbia only (n = 100). The sheet included questions about people’s awareness of high out-migration, their own daily life strategies or challenges and their social network situation (data about the social network was necessary here since it is supposed to be an influencing variable for resilience). Furthermore, test persons were asked to make different choices of coping with high out-migration in a self-designed choice experiment. Additionally, the study included qualitative interviews asking citizens from rural areas of Serbia. The topics asked during the interview focused on their awareness of high out-migration, their daily life strategies, and challenges as well as their social network situation. Results have shown the following major findings. The awareness of high out-migration is not the same with all test persons. Some declare it as something positive for their own life, others as negative or not effecting at all. The way of coping generally depended – maybe not surprising – on the people’s social network. However – and this might be the most important finding - not everybody with a certain number of contacts had better coping strategies and was, therefore, more resilient. Here the results show that especially people with high affiliation and proximity inside their network were able to cope better and shew higher resilience skills. The study took one step forward in terms of knowledge about societal resilience as well as coping strategies of societies in rural areas. It has shown part of the other side of nowadays migration‘s coin and gives a hint for a more sustainable rural development and community empowerment.

Keywords: coping, out-migration, resilience, rural development, social networks, south-east Europe

Procedia PDF Downloads 111
295 Digimesh Wireless Sensor Network-Based Real-Time Monitoring of ECG Signal

Authors: Sahraoui Halima, Dahani Ameur, Tigrine Abedelkader

Abstract:

DigiMesh technology represents a pioneering advancement in wireless networking, offering cost-effective and energy-efficient capabilities. Its inherent simplicity and adaptability facilitate the seamless transfer of data between network nodes, extending the range and ensuring robust connectivity through autonomous self-healing mechanisms. In light of these advantages, this study introduces a medical platform harnessed with DigiMesh wireless network technology characterized by low power consumption, immunity to interference, and user-friendly operation. The primary application of this platform is the real-time, long-distance monitoring of Electrocardiogram (ECG) signals, with the added capacity for simultaneous monitoring of ECG signals from multiple patients. The experimental setup comprises key components such as Raspberry Pi, E-Health Sensor Shield, and Xbee DigiMesh modules. The platform is composed of multiple ECG acquisition devices labeled as Sensor Node 1 and Sensor Node 2, with a Raspberry Pi serving as the central hub (Sink Node). Two communication approaches are proposed: Single-hop and multi-hop. In the Single-hop approach, ECG signals are directly transmitted from a sensor node to the sink node through the XBee3 DigiMesh RF Module, establishing peer-to-peer connections. This approach was tested in the first experiment to assess the feasibility of deploying wireless sensor networks (WSN). In the multi-hop approach, two sensor nodes communicate with the server (Sink Node) in a star configuration. This setup was tested in the second experiment. The primary objective of this research is to evaluate the performance of both Single-hop and multi-hop approaches in diverse scenarios, including open areas and obstructed environments. Experimental results indicate the DigiMesh network's effectiveness in Single-hop mode, with reliable communication over distances of approximately 300 meters in open areas. In the multi-hop configuration, the network demonstrated robust performance across approximately three floors, even in the presence of obstacles, without the need for additional router devices. This study offers valuable insights into the capabilities of DigiMesh wireless technology for real-time ECG monitoring in healthcare applications, demonstrating its potential for use in diverse medical scenarios.

Keywords: DigiMesh protocol, ECG signal, real-time monitoring, medical platform

Procedia PDF Downloads 58
294 Systematic Study of Structure Property Relationship in Highly Crosslinked Elastomers

Authors: Natarajan Ramasamy, Gurulingamurthy Haralur, Ramesh Nivarthu, Nikhil Kumar Singha

Abstract:

Elastomers are polymeric materials with varied backbone architectures ranging from linear to dendrimeric structures and wide varieties of monomeric repeat units. These elastomers show strongly viscous and weakly elastic when it is not cross-linked. But when crosslinked, based on the extent the properties of these elastomers can range from highly flexible to highly stiff nature. Lightly cross-linked systems are well studied and reported. Understanding the nature of highly cross-linked rubber based upon chemical structure and architecture is critical for varieties of applications. One of the critical parameters is cross-link density. In the current work, we have studied the highly cross-linked state of linear, lightly branched to star-shaped branched elastomers and determined the cross-linked density by using different models. Change in hardness, shift in Tg, change in modulus and swelling behavior were measured experimentally as a function of the extent of curing. These properties were analyzed using varied models to determine cross-link density. We used hardness measurements to examine cure time. Hardness to the extent of curing relationship is determined. It is well known that micromechanical transitions like Tg and storage modulus are related to the extent of crosslinking. The Tg of the elastomer in different crosslinked state was determined by DMA, and based on plateau modulus the crosslink density is estimated by using Nielsen’s model. Usually for lightly crosslinked systems, based on equilibrium swelling ratio in solvent the cross link density is estimated by using Flory–Rhener model. When it comes to highly crosslinked system, Flory-Rhener model is not valid because of smaller chain length. So models based on the assumption of polymer as a Non-Gaussian chain like 1) Helmis–Heinrich–Straube (HHS) model, 2) Gloria M.gusler and Yoram Cohen Model, 3) Barbara D. Barr-Howell and Nikolaos A. Peppas model is used for estimating crosslink density. In this work, correction factors are determined to the existing models and based upon it structure-property relationship of highly crosslinked elastomers was studied.

Keywords: dynamic mechanical analysis, glass transition temperature, parts per hundred grams of rubber, crosslink density, number of networks per unit volume of elastomer

Procedia PDF Downloads 150
293 Vehicle Timing Motion Detection Based on Multi-Dimensional Dynamic Detection Network

Authors: Jia Li, Xing Wei, Yuchen Hong, Yang Lu

Abstract:

Detecting vehicle behavior has always been the focus of intelligent transportation, but with the explosive growth of the number of vehicles and the complexity of the road environment, the vehicle behavior videos captured by traditional surveillance have been unable to satisfy the study of vehicle behavior. The traditional method of manually labeling vehicle behavior is too time-consuming and labor-intensive, but the existing object detection and tracking algorithms have poor practicability and low behavioral location detection rate. This paper proposes a vehicle behavior detection algorithm based on the dual-stream convolution network and the multi-dimensional video dynamic detection network. In the videos, the straight-line behavior of the vehicle will default to the background behavior. The Changing lanes, turning and turning around are set as target behaviors. The purpose of this model is to automatically mark the target behavior of the vehicle from the untrimmed videos. First, the target behavior proposals in the long video are extracted through the dual-stream convolution network. The model uses a dual-stream convolutional network to generate a one-dimensional action score waveform, and then extract segments with scores above a given threshold M into preliminary vehicle behavior proposals. Second, the preliminary proposals are pruned and identified using the multi-dimensional video dynamic detection network. Referring to the hierarchical reinforcement learning, the multi-dimensional network includes a Timer module and a Spacer module, where the Timer module mines time information in the video stream and the Spacer module extracts spatial information in the video frame. The Timer and Spacer module are implemented by Long Short-Term Memory (LSTM) and start from an all-zero hidden state. The Timer module uses the Transformer mechanism to extract timing information from the video stream and extract features by linear mapping and other methods. Finally, the model fuses time information and spatial information and obtains the location and category of the behavior through the softmax layer. This paper uses recall and precision to measure the performance of the model. Extensive experiments show that based on the dataset of this paper, the proposed model has obvious advantages compared with the existing state-of-the-art behavior detection algorithms. When the Time Intersection over Union (TIoU) threshold is 0.5, the Average-Precision (MP) reaches 36.3% (the MP of baselines is 21.5%). In summary, this paper proposes a vehicle behavior detection model based on multi-dimensional dynamic detection network. This paper introduces spatial information and temporal information to extract vehicle behaviors in long videos. Experiments show that the proposed algorithm is advanced and accurate in-vehicle timing behavior detection. In the future, the focus will be on simultaneously detecting the timing behavior of multiple vehicles in complex traffic scenes (such as a busy street) while ensuring accuracy.

Keywords: vehicle behavior detection, convolutional neural network, long short-term memory, deep learning

Procedia PDF Downloads 111
292 The Role of Risk Attitudes and Networks on the Migration Decision: Empirical Evidence from the United States

Authors: Tamanna Rimi

Abstract:

A large body of literature has discussed the determinants of migration decision. However, the potential role of individual risk attitudes on migration decision has so far been overlooked. The research on migration literature has studied how the expected income differential influences migration flows for a risk neutral individual. However, migration takes place when there is no expected income differential or even the variability of income appears as lower than in the current location. This migration puzzle motivates a recent trend in the literature that analyzes how attitudes towards risk influence the decision to migrate. However, the significance of risk attitudes on migration decision has been addressed mostly in a theoretical perspective in the mainstream migration literature. The efficient outcome of labor market and overall economy are largely influenced by migration in many countries. Therefore, attitudes towards risk as a determinant of migration should get more attention in empirical studies. To author’s best knowledge, this is the first study that has examined the relationship between relative risk aversion and migration decision in US market. This paper considers movement across United States as a means of migration. In addition, this paper also explores the network effect due to the increasing size of one’s own ethnic group to a source location on the migration decision and how attitudes towards risk vary with network effect. Two ethnic groups (i.e. Asian and Hispanic) have been considered in this regard. For the empirical estimation, this paper uses two sources of data: 1) U.S. census data for social, economic, and health research, 2010 (IPUMPS) and 2) University of Michigan Health and Retirement Study, 2010 (HRS). In order to measure relative risk aversion, this study uses the ‘Two Sample Two-Stage Instrumental Variable (TS2SIV)’ technique. This is a similar method of Angrist (1990) and Angrist and Kruegers’ (1992) ‘Two Sample Instrumental Variable (TSIV)’ technique. Using a probit model, the empirical investigation yields the following results: (i) risk attitude has a significantly large impact on migration decision where more risk averse people are less likely to migrate; (ii) the impact of risk attitude on migration varies by other demographic characteristics such as age and sex; (iii) people with higher concentration of same ethnic households living in a particular place are expected to migrate less from their current place; (iv) the risk attitudes on migration vary with network effect. The overall findings of this paper relating risk attitude, migration decision and network effect can be a significant contribution addressing the gap between migration theory and empirical study in migration literature.

Keywords: migration, network effect, risk attitude, U.S. market

Procedia PDF Downloads 147