Search results for: psychophysiological adaptive automation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1435

Search results for: psychophysiological adaptive automation

985 Real-Time Multi-Vehicle Tracking Application at Intersections Based on Feature Selection in Combination with Color Attribution

Authors: Qiang Zhang, Xiaojian Hu

Abstract:

In multi-vehicle tracking, based on feature selection, the tracking system efficiently tracks vehicles in a video with minimal error in combination with color attribution, which focuses on presenting a simple and fast, yet accurate and robust solution to the problem such as inaccurately and untimely responses of statistics-based adaptive traffic control system in the intersection scenario. In this study, a real-time tracking system is proposed for multi-vehicle tracking in the intersection scene. Considering the complexity and application feasibility of the algorithm, in the object detection step, the detection result provided by virtual loops were post-processed and then used as the input for the tracker. For the tracker, lightweight methods were designed to extract and select features and incorporate them into the adaptive color tracking (ACT) framework. And the approbatory online feature selection algorithms are integrated on the mature ACT system with good compatibility. The proposed feature selection methods and multi-vehicle tracking method are evaluated on KITTI datasets and show efficient vehicle tracking performance when compared to the other state-of-the-art approaches in the same category. And the system performs excellently on the video sequences recorded at the intersection. Furthermore, the presented vehicle tracking system is suitable for surveillance applications.

Keywords: real-time, multi-vehicle tracking, feature selection, color attribution

Procedia PDF Downloads 139
984 A Hybrid Algorithm Based on Greedy Randomized Adaptive Search Procedure and Chemical Reaction Optimization for the Vehicle Routing Problem with Hard Time Windows

Authors: Imen Boudali, Marwa Ragmoun

Abstract:

The Vehicle Routing Problem with Hard Time Windows (VRPHTW) is a basic distribution management problem that models many real-world problems. The objective of the problem is to deliver a set of customers with known demands on minimum-cost vehicle routes while satisfying vehicle capacity and hard time windows for customers. In this paper, we propose to deal with our optimization problem by using a new hybrid stochastic algorithm based on two metaheuristics: Chemical Reaction Optimization (CRO) and Greedy Randomized Adaptive Search Procedure (GRASP). The first method is inspired by the natural process of chemical reactions enabling the transformation of unstable substances with excessive energy to stable ones. During this process, the molecules interact with each other through a series of elementary reactions to reach minimum energy for their existence. This property is embedded in CRO to solve the VRPHTW. In order to enhance the population diversity throughout the search process, we integrated the GRASP in our method. Simulation results on the base of Solomon’s benchmark instances show the very satisfactory performances of the proposed approach.

Keywords: Benchmark Problems, Combinatorial Optimization, Vehicle Routing Problem with Hard Time Windows, Meta-heuristics, Hybridization, GRASP, CRO

Procedia PDF Downloads 385
983 Blood Volume Pulse Extraction for Non-Contact Photoplethysmography Measurement from Facial Images

Authors: Ki Moo Lim, Iman R. Tayibnapis

Abstract:

According to WHO estimation, 38 out of 56 million (68%) global deaths in 2012, were due to noncommunicable diseases (NCDs). To avert NCD, one of the solutions is early detection of diseases. In order to do that, we developed 'U-Healthcare Mirror', which is able to measure vital sign such as heart rate (HR) and respiration rate without any physical contact and consciousness. To measure HR in the mirror, we utilized digital camera. The camera records red, green, and blue (RGB) discoloration from user's facial image sequences. We extracted blood volume pulse (BVP) from the RGB discoloration because the discoloration of the facial skin is accordance with BVP. We used blind source separation (BSS) to extract BVP from the RGB discoloration and adaptive filters for removing noises. We utilized singular value decomposition (SVD) method to implement the BSS and the adaptive filters. HR was estimated from the obtained BVP. We did experiment for HR measurement by using our method and previous method that used independent component analysis (ICA) method. We compared both of them with HR measurement from commercial oximeter. The experiment was conducted under various distance between 30~110 cm and light intensity between 5~2000 lux. For each condition, we did measurement 7 times. The estimated HR showed 2.25 bpm of mean error and 0.73 of pearson correlation coefficient. The accuracy has improved compared to previous work. The optimal distance between the mirror and user for HR measurement was 50 cm with medium light intensity, around 550 lux.

Keywords: blood volume pulse, heart rate, photoplethysmography, independent component analysis

Procedia PDF Downloads 313
982 Optimal Simultaneous Sizing and Siting of DGs and Smart Meters Considering Voltage Profile Improvement in Active Distribution Networks

Authors: T. Sattarpour, D. Nazarpour

Abstract:

This paper investigates the effect of simultaneous placement of DGs and smart meters (SMs), on voltage profile improvement in active distribution networks (ADNs). A substantial center of attention has recently been on responsive loads initiated in power system problem studies such as distributed generations (DGs). Existence of responsive loads in active distribution networks (ADNs) would have undeniable effect on sizing and siting of DGs. For this reason, an optimal framework is proposed for sizing and siting of DGs and SMs in ADNs. SMs are taken into consideration for the sake of successful implementing of demand response programs (DRPs) such as direct load control (DLC) with end-side consumers. Looking for voltage profile improvement, the optimization procedure is solved by genetic algorithm (GA) and tested on IEEE 33-bus distribution test system. Different scenarios with variations in the number of DG units, individual or simultaneous placing of DGs and SMs, and adaptive power factor (APF) mode for DGs to support reactive power have been established. The obtained results confirm the significant effect of DRPs and APF mode in determining the optimal size and site of DGs to be connected in ADN resulting to the improvement of voltage profile as well.

Keywords: active distribution network (ADN), distributed generations (DGs), smart meters (SMs), demand response programs (DRPs), adaptive power factor (APF)

Procedia PDF Downloads 275
981 Adaptive Approach Towards Comprehensive Urban Development Simulation in Coastal Regions: Case Study of New Alamein City, Egypt

Authors: Nada Mohamed, Abdel Aziz Mohamed

Abstract:

Climate change in coastal areas is a global issue that can be felt on local scale and will be around for decades and centuries to come to an end; it also has critical risks on the city’s economy, communities, and the natural environment. One of these changes that cause a huge risk on coastal cities is the sea level rise (SLR). SLR is a result of scarcity and reduction in global environmental system. The main cause of climate change and global warming is the countries with high development index (HDI) as Japan and Germany while the medium and low HDI countries as Egypt does not have enough awareness and advanced tactics to adapt with this changes that destroy urban areas and cause loss in land and economy. This is why Climate Resilience is one of the UN sustainable development goals 2030, which is calling for actions to strengthen climate change resilience through mitigation and adaptation. For many reasons, adaptation has received less attention than mitigation and it is only recently that adaptation has become a focal global point of attention. This adaption can be achieved through some actions such as upgrading the use and the design of the land, adjusting business and activities of people, and increasing community understanding of climate risks. To reach the adaption goals, and we have to apply a strategic pathway to Climate Resilience, which is the Urban Bioregionalism Paradigm. Resiliency has been framed as persistence, adaptation, and transformation. Climate Resilience decision support system includes a visualization platform where ecological, social, and economic information can be viewed alongside with specific geographies that's why Urban Bioregionalism is a socio-ecological system which is defined as a paradigm that has potential to help move social attitudes toward environmental understanding and deepen human-environment connections within ecological development. The research aim is to achieve an adaptive integrated urban development model throughout the analyses of tactics and strategies that can be used to adapt urban areas and coastal communities to the challenges of climate changes especially SLR and also simulation model using advanced technological software for a coastal city corridor to elaborates the suitable strategy to apply.

Keywords: climate resilience, sea level rise, SLR, coastal resilience, adaptive development simulation

Procedia PDF Downloads 112
980 Substation Automation, Digitization, Cyber Risk and Chain Risk Management Reliability

Authors: Serzhan Ashirov, Dana Nour, Rafat Rob, Khaled Alotaibi

Abstract:

There has been a fast growth in the introduction and use of communications, information, monitoring, and sensing technologies. The new technologies are making their way to the Industrial Control Systems as embedded in products, software applications, IT services, or commissioned to enable integration and automation of increasingly global supply chains. As a result, the lines that separated the physical, digital, and cyber world have diminished due to the vast implementation of the new, disruptive digital technologies. The variety and increased use of these technologies introduce many cybersecurity risks affecting cyber-resilience of the supply chain, both in terms of the product or service delivered to a customer and members of the supply chain operation. US department of energy considers supply chain in the IR4 space to be the weakest link in cybersecurity. The IR4 identified the digitization of the field devices, followed by digitalization that eventually moved through the digital transformation space with little care for the new introduced cybersecurity risks. This paper will examine the best methodologies for securing the electrical substations from cybersecurity attacks due to supply chain risks, and due to digitization effort. SCADA systems are the most vulnerable part of the power system infrastructure due to digitization and due to the weakness and vulnerabilities in the supply chain security. The paper will discuss in details how create a secure supply chain methodology, secure substations, and mitigate the risks due to digitization

Keywords: cybersecurity, supply chain methodology, secure substation, digitization

Procedia PDF Downloads 45
979 Detecting Tomato Flowers in Greenhouses Using Computer Vision

Authors: Dor Oppenheim, Yael Edan, Guy Shani

Abstract:

This paper presents an image analysis algorithm to detect and count yellow tomato flowers in a greenhouse with uneven illumination conditions, complex growth conditions and different flower sizes. The algorithm is designed to be employed on a drone that flies in greenhouses to accomplish several tasks such as pollination and yield estimation. Detecting the flowers can provide useful information for the farmer, such as the number of flowers in a row, and the number of flowers that were pollinated since the last visit to the row. The developed algorithm is designed to handle the real world difficulties in a greenhouse which include varying lighting conditions, shadowing, and occlusion, while considering the computational limitations of the simple processor in the drone. The algorithm identifies flowers using an adaptive global threshold, segmentation over the HSV color space, and morphological cues. The adaptive threshold divides the images into darker and lighter images. Then, segmentation on the hue, saturation and volume is performed accordingly, and classification is done according to size and location of the flowers. 1069 images of greenhouse tomato flowers were acquired in a commercial greenhouse in Israel, using two different RGB Cameras – an LG G4 smartphone and a Canon PowerShot A590. The images were acquired from multiple angles and distances and were sampled manually at various periods along the day to obtain varying lighting conditions. Ground truth was created by manually tagging approximately 25,000 individual flowers in the images. Sensitivity analyses on the acquisition angle of the images, periods throughout the day, different cameras and thresholding types were performed. Precision, recall and their derived F1 score were calculated. Results indicate better performance for the view angle facing the flowers than any other angle. Acquiring images in the afternoon resulted with the best precision and recall results. Applying a global adaptive threshold improved the median F1 score by 3%. Results showed no difference between the two cameras used. Using hue values of 0.12-0.18 in the segmentation process provided the best results in precision and recall, and the best F1 score. The precision and recall average for all the images when using these values was 74% and 75% respectively with an F1 score of 0.73. Further analysis showed a 5% increase in precision and recall when analyzing images acquired in the afternoon and from the front viewpoint.

Keywords: agricultural engineering, image processing, computer vision, flower detection

Procedia PDF Downloads 302
978 MLProxy: SLA-Aware Reverse Proxy for Machine Learning Inference Serving on Serverless Computing Platforms

Authors: Nima Mahmoudi, Hamzeh Khazaei

Abstract:

Serving machine learning inference workloads on the cloud is still a challenging task at the production level. The optimal configuration of the inference workload to meet SLA requirements while optimizing the infrastructure costs is highly complicated due to the complex interaction between batch configuration, resource configurations, and variable arrival process. Serverless computing has emerged in recent years to automate most infrastructure management tasks. Workload batching has revealed the potential to improve the response time and cost-effectiveness of machine learning serving workloads. However, it has not yet been supported out of the box by serverless computing platforms. Our experiments have shown that for various machine learning workloads, batching can hugely improve the system’s efficiency by reducing the processing overhead per request. In this work, we present MLProxy, an adaptive reverse proxy to support efficient machine learning serving workloads on serverless computing systems. MLProxy supports adaptive batching to ensure SLA compliance while optimizing serverless costs. We performed rigorous experiments on Knative to demonstrate the effectiveness of MLProxy. We showed that MLProxy could reduce the cost of serverless deployment by up to 92% while reducing SLA violations by up to 99% that can be generalized across state-of-the-art model serving frameworks.

Keywords: serverless computing, machine learning, inference serving, Knative, google cloud run, optimization

Procedia PDF Downloads 153
977 Metabolic and Adaptive Laboratory Evolutionary Engineering (ALE) of Saccharomyces cerevisiae for Second Generation Biofuel Production

Authors: Farnaz Yusuf, Naseem A. Gaur

Abstract:

The increase in environmental concerns, rapid depletion of fossil fuel reserves and intense interest in achieving energy security has led to a global research effort towards developing renewable sources of fuels. Second generation biofuels have attracted more attention recently as the use of lignocellulosic biomass can reduce fossil fuel dependence and is environment-friendly. Xylose is the main pentose and second most abundant sugar after glucose in lignocelluloses. Saccharomyces cerevisiae does not readily uptake and use pentose sugars. For an economically feasible biofuel production, both hexose and pentose sugars must be fermented to ethanol. Therefore, it is important to develop S. cerevisiae host platforms with more efficient xylose utilization. This work aims to construct a xylose fermenting yeast strains with engineered oxido-reductative pathway for xylose metabolism. Engineered strain was further improved by adaptive evolutionary engineering approach. The engineered strain is able to grow on xylose as sole carbon source with the maximum ethanol yield of 0.39g/g xylose and productivity of 0.139g/l/h at 96 hours. The further improvement in strain development involves over expression of pentose phosphate pathway and protein engineering of xylose reductase/xylitol dehydrogenase to change their cofactor specificity in order to reduce xylitol accumulation.

Keywords: biofuel, lignocellulosic biomass, saccharomyces cerevisiae, xylose

Procedia PDF Downloads 188
976 Lockit: A Logic Locking Automation Software

Authors: Nemanja Kajtez, Yue Zhan, Basel Halak

Abstract:

The significant rise in the cost of manufacturing of nanoscale integrated circuits (IC) has led the majority of IC design companies to outsource the fabrication of their products to other companies, often located in different countries. This multinational nature of the hardware supply chain has led to a host of security threats, including IP piracy, IC overproduction, and Trojan insertion. To combat that, researchers have proposed logic locking techniques to protect the intellectual properties of the design and increase the difficulty of malicious modification of its functionality. However, the adoption of logic locking approaches is rather slow due to the lack of the integration with IC production process and the lack of efficacy of existing algorithms. This work automates the logic locking process by developing software using Python that performs the locking on a gate-level netlist and can be integrated with the existing digital synthesis tools. Analysis of the latest logic locking algorithms has demonstrated that the SFLL-HD algorithm is one of the most secure and versatile in trading-off levels of protection against different types of attacks and was thus selected for implementation. The presented tool can also be expanded to incorporate the latest locking mechanisms to keep up with the fast-paced development in this field. The paper also presents a case study to demonstrate the functionality of the tool and how it could be used to explore the design space and compare different locking solutions. The source code of this tool is available freely from (https://www.researchgate.net/publication/353195333_Source_Code_for_The_Lockit_Tool).

Keywords: design automation, hardware security, IP piracy, logic locking

Procedia PDF Downloads 157
975 Diversity for Safety and Security of Autonomous Vehicles against Accidental and Deliberate Faults

Authors: Anil Ranjitbhai Patel, Clement John Shaji, Peter Liggesmeyer

Abstract:

Safety and security of autonomous vehicles (AVs) is a growing concern, first, due to the increased number of safety-critical functions taken over by automotive embedded systems; second, due to the increased exposure of the software-intensive systems to potential attackers; third, due to dynamic interaction in an uncertain and unknown environment at runtime which results in changed functional and non-functional properties of the system. Frequently occurring environmental uncertainties, random component failures, and compromise security of the AVs might result in hazardous events, sometimes even in an accident, if left undetected. Beyond these technical issues, we argue that the safety and security of AVs against accidental and deliberate faults are poorly understood and rarely implemented. One possible way to overcome this is through a well-known diversity approach. As an effective approach to increase safety and security, diversity has been widely used in the aviation, railway, and aerospace industries. Thus, the paper proposes fault-tolerance by diversity model takes into consideration the mitigation of accidental and deliberate faults by application of structure and variant redundancy. The model can be used to design the AVs with various types of diversity in hardware and software-based multi-version system. The paper evaluates the presented approach by employing an example from adaptive cruise control, followed by discussing the case study with initial findings.

Keywords: autonomous vehicles, diversity, fault-tolerance, adaptive cruise control, safety, security

Procedia PDF Downloads 107
974 The Influence of Training and Competition on Cortisol Levels and Sleep in Elite Female Athletes

Authors: Shannon O’Donnell, Matthew Driller, Gregory Jacobson, Steve Bird

Abstract:

Stress hormone levels in a competition vs. training setting are yet to be evaluated in elite female athletes. The effect that these levels of stress have on subsequent sleep quality and quantity is also yet to be investigated. The aim of the current study was to evaluate different psychophysiological stress markers in competition and training environments and the subsequent effect on sleep indices in an elite female athlete population. The study involved 10 elite female netball athletes (mean ± SD; age = 23 ± 6 yrs) providing multiple salivary hormone measures and having their sleep monitored on two occasions; a match day, and a training day. The training and match were performed at the same time of day and were matched for intensity and duration. Saliva samples were collected immediately pre (5:00 pm) and post session (7:15 pm), and at 10:00 pm and were analysed for cortisol concentrations. Sleep monitoring was performed using wrist actigraphy to assess total sleep time (TST), sleep efficiency (SE%) and sleep latency (SL). Cortisol levels were significantly higher (p < 0.01) immediately post the match vs post training (mean ± SD; 0.925 ± 0.341 μg/dL and 0.239 ± 0.284 μg/dL, respectively) and at 10:00pm (0.143 ± 0.085 μg/dL and 0.072 ± 0.064 μg/dL, respectively, p < 0.01). The difference between trials was associated with a very large effect (ES: 2.23) immediately post (7:15 pm) and a large effect (ES: 1.02) at 10:00 pm. There was a significant reduction in TST (mean ± SD; -117.9 ± 111.9 minutes, p < 0.01, ES: -1.89) and SE% (-7.7 ± 8.5%, p < 0.02, ES: -0.79) on the night following the netball match compared to the training session. Although not significant (p > 0.05), there was an increase in SL following the netball match v the training session (67.0 ± 51.9 minutes and 38.5 ± 29.3 minutes, respectively), which was associated with a moderate effect (ES: 0.80). The current study reports that cortisol levels are significantly higher and subsequent sleep quantity and quality is significantly reduced in elite female athletes following a match compared to a training session.

Keywords: cortisol, netball, performance, recovery

Procedia PDF Downloads 239
973 Parkinson’s Disease Detection Analysis through Machine Learning Approaches

Authors: Muhtasim Shafi Kader, Fizar Ahmed, Annesha Acharjee

Abstract:

Machine learning and data mining are crucial in health care, as well as medical information and detection. Machine learning approaches are now being utilized to improve awareness of a variety of critical health issues, including diabetes detection, neuron cell tumor diagnosis, COVID 19 identification, and so on. Parkinson’s disease is basically a disease for our senior citizens in Bangladesh. Parkinson's Disease indications often seem progressive and get worst with time. People got affected trouble walking and communicating with the condition advances. Patients can also have psychological and social vagaries, nap problems, hopelessness, reminiscence loss, and weariness. Parkinson's disease can happen in both men and women. Though men are affected by the illness at a proportion that is around partial of them are women. In this research, we have to get out the accurate ML algorithm to find out the disease with a predictable dataset and the model of the following machine learning classifiers. Therefore, nine ML classifiers are secondhand to portion study to use machine learning approaches like as follows, Naive Bayes, Adaptive Boosting, Bagging Classifier, Decision Tree Classifier, Random Forest classifier, XBG Classifier, K Nearest Neighbor Classifier, Support Vector Machine Classifier, and Gradient Boosting Classifier are used.

Keywords: naive bayes, adaptive boosting, bagging classifier, decision tree classifier, random forest classifier, XBG classifier, k nearest neighbor classifier, support vector classifier, gradient boosting classifier

Procedia PDF Downloads 108
972 Crisis In/Out, Emergent, and Adaptive Urban Organisms

Authors: Alessandra Swiny, Michalis Georgiou, Yiorgos Hadjichristou

Abstract:

This paper focuses on the questions raised through the work of Unit 5: ‘In/Out of crisis, emergent and adaptive’; an architectural research-based studio at the University of Nicosia. It focusses on sustainable architectural and urban explorations tackling with the ever growing crises in its various types, phases and locations. ‘Great crisis situations’ are seen as ‘great chances’ that trigger investigations for further development and evolution of the built environment in an ultimate sustainable approach. The crisis is taken as an opportunity to rethink the urban and architectural directions as new forces for inventions leading to emergent and adaptive built environments. The Unit 5’s identity and environment facilitates the students to respond optimistically, alternatively and creatively towards the global current crisis. Mark Wigley’s notion that “crises are ultimately productive” and “They force invention” intrigued and defined the premises of the Unit. ‘Weather and nature are coauthors of the built environment’ Jonathan Hill states in his ‘weather architecture’ discourse. The weather is constantly changing and new environments, the subnatures are created which derived from the human activities David Gissen explains. The above set of premises triggered innovative responses by the Unit’s students. They thoroughly investigated the various kinds of crisis and their causes in relation to their various types of Terrains. The tools used for the research and investigation were chosen in contradictive pairs to generate further crisis situations: The re-used/salvaged competed with the new, the handmade rivalling with the fabrication, the analogue juxtaposed with digital. Students were asked to delve into state of art technologies in order to propose sustainable emergent and adaptive architectures and Urbanities, having though always in mind that the human and the social aspects of the community should be the core of the investigation. The resulting unprecedented spatial conditions and atmospheres of the emergent new ways of living are deemed to be the ultimate aim of the investigation. Students explored a variety of sites and crisis conditions such as: The vague terrain of the Green Line in Nicosia, the lost footprints of the sinking Venice, the endangered Australian coral reefs, the earthquake torn town of Crevalcore, and the decaying concrete urbanscape of Athens. Among other projects, ‘the plume project’ proposes a cloud-like, floating and almost dream-like living environment with unprecedented spatial conditions to the inhabitants of the coal mine of Centralia, USA, not just to enable them to survive but even to prosper in this unbearable environment due to the process of the captured plumes of smoke and heat. Existing water wells inspire inversed vertical structures creating a new living underground network, protecting the nomads from catastrophic sand storms in the Araoune of Mali. “Inverted utopia: Lost things in the sand”, weaves a series of tea-houses and a library holding lost artifacts and transcripts into a complex underground labyrinth by the utilization of the sand solidification technology. Within this methodology, crisis is seen as a mechanism for allowing an emergence of new and fascinating ultimate sustainable future cultures and cities.

Keywords: adaptive built environments, crisis as opportunity, emergent urbanities, forces for inventions

Procedia PDF Downloads 413
971 Disaster Adaptation Mechanism and Disaster Prevention Adaptation Planning Strategies for Industrial Parks in Response to Climate Change and Different Socio-Economic Disasters

Authors: Jen-Te Pai, Jao-Heng Liu, Shin-En Pai

Abstract:

The impact of climate change has intensified in recent years, causing Taiwan to face higher frequency and serious natural disasters. Therefore, it is imperative for industrial parks manufacturers to promote adaptation policies in response to climate change. On the other hand, with the rise of the international anti-terrorism situation, once a terrorist attack occurs, it will attract domestic and international media attention, especially the strategic and economic status of the science park. Thus, it is necessary to formulate adaptation and mitigation strategies under climate change and social economic disasters. After reviewed the literature about climate change, urban disaster prevention, vulnerability assessment, and risk communication, the study selected 62 industrial parks compiled by the Industrial Bureau of the Ministry of Economic Affairs of Taiwan as the research object. This study explored the vulnerability and disaster prevention and disaster relief functional assessment of these industrial parks facing of natural and socio-economic disasters. Furthermore, this study explored planned adaptation of industrial parks management section and autonomous adaptation of corporate institutions in the park. The conclusion of this study is that Taiwan industrial parks with a higher vulnerability to natural and socio-economic disasters should employ positive adaptive behaviours.

Keywords: adaptive behaviours, analytic network process, vulnerability, industrial parks

Procedia PDF Downloads 119
970 Fueling Efficient Reporting And Decision-Making In Public Health With Large Data Automation In Remote Areas, Neno Malawi

Authors: Wiseman Emmanuel Nkhomah, Chiyembekezo Kachimanga, Julia Huggins, Fabien Munyaneza

Abstract:

Background: Partners In Health – Malawi introduced one of Operational Researches called Primary Health Care (PHC) Surveys in 2020, which seeks to assess progress of delivery of care in the district. The study consists of 5 long surveys, namely; Facility assessment, General Patient, Provider, Sick Child, Antenatal Care (ANC), primarily conducted in 4 health facilities in Neno district. These facilities include Neno district hospital, Dambe health centre, Chifunga and Matope. Usually, these annual surveys are conducted from January, and the target is to present final report by June. Once data is collected and analyzed, there are a series of reviews that take place before reaching final report. In the first place, the manual process took over 9 months to present final report. Initial findings reported about 76.9% of the data that added up when cross-checked with paper-based sources. Purpose: The aim of this approach is to run away from manually pulling the data, do fresh analysis, and reporting often associated not only with delays in reporting inconsistencies but also with poor quality of data if not done carefully. This automation approach was meant to utilize features of new technologies to create visualizations, reports, and dashboards in Power BI that are directly fished from the data source – CommCare hence only require a single click of a ‘refresh’ button to have the updated information populated in visualizations, reports, and dashboards at once. Methodology: We transformed paper-based questionnaires into electronic using CommCare mobile application. We further connected CommCare Mobile App directly to Power BI using Application Program Interface (API) connection as data pipeline. This provided chance to create visualizations, reports, and dashboards in Power BI. Contrary to the process of manually collecting data in paper-based questionnaires, entering them in ordinary spreadsheets, and conducting analysis every time when preparing for reporting, the team utilized CommCare and Microsoft Power BI technologies. We utilized validations and logics in CommCare to capture data with less errors. We utilized Power BI features to host the reports online by publishing them as cloud-computing process. We switched from sharing ordinary report files to sharing the link to potential recipients hence giving them freedom to dig deep into extra findings within Power BI dashboards and also freedom to export to any formats of their choice. Results: This data automation approach reduced research timelines from the initial 9 months’ duration to 5. It also improved the quality of the data findings from the original 76.9% to 98.9%. This brought confidence to draw conclusions from the findings that help in decision-making and gave opportunities for further researches. Conclusion: These results suggest that automating the research data process has the potential of reducing overall amount of time spent and improving the quality of the data. On this basis, the concept of data automation should be taken into serious consideration when conducting operational research for efficiency and decision-making.

Keywords: reporting, decision-making, power BI, commcare, data automation, visualizations, dashboards

Procedia PDF Downloads 100
969 Development of Sustainable Building Environmental Model (SBEM) in Hong Kong

Authors: Kwok W. Mui, Ling T. Wong, F. Xiao, Chin T. Cheung, Ho C. Yu

Abstract:

This study addresses a concept of the Sustainable Building Environmental Model (SBEM) developed to optimize energy consumption in air conditioning and ventilation (ACV) systems without any deterioration of indoor environmental quality (IEQ). The SBEM incorporates two main components: an adaptive comfort temperature control module (ACT) and a new carbon dioxide demand control module (nDCV). These two modules take an innovative approach to maintain satisfaction of the Indoor Environmental Quality (IEQ) with optimum energy consumption, they provide a rational basis of effective control. A total of 2133 sets of measurement data of indoor air temperature (Ta), relative humidity (Rh) and carbon dioxide concentration (CO2) were conducted in some Hong Kong offices to investigate the potential of integrating the SBEM. A simulation was used to evaluate the dynamic performance of the energy and air conditioning system with the integration of the SBEM in an air-conditioned building. It allows us make a clear picture of the control strategies and performed any pre-tuned of controllers before utilized in real systems. With the integration of SBEM, it was able to save up to 12.3% in simulation and 15% in field measurement of overall electricity consumption, and maintain the average carbon dioxide concentration within 1000ppm and occupant dissatisfaction in 20%.

Keywords: sustainable building environmental model (SBEM), adaptive comfort temperature (ACT), new demand control ventilation (nDCV), energy saving

Procedia PDF Downloads 623
968 The Response of Adaptive Mechanism of Fluorescent Proteins from Coral Species and Target Cell Properties on Signalling Capacity as Biosensor

Authors: Elif Tugce Aksun Tumerkan

Abstract:

Fluorescent proteins (FPs) have become very popular since green fluorescent protein discovered from crystal jellyfish. It is known that Anthozoa species have a wide range of chromophore organisms, and the initial crystal structure for non-fluorescent chromophores obtained from the reef-building coral has been determined. There are also differently coloured pigments in non-bioluminescent Anthozoa zooxanthellate and azooxanthellate which are frequently members of the GFP-like protein family. The development of fluorescent proteins (FPs) and their applications is an outstanding example of basic science leading to practical biotechnological and medical applications. Fluorescent proteins have several applications in science and are used as important indicators in molecular biology and cell-based research. With rising interest in cell biology, FPs have used as biosensor indicators and probes in pharmacology and cell biology. Using fluorescent proteins in genetically encoded metabolite sensors has many advantages than chemical probes for metabolites such as easily introduced into any cell or organism in any sub-cellular localization and giving chance to fixing to fluoresce of different colours or characteristics. There are different factors effects to signalling mechanism when they used as a biosensor. While there are wide ranges of research have been done on the significance and applications of fluorescent proteins, the cell signalling response of FPs and target cell are less well understood. In this study, it was aimed to clarify the response of adaptive mechanisms of coral species such as pH, temperature and symbiotic relationship and target cells properties on the signalling capacity. Corals are a rich natural source of fluorescent proteins that change with environmental conditions such as light, heat stress and injury. Adaptation mechanism of coral species to these types of environmental variations is important factor due to FPs properties have affected by this mechanism. Since fluorescent proteins obtained from nature, their own ecological property like the symbiotic relationship is observed very commonly in coral species and living conditions have the impact on FPs efficiency. Target cell properties also have an effect on signalling and visualization. The dynamicity of detector that used for reading fluorescence and the level of background fluorescence are key parameters for the quality of the fluorescent signal. Among the factors, it can be concluded that coral species adaptive characteristics have the strongest effect on FPs signalling capacity.

Keywords: biosensor, cell biology, environmental conditions, fluorescent protein, sea anemone

Procedia PDF Downloads 151
967 Application of Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Multipoint Optimal Minimum Entropy Deconvolution in Railway Bearings Fault Diagnosis

Authors: Yao Cheng, Weihua Zhang

Abstract:

Although the measured vibration signal contains rich information on machine health conditions, the white noise interferences and the discrete harmonic coming from blade, shaft and mash make the fault diagnosis of rolling element bearings difficult. In order to overcome the interferences of useless signals, a new fault diagnosis method combining Complete Ensemble Empirical Mode Decomposition with adaptive noise (CEEMDAN) and Multipoint Optimal Minimum Entropy Deconvolution (MOMED) is proposed for the fault diagnosis of high-speed train bearings. Firstly, the CEEMDAN technique is applied to adaptively decompose the raw vibration signal into a series of finite intrinsic mode functions (IMFs) and a residue. Compared with Ensemble Empirical Mode Decomposition (EEMD), the CEEMDAN can provide an exact reconstruction of the original signal and a better spectral separation of the modes, which improves the accuracy of fault diagnosis. An effective sensitivity index based on the Pearson's correlation coefficients between IMFs and raw signal is adopted to select sensitive IMFs that contain bearing fault information. The composite signal of the sensitive IMFs is applied to further analysis of fault identification. Next, for propose of identifying the fault information precisely, the MOMED is utilized to enhance the periodic impulses in composite signal. As a non-iterative method, the MOMED has better deconvolution performance than the classical deconvolution methods such Minimum Entropy Deconvolution (MED) and Maximum Correlated Kurtosis Deconvolution (MCKD). Third, the envelope spectrum analysis is applied to detect the existence of bearing fault. The simulated bearing fault signals with white noise and discrete harmonic interferences are used to validate the effectiveness of the proposed method. Finally, the superiorities of the proposed method are further demonstrated by high-speed train bearing fault datasets measured from test rig. The analysis results indicate that the proposed method has strong practicability.

Keywords: bearing, complete ensemble empirical mode decomposition with adaptive noise, fault diagnosis, multipoint optimal minimum entropy deconvolution

Procedia PDF Downloads 348
966 Material Handling Equipment Selection Using Fuzzy AHP Approach

Authors: Priyanka Verma, Vijaya Dixit, Rishabh Bajpai

Abstract:

This research paper is aimed at selecting appropriate material handling equipment among the given choices so that the automation level in material handling can be enhanced. This work is a practical case scenario of material handling systems in consumer electronic appliances manufacturing organization. The choices of material handling equipment among which the decision has to be made are Automated Guided Vehicle’s (AGV), Autonomous Mobile Robots (AMR), Overhead Conveyer’s (OC) and Battery Operated Trucks/Vehicle’s (BOT). There is a need of attaining a certain level of automation in order to reduce human interventions in the organization. This requirement of achieving certain degree of automation can be attained by material handling equipment’s mentioned above. The main motive for selecting above equipment’s for study was solely based on corporate financial strategy of investment and return obtained through that investment made in stipulated time framework. Since the low cost automation with respect to material handling devices has to be achieved hence these equipment’s were selected. Investment to be done on each unit of this equipment is less than 20 lakh rupees (INR) and the recovery period is less than that of five years. Fuzzy analytic hierarchic process (FAHP) is applied here for selecting equipment where the four choices are evaluated on basis of four major criteria’s and 13 sub criteria’s, and are prioritized on the basis of weight obtained. The FAHP used here make use of triangular fuzzy numbers (TFN). The inability of the traditional AHP in order to deal with the subjectiveness and impreciseness in the pair-wise comparison process has been improved in the FAHP. The range of values for general rating purposes for all decision making parameters is kept between 0 and 1 on the basis of expert opinions captured on shop floor. These experts were familiar with operating environment and shop floor activity control. Instead of generating exact value the FAHP generates the ranges of values to accommodate the uncertainty in decision-making process. The four major criteria’s selected for the evaluation of choices of material handling equipment’s available are materials, technical capabilities, cost and other features. The thirteen sub criteria’s listed under these following four major criteria’s are weighing capacity, load per hour, material compatibility, capital cost, operating cost and maintenance cost, speed, distance moved, space required, frequency of trips, control required, safety and reliability issues. The key finding shows that among the four major criteria selected, cost is emerged as the most important criteria and is one of the key decision making aspect on the basis of which material equipment selection is based on. While further evaluating the choices of equipment available for each sub criteria it is found that AGV scores the highest weight in most of the sub-criteria’s. On carrying out complete analysis the research shows that AGV is the best material handling equipment suiting all decision criteria’s selected in FAHP and therefore it is beneficial for the organization to carry out automated material handling in the facility using AGV’s.

Keywords: fuzzy analytic hierarchy process (FAHP), material handling equipment, subjectiveness, triangular fuzzy number (TFN)

Procedia PDF Downloads 417
965 Emotional Awareness and Working Memory as Predictive Factors for the Habitual Use of Cognitive Reappraisal among Adolescents

Authors: Yuri Kitahara

Abstract:

Background: Cognitive reappraisal refers to an emotion regulation strategy in which one changes the interpretation of emotion-eliciting events. Numerous studies show that cognitive reappraisal is associated with mental health and better social functioning. However the examination of the predictive factors of adaptive emotion regulation remains as an issue. The present study examined the factors contributing to the habitual use of cognitive reappraisal, with a focus on emotional awareness and working memory. Methods: Data was collected from 30 junior high school students, using a Japanese version of the Emotion Regulation Questionnaire (ERQ), the Levels of Emotional Awareness Scale for Children (LEAS-C), and N-back task. Results: A positive correlation between emotional awareness and cognitive reappraisal was observed in the high-working-memory group (r = .54, p < .05), whereas no significant relationship was found in the low-working-memory group. In addition, the results of the analysis of variance (ANOVA) showed a significant interaction between emotional awareness and working memory capacity (F(1, 26) = 7.74, p < .05). Subsequent analysis of simple main effects confirmed that high working memory capacity significantly increases the use of cognitive reappraisal for high-emotional-awareness subjects, and significantly decreases the use of cognitive reappraisal for low-emotional-awareness subjects. Discussion: These results indicate that under the condition when one has an adequate ability for simultaneous processing of information, explicit understanding of emotion would contribute to adaptive cognitive emotion regulation. The findings are discussed along with neuroscientific claims.

Keywords: cognitive reappraisal, emotional awareness, emotion regulation, working memory

Procedia PDF Downloads 203
964 Bridging Healthcare Information Systems and Customer Relationship Management for Effective Pandemic Response

Authors: Sharda Kumari

Abstract:

As the Covid-19 pandemic continues to leave its mark on the global business landscape, companies have had to adapt to new realities and find ways to sustain their operations amid social distancing measures, government restrictions, and heightened public health concerns. This unprecedented situation has placed considerable stress on both employees and employers, underscoring the need for innovative approaches to manage the risks associated with Covid-19 transmission in the workplace. In response to these challenges, the pandemic has accelerated the adoption of digital technologies, with an increasing preference for remote interactions and virtual collaboration. Customer relationship management (CRM) systems have risen to prominence as a vital resource for organizations navigating the post-pandemic world, providing a range of benefits that include acquiring new customers, generating insightful consumer data, enhancing customer relationships, and growing market share. In the context of pandemic management, CRM systems offer three primary advantages: (1) integration features that streamline operations and reduce the need for multiple, costly software systems; (2) worldwide accessibility from any internet-enabled device, facilitating efficient remote workforce management during a pandemic; and (3) the capacity for rapid adaptation to changing business conditions, given that most CRM platforms boast a wide array of remotely deployable business growth solutions, a critical attribute when dealing with a dispersed workforce in a pandemic-impacted environment. These advantages highlight the pivotal role of CRM systems in helping organizations remain resilient and adaptive in the face of ongoing global challenges.

Keywords: healthcare, CRM, customer relationship management, customer experience, digital transformation, pandemic response, patient monitoring, patient management, healthcare automation, electronic health record, patient billing, healthcare information systems, remote workforce, virtual collaboration, resilience, adaptable business models, integration features, CRM in healthcare, telehealth, pandemic management

Procedia PDF Downloads 84
963 Factors Affecting Employee Decision Making in an AI Environment

Authors: Yogesh C. Sharma, A. Seetharaman

Abstract:

The decision-making process in humans is a complicated system influenced by a variety of intrinsic and extrinsic factors. Human decisions have a ripple effect on subsequent decisions. In this study, the scope of human decision making is limited to employees. In an organisation, a person makes a variety of decisions from the time they are hired to the time they retire. The goal of this research is to identify various elements that influence decision-making. In addition, the environment in which a decision is made is a significant aspect of the decision-making process. Employees in today's workplace use artificial intelligence (AI) systems for automation and decision augmentation. The impact of AI systems on the decision-making process is examined in this study. This research is designed based on a systematic literature review. Based on gaps in the literature, limitations and the scope of future research have been identified. Based on these findings, a research framework has been designed to identify various factors affecting employee decision making. Employee decision making is influenced by technological advancement, data-driven culture, human trust, decision automation-augmentation, and workplace motivation. Hybrid human-AI systems require the development of new skill sets and organisational design. Employee psychological safety and supportive leadership influences overall job satisfaction.

Keywords: employee decision making, artificial intelligence (AI) environment, human trust, technology innovation, psychological safety

Procedia PDF Downloads 86
962 Multi-Modal Film Boiling Simulations on Adaptive Octree Grids

Authors: M. Wasy Akhtar

Abstract:

Multi-modal film boiling simulations are carried out on adaptive octree grids. The liquid-vapor interface is captured using the volume-of-fluid framework adjusted to account for exchanges of mass, momentum, and energy across the interface. Surface tension effects are included using a volumetric source term in the momentum equations. The phase change calculations are conducted based on the exact location and orientation of the interface; however, the source terms are calculated using the mixture variables to be consistent with the one field formulation used to represent the entire fluid domain. The numerical model on octree representation of the computational grid is first verified using test cases including advection tests in severely deforming velocity fields, gravity-based instabilities and bubble growth in uniformly superheated liquid under zero gravity. The model is then used to simulate both single and multi-modal film boiling simulations. The octree grid is dynamically adapted in order to maintain the highest grid resolution on the instability fronts using markers of interface location, volume fraction, and thermal gradients. The method thus provides an efficient platform to simulate fluid instabilities with or without phase change in the presence of body forces like gravity or shear layer instabilities.

Keywords: boiling flows, dynamic octree grids, heat transfer, interface capturing, phase change

Procedia PDF Downloads 228
961 Farmers' Perception of the Effects of Climate Change on Rice Production in Nasarawa State, Nigeria

Authors: P. O. Fatoki, R. S. Olaleye, B. O. Adeniji

Abstract:

The study investigated farmers’ perception of the effects of climate change on rice production in Nasarawa State, Nigeria. Multi-stage sampling technique was used in selecting a total of 248 rice farmers from the study area. Data for the study were collected through the use of interview schedule. The data were analysed using both descriptive and inferential statistics. Results showed that majority (71.8%) of the respondents were married and the mean age of the respondents was 44.54 years. The results also showed that most adapted strategies for mitigating the effects of climate change on rice production were change of planting and harvesting date (67.7%), movement to another site (63.7%) and increased or reduced land size (58.5%). Relationship between the roles of extension agents in mitigating climate change effects on rice production and farmers’ perception were significant as revealed Chi-Square analysis from the study ; Dissemination of information ( = 2.16, P < 0.05) and use of demonstration methods ( = 2.15, P < 0.05). Poisson regression analysis revealed that educational status, farm size, experience and yield had significant relationship with the perception of the effects of climate change at 0.01 significance level while household size was as well significant at 0.05. It is recommended that some of the adaptive strategies and practices for mitigating the effects of climate change in rice production should be improved, while the extension outfits should be strengthened to ensure adequate dissemination of relevant information on climate change with a view to mitigate its effects on rice production.

Keywords: perception, rice farmers, climate change, mitigation, adaptive strategies

Procedia PDF Downloads 331
960 Choice Analysis of Ground Access to São Paulo/Guarulhos International Airport Using Adaptive Choice-Based Conjoint Analysis (ACBC)

Authors: Carolina Silva Ansélmo

Abstract:

Airports are demand-generating poles that affect the flow of traffic around them. The airport access system must be fast, convenient, and adequately planned, considering its potential users. An airport with good ground access conditions can provide the user with a more satisfactory access experience. When several transport options are available, service providers must understand users' preferences and the expected quality of service. The present study focuses on airport access in a comparative scenario between bus, private vehicle, subway, taxi and urban mobility transport applications to São Paulo/Guarulhos International Airport. The objectives are (i) to identify the factors that influence the choice, (ii) to measure Willingness to Pay (WTP), and (iii) to estimate the market share for each modal. The applied method was Adaptive Choice-based Conjoint Analysis (ACBC) technique using Sawtooth Software. Conjoint analysis, rooted in Utility Theory, is a survey technique that quantifies the customer's perceived utility when choosing alternatives. Assessing user preferences provides insights into their priorities for product or service attributes. An additional advantage of conjoint analysis is its requirement for a smaller sample size compared to other methods. Furthermore, ACBC provides valuable insights into consumers' preferences, willingness to pay, and market dynamics, aiding strategic decision-making to provide a better customer experience, pricing, and market segmentation. In the present research, the ACBC questionnaire had the following variables: (i) access time to the boarding point, (ii) comfort in the vehicle, (iii) number of travelers together, (iv) price, (v) supply power, and (vi) type of vehicle. The case study questionnaire reached 213 valid responses considering the scenario of access from the São Paulo city center to São Paulo/Guarulhos International Airport. As a result, the price and the number of travelers are the most relevant attributes for the sample when choosing airport access. The market share of the selection is mainly urban mobility transport applications, followed by buses, private vehicles, taxis and subways.

Keywords: adaptive choice-based conjoint analysis, ground access to airport, market share, willingness to pay

Procedia PDF Downloads 54
959 General Architecture for Automation of Machine Learning Practices

Authors: U. Borasi, Amit Kr. Jain, Rakesh, Piyush Jain

Abstract:

Data collection, data preparation, model training, model evaluation, and deployment are all processes in a typical machine learning workflow. Training data needs to be gathered and organised. This often entails collecting a sizable dataset and cleaning it to remove or correct any inaccurate or missing information. Preparing the data for use in the machine learning model requires pre-processing it after it has been acquired. This often entails actions like scaling or normalising the data, handling outliers, selecting appropriate features, reducing dimensionality, etc. This pre-processed data is then used to train a model on some machine learning algorithm. After the model has been trained, it needs to be assessed by determining metrics like accuracy, precision, and recall, utilising a test dataset. Every time a new model is built, both data pre-processing and model training—two crucial processes in the Machine learning (ML) workflow—must be carried out. Thus, there are various Machine Learning algorithms that can be employed for every single approach to data pre-processing, generating a large set of combinations to choose from. Example: for every method to handle missing values (dropping records, replacing with mean, etc.), for every scaling technique, and for every combination of features selected, a different algorithm can be used. As a result, in order to get the optimum outcomes, these tasks are frequently repeated in different combinations. This paper suggests a simple architecture for organizing this largely produced “combination set of pre-processing steps and algorithms” into an automated workflow which simplifies the task of carrying out all possibilities.

Keywords: machine learning, automation, AUTOML, architecture, operator pool, configuration, scheduler

Procedia PDF Downloads 35
958 An Adaptive Decomposition for the Variability Analysis of Observation Time Series in Geophysics

Authors: Olivier Delage, Thierry Portafaix, Hassan Bencherif, Guillaume Guimbretiere

Abstract:

Most observation data sequences in geophysics can be interpreted as resulting from the interaction of several physical processes at several time and space scales. As a consequence, measurements time series in geophysics have often characteristics of non-linearity and non-stationarity and thereby exhibit strong fluctuations at all time-scales and require a time-frequency representation to analyze their variability. Empirical Mode Decomposition (EMD) is a relatively new technic as part of a more general signal processing method called the Hilbert-Huang transform. This analysis method turns out to be particularly suitable for non-linear and non-stationary signals and consists in decomposing a signal in an auto adaptive way into a sum of oscillating components named IMFs (Intrinsic Mode Functions), and thereby acts as a bank of bandpass filters. The advantages of the EMD technic are to be entirely data driven and to provide the principal variability modes of the dynamics represented by the original time series. However, the main limiting factor is the frequency resolution that may give rise to the mode mixing phenomenon where the spectral contents of some IMFs overlap each other. To overcome this problem, J. Gilles proposed an alternative entitled “Empirical Wavelet Transform” (EWT) which consists in building from the segmentation of the original signal Fourier spectrum, a bank of filters. The method used is based on the idea utilized in the construction of both Littlewood-Paley and Meyer’s wavelets. The heart of the method lies in the segmentation of the Fourier spectrum based on the local maxima detection in order to obtain a set of non-overlapping segments. Because linked to the Fourier spectrum, the frequency resolution provided by EWT is higher than that provided by EMD and therefore allows to overcome the mode-mixing problem. On the other hand, if the EWT technique is able to detect the frequencies involved in the original time series fluctuations, EWT does not allow to associate the detected frequencies to a specific mode of variability as in the EMD technic. Because EMD is closer to the observation of physical phenomena than EWT, we propose here a new technic called EAWD (Empirical Adaptive Wavelet Decomposition) based on the coupling of the EMD and EWT technics by using the IMFs density spectral content to optimize the segmentation of the Fourier spectrum required by EWT. In this study, EMD and EWT technics are described, then EAWD technic is presented. Comparison of results obtained respectively by EMD, EWT and EAWD technics on time series of ozone total columns recorded at Reunion island over [1978-2019] period is discussed. This study was carried out as part of the SOLSTYCE project dedicated to the characterization and modeling of the underlying dynamics of time series issued from complex systems in atmospheric sciences

Keywords: adaptive filtering, empirical mode decomposition, empirical wavelet transform, filter banks, mode-mixing, non-linear and non-stationary time series, wavelet

Procedia PDF Downloads 115
957 Distributed Automation System Based Remote Monitoring of Power Quality Disturbance on LV Network

Authors: Emmanuel D. Buedi, K. O. Boateng, Griffith S. Klogo

Abstract:

Electrical distribution networks are prone to power quality disturbances originating from the complexity of the distribution network, mode of distribution (overhead or underground) and types of loads used by customers. Data on the types of disturbances present and frequency of occurrence is needed for economic evaluation and hence finding solution to the problem. Utility companies have resorted to using secondary power quality devices such as smart meters to help gather the required data. Even though this approach is easier to adopt, data gathered from these devices may not serve the required purpose, since the installation of these devices in the electrical network usually does not conform to available PQM placement methods. This paper presents a design of a PQM that is capable of integrating into an existing DAS infrastructure to take advantage of available placement methodologies. The monitoring component of the design is implemented and installed to monitor an existing LV network. Data from the monitor is analyzed and presented. A portion of the LV network of the Electricity Company of Ghana is modeled in MATLAB-Simulink and analyzed under various earth fault conditions. The results presented show the ability of the PQM to detect and analyze PQ disturbance such as voltage sag and overvoltage. By adopting a placement methodology and installing these nodes, utilities are assured of accurate and reliable information with respect to the quality of power delivered to consumers.

Keywords: power quality, remote monitoring, distributed automation system, economic evaluation, LV network

Procedia PDF Downloads 330
956 Sample Hospital Buildings as Modern Health Facilities in Early Republican Turkey

Authors: Mehmet Sener, Emre Kishali

Abstract:

The establishment of republic brought radical changes related to the modernization of life in early republican Turkey considering the revolutions in socio-economical, cultural and political aspects. These changes also had many influences on the formation of city planning and architectural medium that the arrangements related with health facility production had an important place amongst them. While the health services were witnessing great transformations with all its sides, socio-cultural and architectural framework of these facilities necessitated the adaption of new conceptual approaches which led to the construction new hospital buildings by the republican state with a name ‘Sample Hospital’. In this period, the state constructed sample hospitals in some cities (Adana, Ankara, Erzurum, İstanbul, Konya, Sivas and Trabzon) for the aim of being a good example for further hospitals sheltering all the characteristics of a contemporary health complex for that day. In this study, these six hospitals will firstly be elucidated considering their historical evaluations and current situations. Then, being one of the most significant modern heritages of republican history, the ways to provide the interrelationship of these complexes with the rapidly evolving current world will be discussed by proposing solutions or approaches coming from the fields of city planning, architectural preservation, engineering and architectural history together with an awareness of the socio-economic conditions, health services and architectural medium of Turkey. These hospitals are complexes composed of building ensembles which have functional relationships with each other. So, some strategies will be proposed for the preservation, renovation, and refurbishment of these complexes with an awareness of the possibility of the conflict between conservation practices and today’s health facility standards. Accordingly, the addition or removal of some elements in the complex or the suggestion of some architectural changes for the modernization of these health facilities will be investigated considering the requirements of the contemporary architectural design of health facilities. Since these hospitals are highly complex structures and have vastly changing design and construction standards, they cannot be used without adopting necessary architectural and technological interventions. So, the adaptive re-use of these buildings instead of demolition or the preservation of their overall characteristics becomes inevitable for the sustaining of these health facility heritages in Turkey. In this context, a multidisciplinary analysis will be made in this study on ‘Sample Hospital’ concept and buildings existing in Turkish modern architectural history within the framework of the adaptive reuse of these health complexes.

Keywords: adaptive re-use, conservation, early republican Turkey, sample hospital

Procedia PDF Downloads 223