Search results for: operator pool
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 736

Search results for: operator pool

256 IoT Based Agriculture Monitoring Framework for Sustainable Rice Production

Authors: Armanul Hoque Shaon, Md Baizid Mahmud, Askander Nobi, Md. Raju Ahmed, Md. Jiabul Hoque

Abstract:

In the Internet of Things (IoT), devices are linked to the internet through a wireless network, allowing them to collect and transmit data without the need for a human operator. Agriculture relies heavily on wireless sensors, which are a vital component of the Internet of Things (IoT). This kind of wireless sensor network monitors physical or environmental variables like temperatures, sound, vibration, pressure, or motion without relying on a central location or sink and collaboratively passes its data across the network to be analyzed. As the primary source of plant nutrients, the soil is critical to the agricultural industry's continued growth. We're excited about the prospect of developing an Internet of Things (IoT) solution. To arrange the network, the sink node collects groundwater levels and sends them to the Gateway, which centralizes the data and forwards it to the sensor nodes. The sink node gathers soil moisture data, transmits the mean to the Gateways, and then forwards it to the website for dissemination. The web server is in charge of storing and presenting the moisture in the soil data to the web application's users. Soil characteristics may be collected using a networked method that we developed to improve rice production. Paddy land is running out as the population of our nation grows. The success of this project will be dependent on the appropriate use of the existing land base.

Keywords: IoT based agriculture monitoring, intelligent irrigation, communicating network, rice production

Procedia PDF Downloads 152
255 The Non-Stationary BINARMA(1,1) Process with Poisson Innovations: An Application on Accident Data

Authors: Y. Sunecher, N. Mamode Khan, V. Jowaheer

Abstract:

This paper considers the modelling of a non-stationary bivariate integer-valued autoregressive moving average of order one (BINARMA(1,1)) with correlated Poisson innovations. The BINARMA(1,1) model is specified using the binomial thinning operator and by assuming that the cross-correlation between the two series is induced by the innovation terms only. Based on these assumptions, the non-stationary marginal and joint moments of the BINARMA(1,1) are derived iteratively by using some initial stationary moments. As regards to the estimation of parameters of the proposed model, the conditional maximum likelihood (CML) estimation method is derived based on thinning and convolution properties. The forecasting equations of the BINARMA(1,1) model are also derived. A simulation study is also proposed where BINARMA(1,1) count data are generated using a multivariate Poisson R code for the innovation terms. The performance of the BINARMA(1,1) model is then assessed through a simulation experiment and the mean estimates of the model parameters obtained are all efficient, based on their standard errors. The proposed model is then used to analyse a real-life accident data on the motorway in Mauritius, based on some covariates: policemen, daily patrol, speed cameras, traffic lights and roundabouts. The BINARMA(1,1) model is applied on the accident data and the CML estimates clearly indicate a significant impact of the covariates on the number of accidents on the motorway in Mauritius. The forecasting equations also provide reliable one-step ahead forecasts.

Keywords: non-stationary, BINARMA(1, 1) model, Poisson innovations, conditional maximum likelihood, CML

Procedia PDF Downloads 128
254 Evaluating and Reducing Aircraft Technical Delays and Cancellations Impact on Reliability Operational: Case Study of Airline Operator

Authors: Adel A. Ghobbar, Ahmad Bakkar

Abstract:

Although special care is given to maintenance, aircraft systems fail, and these failures cause delays and cancellations. The occurrence of Delays and Cancellations affects operators and manufacturers negatively. To reduce technical delays and cancellations, one should be able to determine the important systems causing them. The goal of this research is to find a method to define the most expensive delays and cancellations systems for Airline operators. A predictive model was introduced to forecast the failure and their impact after carrying out research that identifies relevant information to tackle the problems faced while answering the questions of this paper. Data were obtained from the manufacturers’ services reliability team database. Subsequently, delays and cancellations evaluation methods were identified. No cost estimation methods were used due to their complexity. The model was developed, and it takes into account the frequency of delays and cancellations and uses weighting factors to give an indication of the severity of their duration. The weighting factors are based on customer experience. The data Analysis approach has shown that delays and cancellations events are not seasonal and do not follow any specific trends. The use of weighting factor does have an influence on the shortlist over short periods (Monthly) but not the analyzed period of three years. Landing gear and the navigation system are among the top 3 factors causing delays and cancellations for all three aircraft types. The results did confirm that the cooperation between certain operators and manufacture reduce the impact of delays and cancellations.

Keywords: reliability, availability, delays & cancellations, aircraft maintenance

Procedia PDF Downloads 130
253 Comparison of Two Anesthetic Methods during Interventional Neuroradiology Procedure: Propofol versus Sevoflurane Using Patient State Index

Authors: Ki Hwa Lee, Eunsu Kang, Jae Hong Park

Abstract:

Background: Interventional neuroradiology (INR) has been a rapidly growing and evolving neurosurgical part during the past few decades. Sevoflurane and propofol are both suitable anesthetics for INR procedure. Monitoring of depth of anesthesia is being used very widely. SEDLine™ monitor, a 4-channel processed EEG monitor, uses a proprietary algorithm to analyze the raw EEG signal and displays the Patient State Index (PSI) values. There are only a fewer studies examining the PSI in the neuro-anesthesia. We aimed to investigate the difference of PSI values and hemodynamic variables between sevoflurane and propofol anesthesia during INR procedure. Methods: We reviewed the medical records of patients who scheduled to undergo embolization of non-ruptured intracranial aneurysm by a single operator from May 2013 to December 2014, retrospectively. Sixty-five patients were categorized into two groups; sevoflurane (n = 33) vs propofol (n = 32) group. The PSI values, hemodynamic variables, and the use of hemodynamic drugs were analyzed. Results: Significant differences were seen between PSI values obtained during different perioperative stages in both two groups (P < 0.0001). The PSI values of propofol group were lower than that of sevoflurane group during INR procedure (P < 0.01). The patients in propofol group had more prolonged time of extubation and more phenylephrine requirement than sevoflurane group (p < 0.05). Anti-hypertensive drug was more administered to the patients during extubation in sevoflurane group (p < 0.05). Conclusions: The PSI can detect depth of anesthesia and changes of concentration of anesthetics during INR procedure. Extubation was faster in sevoflurane group, but smooth recovery was shown in propofol group.

Keywords: interventional neuroradiology, patient state index, propofol, sevoflurane

Procedia PDF Downloads 179
252 Mindfulness and Mental Resilience Training for Pilots: Enhancing Cognitive Performance and Stress Management

Authors: Nargiza Nuralieva

Abstract:

The study delves into assessing the influence of mindfulness and mental resilience training on the cognitive performance and stress management of pilots. Employing a meticulous literature search across databases such as Medline and Google Scholar, the study used specific keywords to target a wide array of studies. Inclusion criteria were stringent, focusing on peer-reviewed studies in English that utilized designs like randomized controlled trials, with a specific interest in interventions related to mindfulness or mental resilience training for pilots and measured outcomes pertaining to cognitive performance and stress management. The initial literature search identified a pool of 123 articles, with subsequent screening resulting in the exclusion of 77 based on title and abstract. The remaining 54 articles underwent a more rigorous full-text screening, leading to the exclusion of 41. Additionally, five studies were selected from the World Health Organization's clinical trials database. A total of 11 articles from meta-analyses were retained for examination, underscoring the study's dedication to a meticulous and robust inclusion process. The interventions varied widely, incorporating mixed approaches, Cognitive behavioral Therapy (CBT)-based, and mindfulness-based techniques. The analysis uncovered positive effects across these interventions. Specifically, mixed interventions demonstrated a Standardized Mean Difference (SMD) of 0.54, CBT-based interventions showed an SMD of 0.29, and mindfulness-based interventions exhibited an SMD of 0.43. Long-term effects at a 6-month follow-up suggested sustained impacts for both mindfulness-based (SMD: 0.63) and CBT-based interventions (SMD: 0.73), albeit with notable heterogeneity.

Keywords: mindfulness, mental resilience, pilots, cognitive performance, stress management

Procedia PDF Downloads 54
251 The Research On The Necessity Of Launching Environmental Programs For Studies In Universities As Well As Training Specialists In This Sphere.

Authors: Anastasia V. Lazareva

Abstract:

Nowadays in the light of the evolving multifocal challenges in the sphere of environmental and social difficulties and despite the strong opposition of globalist and anti-globalist movements, we are facing the urgent need of the creation of a vast pool of educated environmentalists through the implementation of relevant university faculties and programs. Considering the threats humanity has tackled these years portrayed in every tiny detail in AGENDA – 2030 –namely, poverty, biodiversity loss, marine and terrestrial pollution, lack of sanitation, and equal rights for all, we must admit that professionals are required to address them all. With this purpose, we have conducted research based on the questionnaires of students, faculty members, and companies’ chief executives and human resources managers on what particular disciplines should be incorporated into the programs in universities and higher institutions to meet the millennium goals and tests. The research is based on the Linkert scale and covers various age groups of students. The topicality of this issue is predetermined by modern reality. The subject of the research is a questionnaire database filled in by 97 students, 17 faculty members of MGIMO University, and 14 companies’ representatives concerning their attitudes towards the implementation of environmental programs of studies in universities and the choice of disciplines required. The study has a limitation -it is based only on one university students' and faculty members’ questionnaires. The methods applied are a questionnaire, content analysis, sampling, and categorization. The findings of this survey imply that all three groups of respondents admit the necessity of implementing environmental programs for studies in higher education. Nevertheless, different groups favor various programs and disciplines to be incorporated into the curriculum.

Keywords: ecology, university studies, environmentalists, education, global challenges

Procedia PDF Downloads 16
250 An Experimental Approach to the Influence of Tipping Points and Scientific Uncertainties in the Success of International Fisheries Management

Authors: Jules Selles

Abstract:

The Atlantic and Mediterranean bluefin tuna fishery have been considered as the archetype of an overfished and mismanaged fishery. This crisis has demonstrated the role of public awareness and the importance of the interactions between science and management about scientific uncertainties. This work aims at investigating the policy making process associated with a regional fisheries management organization. We propose a contextualized computer-based experimental approach, in order to explore the effects of key factors on the cooperation process in a complex straddling stock management setting. Namely, we analyze the effects of the introduction of a socio-economic tipping point and the uncertainty surrounding the estimation of the resource level. Our approach is based on a Gordon-Schaefer bio-economic model which explicitly represents the decision making process. Each participant plays the role of a stakeholder of ICCAT and represents a coalition of fishing nations involved in the fishery and decide unilaterally a harvest policy for the coming year. The context of the experiment induces the incentives for exploitation and collaboration to achieve common sustainable harvest plans at the Atlantic bluefin tuna stock scale. Our rigorous framework allows testing how stakeholders who plan the exploitation of a fish stock (a common pool resource) respond to two kinds of effects: i) the inclusion of a drastic shift in the management constraints (beyond a socio-economic tipping point) and ii) an increasing uncertainty in the scientific estimation of the resource level.

Keywords: economic experiment, fisheries management, game theory, policy making, Atlantic Bluefin tuna

Procedia PDF Downloads 253
249 The Effect of Vitamin D Supplementation on Prostate Cancer: A Systematic Review and Meta-Analysis of Clinical Trials

Authors: Simin Shahvazi, Sepideh Soltani, Seyed Mehdi Ahmadi, Russell J. De Souza, Amin Salehi-Abargouei

Abstract:

Background and Objectives: Vitamin D has received attention for its potential to disrupt cancer processes such as attenuating cell proliferation and exacerbating differentiation and apoptosis. However, whether there exists a role for vitamin D in the treatment of prostate cancer specifically remains controversial. We systematically review the literature to assess whether supplementation with vitamin D influences PSA response and overall survival in patients with prostate cancer. Methods: We searched PubMed, Scopus, ISI Web of Science and Google scholar from inception through up to 10 September 2017 for both before-and-after and randomized trials that evaluated the effect of vitamin D supplementation on the prostate specific antigen (PSA) response rate in participants with prostate cancer. The DerSimonian and Laird, inverse-weighted random-effects model was used to pool effect estimates from the studies. Heterogeneity and potential publication bias were evaluated. Subgroup analyses were also performed. Results: Twenty-two studies (16 before-after and 6 randomized controlled trials) were found and included in meta-analysis. The analysis on controlled clinical trials revealed that PSA change from baseline [weighted mean difference (WMD) = -1.66 ng/ml, 95%CI: -0.69, 0.36, P= 0.543)], PSA response (RR=1.18, 95%CI: 0.97, 1.45, P=0.104) and mortality rate (risk ratio (RR) = 1.05, 95% CI: 0.81-1.36; P=0.713) was not significantly different between vitamin D supplementation and placebo groups. Single arm trials revealed that vitamin D supplementation had had a modest effect on PSA response rate: 19% of those enrolled had at least a 50% reduction in PSA by the end of treatment (95% CI: 7% to 31%; p=0.002). Conclusion: We found that vitamin D modestly increases the PSA response rate in single arm studies. No effect on serum PSA levels, PSA response and mortality was seen in randomized controlled clinical trials. It does not seem patients with prostate cancer benefit from vitamin D supplementation.

Keywords: mortality, prostatic neoplasms, PSA response, vitamin D

Procedia PDF Downloads 193
248 Resilience and Mindfulness as Individual Resources Building Communication Skills for Physicians

Authors: Malgorzata Sekulowicz, Krystyna Boron-Krupinska, Paulina Morga, Blazej Cieslik

Abstract:

Burnout is highly prevalent in health care employees, especially in physicians. It significantly reduces the efficiency of these employees, which can have negative consequences for both physicians and patients. Resilience and mindfulness enhancing positive emotions, leading to sustainable development and personal commitment, can have a significant impact on burnout. Therefore, the aim of this study was to determine the relationship between burnout symptoms and mindfulness and resilience among physicians. The authors conducted a cross-sectional survey study among seventy-four polish physicians. Participants filled out the following psychometric tools: the Maslach Burnout Inventory - Human Services Survey (MBI-HSS), Five Facet Mindfulness Questionnaire (FFMQ), Areas of Work-Life Survey (AWS), International Personality Item Pool (IPIP), the Resilience Assessment Scale (SPP-25) and the Mini-COPE Inventory. The relationship between burnout and resilience and mindfulness was assessed with path analysis. Analyzing the relationship between MBI-HSS components and mindfulness, a significant negative correlation of the FFMQ score with emotional exhaustion (-0.50, p < 0.05) and depersonalization (-0.43, p < 0.05) and a positive correlation with personal accomplishment (0.50, p < 0.05) was demonstrated. Analyzing resilience, a statistically significant relationship of SPP-25 with all tested components of MBI-HSS was demonstrated: emotional exhaustion (-0.54, p < 0.05), depersonalization (-0.31, p < 0.05) and personal accomplishment (0.35, p < 0.05). In the group of medical doctors, the higher the level of mindfulness and resilience, the lower the risk of burnout. Furthermore, the more frequently used active coping strategies (planning, acceptance), the lower the risk of burnout, while the use of passive, evasive strategies increases the risk of burnout. It may be worth considering implementing mindfulness intervention to effectively manage burnout symptoms in this group.

Keywords: burnout, medical doctors, mindfulness, physicians, resilience

Procedia PDF Downloads 104
247 Development of a Few-View Computed Tomographic Reconstruction Algorithm Using Multi-Directional Total Variation

Authors: Chia Jui Hsieh, Jyh Cheng Chen, Chih Wei Kuo, Ruei Teng Wang, Woei Chyn Chu

Abstract:

Compressed sensing (CS) based computed tomographic (CT) reconstruction algorithm utilizes total variation (TV) to transform CT image into sparse domain and minimizes L1-norm of sparse image for reconstruction. Different from the traditional CS based reconstruction which only calculates x-coordinate and y-coordinate TV to transform CT images into sparse domain, we propose a multi-directional TV to transform tomographic image into sparse domain for low-dose reconstruction. Our method considers all possible directions of TV calculations around a pixel, so the sparse transform for CS based reconstruction is more accurate. In 2D CT reconstruction, we use eight-directional TV to transform CT image into sparse domain. Furthermore, we also use 26-directional TV for 3D reconstruction. This multi-directional sparse transform method makes CS based reconstruction algorithm more powerful to reduce noise and increase image quality. To validate and evaluate the performance of this multi-directional sparse transform method, we use both Shepp-Logan phantom and a head phantom as the targets for reconstruction with the corresponding simulated sparse projection data (angular sampling interval is 5 deg and 6 deg, respectively). From the results, the multi-directional TV method can reconstruct images with relatively less artifacts compared with traditional CS based reconstruction algorithm which only calculates x-coordinate and y-coordinate TV. We also choose RMSE, PSNR, UQI to be the parameters for quantitative analysis. From the results of quantitative analysis, no matter which parameter is calculated, the multi-directional TV method, which we proposed, is better.

Keywords: compressed sensing (CS), low-dose CT reconstruction, total variation (TV), multi-directional gradient operator

Procedia PDF Downloads 255
246 Water Quality Calculation and Management System

Authors: H. M. B. N Jayasinghe

Abstract:

The water is found almost everywhere on Earth. Water resources contain a lot of pollution. Some diseases can be spread through the water to the living beings. So to be clean water it should undergo a number of treatments necessary to make it drinkable. So it is must to have purification technology for the wastewater. So the waste water treatment plants act a major role in these issues. When considering the procedures taken after the water treatment process was always based on manual calculations and recordings. Water purification plants may interact with lots of manual processes. It means the process taking much time consuming. So the final evaluation and chemical, biological treatment process get delayed. So to prevent those types of drawbacks there are some computerized programmable calculation and analytical techniques going to be introduced to the laboratory staff. To solve this problem automated system will be a solution in which guarantees the rational selection. A decision support system is a way to model data and make quality decisions based upon it. It is widely used in the world for the various kind of process automation. Decision support systems that just collect data and organize it effectively are usually called passive models where they do not suggest a specific decision but only reveal information. This web base system is based on global positioning data adding facility with map location. Most worth feature is SMS and E-mail alert service to inform the appropriate person on a critical issue. The technological influence to the system is HTML, MySQL, PHP, and some other web developing technologies. Current issues in the computerized water chemistry analysis are not much deep in progress. For an example the swimming pool water quality calculator. The validity of the system has been verified by test running and comparison with an existing plant data. Automated system will make the life easier in productively and qualitatively.

Keywords: automated system, wastewater, purification technology, map location

Procedia PDF Downloads 245
245 A Flute Tracking System for Monitoring the Wear of Cutting Tools in Milling Operations

Authors: Hatim Laalej, Salvador Sumohano-Verdeja, Thomas McLeay

Abstract:

Monitoring of tool wear in milling operations is essential for achieving the desired dimensional accuracy and surface finish of a machined workpiece. Although there are numerous statistical models and artificial intelligence techniques available for monitoring the wear of cutting tools, these techniques cannot pin point which cutting edge of the tool, or which insert in the case of indexable tooling, is worn or broken. Currently, the task of monitoring the wear on the tool cutting edges is carried out by the operator who performs a manual inspection, causing undesirable stoppages of machine tools and consequently resulting in costs incurred from lost productivity. The present study is concerned with the development of a flute tracking system to segment signals related to each physical flute of a cutter with three flutes used in an end milling operation. The purpose of the system is to monitor the cutting condition for individual flutes separately in order to determine their progressive wear rates and to predict imminent tool failure. The results of this study clearly show that signals associated with each flute can be effectively segmented using the proposed flute tracking system. Furthermore, the results illustrate that by segmenting the sensor signal by flutes it is possible to investigate the wear in each physical cutting edge of the cutting tool. These findings are significant in that they facilitate the online condition monitoring of a cutting tool for each specific flute without the need for operators/engineers to perform manual inspections of the tool.

Keywords: machining, milling operation, tool condition monitoring, tool wear prediction

Procedia PDF Downloads 299
244 People's Perspective on Water Commons in Trans-Boundary Water Governance: A Case Study from Nepal

Authors: Sristi Silwal

Abstract:

South Asian rivers support ecosystems and sustain well-being of thousands of riparian communities. Rivers however are also sources of conflict between countries and one of the contested issues between governments of the region. Governments have signed treaties to harness some of the rivers but their provisions have not been successful in improving the quality of life of those who depend on water as common property resources. This paper will present a case of the study of the status of the water commons along the lower command areas of Koshi, Gandka and Mahakali rivers. Nepal and India have signed treaties for development and management of these rivers in 1928, 1954 and 1966. The study investigated perceptions of the local community on climate-induced disasters, provision of the treaties such as water for irrigation, participation in decision-making and specific impact of women. It looked at how the local community coped with adversities. The study showed that the common pool resources are gradually getting degraded, flood events increasing while community blame ‘other state’ and state administration for exacerbating these ills. The level of awareness about provisions of existing treatise is poor. Ongoing approach to trans-boundary water management has taken inadequate cognizance of these realities as the dominant narrative perpetuates cooperation between the governments. The paper argues that on-going discourses on trans-boundary water development and management need to use a new metrics of taking cognizance of the condition of the commons and that of the people depended on them for sustenance. In absence of such narratives, the scale of degradation would increase making those already marginalized more vulnerable to impacts of global climate change.

Keywords: climate change vulnerability, conflict, cooperation, water commons

Procedia PDF Downloads 235
243 A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider

Authors: J. C. Cabanillas-Noris, M. I. Martínez-Hernández, I. León-Monzón

Abstract:

The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.

Keywords: AD0, ALICE, DCS, LHC

Procedia PDF Downloads 304
242 A Picture is worth a Billion Bits: Real-Time Image Reconstruction from Dense Binary Pixels

Authors: Tal Remez, Or Litany, Alex Bronstein

Abstract:

The pursuit of smaller pixel sizes at ever increasing resolution in digital image sensors is mainly driven by the stringent price and form-factor requirements of sensors and optics in the cellular phone market. Recently, Eric Fossum proposed a novel concept of an image sensor with dense sub-diffraction limit one-bit pixels (jots), which can be considered a digital emulation of silver halide photographic film. This idea has been recently embodied as the EPFL Gigavision camera. A major bottleneck in the design of such sensors is the image reconstruction process, producing a continuous high dynamic range image from oversampled binary measurements. The extreme quantization of the Poisson statistics is incompatible with the assumptions of most standard image processing and enhancement frameworks. The recently proposed maximum-likelihood (ML) approach addresses this difficulty, but suffers from image artifacts and has impractically high computational complexity. In this work, we study a variant of a sensor with binary threshold pixels and propose a reconstruction algorithm combining an ML data fitting term with a sparse synthesis prior. We also show an efficient hardware-friendly real-time approximation of this inverse operator. Promising results are shown on synthetic data as well as on HDR data emulated using multiple exposures of a regular CMOS sensor.

Keywords: binary pixels, maximum likelihood, neural networks, sparse coding

Procedia PDF Downloads 200
241 An Operators’ Real-sense-based Fire Simulation for Human Factors Validation in Nuclear Power Plants

Authors: Sa-Kil Kim, Jang-Soo Lee

Abstract:

On March 31, 1993, a severe fire accident took place in a nuclear power plant located in Narora in North India. The event involved a major fire in the turbine building of NAPS unit-1 and resulted in a total loss of power to the unit for 17 hours. In addition, there was a heavy ingress of smoke in the control room, mainly through the intake of the ventilation system, forcing the operators to vacate the control room. The Narora fire accident provides us lessons indicating that operators could lose their mind and predictable behaviors during a fire. After the Fukushima accident, which resulted from a natural disaster, unanticipated external events are also required to be prepared and controlled for the ultimate safety of nuclear power plants. From last year, our research team has developed a test and evaluation facility that can simulate external events such as an earthquake and fire based on the operators’ real-sense. As one of the results of the project, we proposed a unit real-sense-based facility that can simulate fire events in a control room for utilizing a test-bed of human factor validation. The test-bed has the operator’s workstation shape and functions to simulate fire conditions such as smoke, heat, and auditory alarms in accordance with the prepared fire scenarios. Furthermore, the test-bed can be used for the operators’ training and experience.

Keywords: human behavior in fire, human factors validation, nuclear power plants, real-sense-based fire simulation

Procedia PDF Downloads 282
240 Oral Lichen Planus a Manifestation of Grinspan's Syndrome or a Lichenoid Reaction to Medication

Authors: Sahar Iqrar, Malik Adeel Anwar, Zain Akram, Maria Noor

Abstract:

Introduction: Oral lichen planus is a chronic inflammatory condition of unknown etiology. Oral lichen planus may be related with several other diseases. Grinspan's Syndrome is characterized by a triad of oral lichen planus, hypertension, and diabetes mellitus. Other associations reported in the literature are with chronic liver disease and, with dyslipidemia. The nature of these associations is still not fully understood. Material and methods: Study was conducted in Department of Oral Medicine, Fatima Memorial Hospital College of Medicine and Dentistry, Lahore, Pakistan. A total of n=89 clinically diagnosed patients of oral lichen planus of both gender and all age groups were recruited and detailed history were recorded in the designed performs. Results: A total of n=89 patients were taken with male to female ratio of 3:8 in which 24 were male and 65 females. Mean age was 48.8 ± 13.8 years. Age range of 10-74 years was seen. Among these patients suffering from oral lichen planus, 41.6% (n=37) had a positive history for hypertension with 59.5% (n=22) of these patients were taking different medication for their condition. Whereas Diabetes Mellitus was found in 24.7% (n=22) patients with 72.7% (n=16) of these patients using the hypoglycemic drug (oral or injectable) to control their blood glucose levels. Out of these n=89 lichen planus patients 21.3% had both hypertension and diabetes mellitus (fulfilling the criteria for Grinspan's Syndrome). Out of this Grinspan's Syndrome pool 94.7% (n=19) were taking drug atleast for one of the two conditions. Conclusion: As noticed form the medical history of the patients, most of them were using hypoglycemic drugs for diabetes mellitus and beta blockers, diuretics and calcium channel blockers for hypertension. These drugs are known for lichenoid reaction. Therefore, it should be ruled out at histopathological/ immunological and molecular level whether these patients are suffering from lichen planus or lichenoid drug reaction to truly declare them as patients with Grinspan’s Syndrome.

Keywords: diabetes mellitus, grinspan's syndrome, lichenoid drug reaction, oral lichen planus

Procedia PDF Downloads 241
239 Robust Heart Rate Estimation from Multiple Cardiovascular and Non-Cardiovascular Physiological Signals Using Signal Quality Indices and Kalman Filter

Authors: Shalini Rankawat, Mansi Rankawat, Rahul Dubey, Mazad Zaveri

Abstract:

Physiological signals such as electrocardiogram (ECG) and arterial blood pressure (ABP) in the intensive care unit (ICU) are often seriously corrupted by noise, artifacts, and missing data, which lead to errors in the estimation of heart rate (HR) and incidences of false alarm from ICU monitors. Clinical support in ICU requires most reliable heart rate estimation. Cardiac activity, because of its relatively high electrical energy, may introduce artifacts in Electroencephalogram (EEG), Electrooculogram (EOG), and Electromyogram (EMG) recordings. This paper presents a robust heart rate estimation method by detection of R-peaks of ECG artifacts in EEG, EMG & EOG signals, using energy-based function and a novel Signal Quality Index (SQI) assessment technique. SQIs of physiological signals (EEG, EMG, & EOG) were obtained by correlation of nonlinear energy operator (teager energy) of these signals with either ECG or ABP signal. HR is estimated from ECG, ABP, EEG, EMG, and EOG signals from separate Kalman filter based upon individual SQIs. Data fusion of each HR estimate was then performed by weighing each estimate by the Kalman filters’ SQI modified innovations. The fused signal HR estimate is more accurate and robust than any of the individual HR estimate. This method was evaluated on MIMIC II data base of PhysioNet from bedside monitors of ICU patients. The method provides an accurate HR estimate even in the presence of noise and artifacts.

Keywords: ECG, ABP, EEG, EMG, EOG, ECG artifacts, Teager-Kaiser energy, heart rate, signal quality index, Kalman filter, data fusion

Procedia PDF Downloads 693
238 Improving Trainings of Mineral Processing Operators Through Gamification and Modelling and Simulation

Authors: Pedro A. S. Bergamo, Emilia S. Streng, Jan Rosenkranz, Yousef Ghorbani

Abstract:

Within the often-hazardous mineral industry, simulation training has speedily gained appreciation as an important method of increasing site safety and productivity through enhanced operator skill and knowledge. Performance calculations related to froth flotation, one of the most important concentration methods, is probably the hardest topic taught during the training of plant operators. Currently, most training teach those skills by traditional methods like slide presentations and hand-written exercises with a heavy focus on memorization. To optimize certain aspects of these pieces of training, we developed “MinFloat”, which teaches the operation formulas of the froth flotation process with the help of gamification. The simulation core based on a first-principles flotation model was implemented in Unity3D and an instructor tutoring system was developed, which presents didactic content and reviews the selected answers. The game was tested by 25 professionals with extensive experience in the mining industry based on a questionnaire formulated for training evaluations. According to their feedback, the game scored well in terms of quality, didactic efficacy and inspiring character. The feedback of the testers on the main target audience and the outlook of the mentioned solution is presented. This paper aims to provide technical background on the construction of educational games for the mining industry besides showing how feedback from experts can more efficiently be gathered thanks to new technologies such as online forms.

Keywords: training evaluation, simulation based training, modelling, and simulation, froth flotation

Procedia PDF Downloads 112
237 Metabolic Profiling of Populus trichocarpa Family 1 UDP-Glycosyltransferases

Authors: Patricia M. B. Saint-Vincent, Anna Furches, Stephanie Galanie, Erica Teixeira Prates, Piet Jones, Nancy Engle, David Kainer, Wellington Muchero, Daniel Jacobson, Timothy J. Tschaplinski

Abstract:

Uridine diphosphate-glycosyltransferases (UGTs) are enzymes that catalyze sugar transfer to a variety of plant metabolites. UGT substrates, which include plant secondary metabolites involved in lignification, demonstrate new activities and incorporation when glycosylated. Knowledge of UGT function, substrate specificity, and enzyme products is important for plant engineering efforts, especially related to increasing plant biomass through lignification. UGTs in Populus trichocarpa, a biofuel feedstock, and model woody plant, were selected from a pool of gene candidates using rapid prioritization strategies. A functional genomics workflow, consisting of a metabolite genome-wide association study (mGWAS), expression of synthetic codon-optimized genes, and high-throughput biochemical assays with mass spectrometry-based analysis, was developed for determining the substrates and products of previously-uncharacterized enzymes. A total of 40 UGTs from P. trichocarpa were profiled, and the biochemical assay results were compared to predicted mGWAS connections. Assay results confirmed seven of 11 leaf mGWAS associations and demonstrated varying levels of substrate specificity among candidate UGTs. P. trichocarpa UGT substrate processing confirms the role of these newly-characterized enzymes in lignan, flavonoid, and phytohormone metabolism, with potential implications for cell wall biosynthesis, nitrogen uptake, and biotic and abiotic stress responses.

Keywords: Populus, metabolite-gene associations, GWAS, bio feedstocks, glycosyltransferase

Procedia PDF Downloads 111
236 A Laboratory Study into the Effects of Surface Waves on Freestyle Swimming

Authors: Scott Draper, Nat Benjanuvatra, Grant Landers, Terry Griffiths, Justin Geldard

Abstract:

Open water swimming has been an Olympic sport since 2008 and is growing in popularity world-wide as a low impact form of exercise. Unlike pool swimming, open water swimmers experience a range of different environmental conditions, including surface waves, variable water temperature, aquatic life, and ocean currents. This presentation will describe experimental research to investigate how freestyle swimming behaviour and performance is influenced by surface waves. A group of 12 swimmers were instructed to swim freestyle in the 54 m long wave flume located at The University of Western Australia’s Coastal and Offshore Engineering Laboratory. A variety of different regular waves were simulated, varying in height (up to 0.3 m), period (1.25 – 4s), and direction (with or against the swimmer). Swimmer’s velocity and acceleration, respectively, were determined from video recording and inertial sensors attached to five different parts of the swimmer’s body. The results illustrate how the swimmers stroke rate and the wave encounter frequency influence their forward speed and how particular wave conditions can benefit or hinder performance. Comparisons to simplified mathematical models provide insight into several aspects of performance, including: (i) how much faster swimmers can travel when swimming with as opposed to against the waves, and (ii) why swimmers of lesser ability are expected to be affected proportionally more by waves than elite swimmers. These findings have implications across the spectrum from elite to ‘weekend’ swimmers, including how they are coached and their ability to win (or just successfully complete) iconic open water events such as the Rottnest Channel Swim held annually in Western Australia.

Keywords: open water, surface waves, wave height/length, wave flume, stroke rate

Procedia PDF Downloads 111
235 Bi-Criteria Vehicle Routing Problem for Possibility Environment

Authors: Bezhan Ghvaberidze

Abstract:

A multiple criteria optimization approach for the solution of the Fuzzy Vehicle Routing Problem (FVRP) is proposed. For the possibility environment the levels of movements between customers are calculated by the constructed simulation interactive algorithm. The first criterion of the bi-criteria optimization problem - minimization of the expectation of total fuzzy travel time on closed routes is constructed for the FVRP. A new, second criterion – maximization of feasibility of movement on the closed routes is constructed by the Choquet finite averaging operator. The FVRP is reduced to the bi-criteria partitioning problem for the so called “promising” routes which were selected from the all admissible closed routes. The convenient selection of the “promising” routes allows us to solve the reduced problem in the real-time computing. For the numerical solution of the bi-criteria partitioning problem the -constraint approach is used. An exact algorithm is implemented based on D. Knuth’s Dancing Links technique and the algorithm DLX. The Main objective was to present the new approach for FVRP, when there are some difficulties while moving on the roads. This approach is called FVRP for extreme conditions (FVRP-EC) on the roads. Also, the aim of this paper was to construct the solving model of the constructed FVRP. Results are illustrated on the numerical example where all Pareto-optimal solutions are found. Also, an approach for more complex model FVRP with time windows was developed. A numerical example is presented in which optimal routes are constructed for extreme conditions on the roads.

Keywords: combinatorial optimization, Fuzzy Vehicle routing problem, multiple objective programming, possibility theory

Procedia PDF Downloads 485
234 Diagnostic Accuracy of the Tuberculin Skin Test for Tuberculosis Diagnosis: Interest of Using ROC Curve and Fagan’s Nomogram

Authors: Nouira Mariem, Ben Rayana Hazem, Ennigrou Samir

Abstract:

Background and aim: During the past decade, the frequency of extrapulmonary forms of tuberculosis has increased. These forms are under-diagnosed using conventional tests. The aim of this study was to evaluate the performance of the Tuberculin Skin Test (TST) for the diagnosis of tuberculosis, using the ROC curve and Fagan’s Nomogram methodology. Methods: This was a case-control, multicenter study in 11 anti-tuberculosis centers in Tunisia, during the period from June to November2014. The cases were adults aged between 18 and 55 years with confirmed tuberculosis. Controls were free from tuberculosis. A data collection sheet was filled out and a TST was performed for each participant. Diagnostic accuracy measures of TST were estimated using ROC curve and Area Under Curve to estimate sensitivity and specificity of a determined cut-off point. Fagan’s nomogram was used to estimate its predictive values. Results: Overall, 1053 patients were enrolled, composed of 339 cases (sex-ratio (M/F)=0.87) and 714 controls (sex-ratio (M/F)=0.99). The mean age was 38.3±11.8 years for cases and 33.6±11 years for controls. The mean diameter of the TST induration was significantly higher among cases than controls (13.7mm vs.6.2mm;p=10-6). Area Under Curve was 0.789 [95% CI: 0.758-0.819; p=0.01], corresponding to a moderate discriminating power for this test. The most discriminative cut-off value of the TST, which were associated with the best sensitivity (73.7%) and specificity (76.6%) couple was about 11 mm with a Youden index of 0.503. Positive and Negative predictive values were 3.11% and 99.52%, respectively. Conclusion: In view of these results, we can conclude that the TST can be used for tuberculosis diagnosis with a good sensitivity and specificity. However, the skin induration measurement and its interpretation is operator dependent and remains difficult and subjective. The combination of the TST with another test such as the Quantiferon test would be a good alternative.

Keywords: tuberculosis, tuberculin skin test, ROC curve, cut-off

Procedia PDF Downloads 66
233 Normalized Enterprises Architectures: Portugal's Public Procurement System Application

Authors: Tiago Sampaio, André Vasconcelos, Bruno Fragoso

Abstract:

The Normalized Systems Theory, which is designed to be applied to software architectures, provides a set of theorems, elements and rules, with the purpose of enabling evolution in Information Systems, as well as ensuring that they are ready for change. In order to make that possible, this work’s solution is to apply the Normalized Systems Theory to the domain of enterprise architectures, using Archimate. This application is achieved through the adaptation of the elements of this theory, making them artifacts of the modeling language. The theorems are applied through the identification of the viewpoints to be used in the architectures, as well as the transformation of the theory’s encapsulation rules into architectural rules. This way, it is possible to create normalized enterprise architectures, thus fulfilling the needs and requirements of the business. This solution was demonstrated using the Portuguese Public Procurement System. The Portuguese government aims to make this system as fair as possible, allowing every organization to have the same business opportunities. The aim is for every economic operator to have access to all public tenders, which are published in any of the 6 existing platforms, independently of where they are registered. In order to make this possible, we applied our solution to the construction of two different architectures, which are able of fulfilling the requirements of the Portuguese government. One of those architectures, TO-BE A, has a Message Broker that performs the communication between the platforms. The other, TO-BE B, represents the scenario in which the platforms communicate with each other directly. Apart from these 2 architectures, we also represent the AS-IS architecture that demonstrates the current behavior of the Public Procurement Systems. Our evaluation is based on a comparison between the AS-IS and the TO-BE architectures, regarding the fulfillment of the rules and theorems of the Normalized Systems Theory and some quality metrics.

Keywords: archimate, architecture, broker, enterprise, evolvable systems, interoperability, normalized architectures, normalized systems, normalized systems theory, platforms

Procedia PDF Downloads 356
232 The Future of Insurance: P2P Innovation versus Traditional Business Model

Authors: Ivan Sosa Gomez

Abstract:

Digitalization has impacted the entire insurance value chain, and the growing movement towards P2P platforms and the collaborative economy is also beginning to have a significant impact. P2P insurance is defined as innovation, enabling policyholders to pool their capital, self-organize, and self-manage their own insurance. In this context, new InsurTech start-ups are emerging as peer-to-peer (P2P) providers, based on a model that differs from traditional insurance. As a result, although P2P platforms do not change the fundamental basis of insurance, they do enable potentially more efficient business models to be established in terms of ensuring the coverage of risk. It is therefore relevant to determine whether p2p innovation can have substantial effects on the future of the insurance sector. For this purpose, it is considered necessary to develop P2P innovation from a business perspective, as well as to build a comparison between a traditional model and a P2P model from an actuarial perspective. Objectives: The objectives are (1) to represent P2P innovation in the business model compared to the traditional insurance model and (2) to establish a comparison between a traditional model and a P2P model from an actuarial perspective. Methodology: The research design is defined as action research in terms of understanding and solving the problems of a collectivity linked to an environment, applying theory and best practices according to the approach. For this purpose, the study is carried out through the participatory variant, which involves the collaboration of the participants, given that in this design, participants are considered experts. For this purpose, prolonged immersion in the field is carried out as the main instrument for data collection. Finally, an actuarial model is developed relating to the calculation of premiums that allows for the establishment of projections of future scenarios and the generation of conclusions between the two models. Main Contributions: From an actuarial and business perspective, we aim to contribute by developing a comparison of the two models in the coverage of risk in order to determine whether P2P innovation can have substantial effects on the future of the insurance sector.

Keywords: Insurtech, innovation, business model, P2P, insurance

Procedia PDF Downloads 92
231 Hysteresis Modeling in Iron-Dominated Magnets Based on a Deep Neural Network Approach

Authors: Maria Amodeo, Pasquale Arpaia, Marco Buzio, Vincenzo Di Capua, Francesco Donnarumma

Abstract:

Different deep neural network architectures have been compared and tested to predict magnetic hysteresis in the context of pulsed electromagnets for experimental physics applications. Modelling quasi-static or dynamic major and especially minor hysteresis loops is one of the most challenging topics for computational magnetism. Recent attempts at mathematical prediction in this context using Preisach models could not attain better than percent-level accuracy. Hence, this work explores neural network approaches and shows that the architecture that best fits the measured magnetic field behaviour, including the effects of hysteresis and eddy currents, is the nonlinear autoregressive exogenous neural network (NARX) model. This architecture aims to achieve a relative RMSE of the order of a few 100 ppm for complex magnetic field cycling, including arbitrary sequences of pseudo-random high field and low field cycles. The NARX-based architecture is compared with the state-of-the-art, showing better performance than the classical operator-based and differential models, and is tested on a reference quadrupole magnetic lens used for CERN particle beams, chosen as a case study. The training and test datasets are a representative example of real-world magnet operation; this makes the good result obtained very promising for future applications in this context.

Keywords: deep neural network, magnetic modelling, measurement and empirical software engineering, NARX

Procedia PDF Downloads 129
230 Diplomatic Assurances in International Law

Authors: William Thomas Worster

Abstract:

Diplomatic assurances issued by states declaring that they will not mistreat individuals returned to them occupy a strange middle ground between being legal and non-legal obligations. States assert that they are non-binding, yet at other times that they are binding. However, this assertion may not be the end of the discussion. The International Court of Justice and other tribunals have concluded that similar instruments were binding, states have disagreed that certain similar instruments were binding, and the Vienna Convention on the Law of Treaties and its travaux prépératoires do not appear to contemplate non-binding instruments. This paper is a case study of diplomatic assurances but, by necessity, touches on the delicate question of whether certain texts are treaties, promises, or non-binding political statements. International law, and law in general, requires a binary approach to obligation. All communications must be binding or not, even if the fit is not precise. Through this study, we will find that some of the obligations in certain assurances can be understood as legal and some not. We will attempt to state the current methodology for determining which obligations are legal under the law of treaties and law on binding unilateral promises. The paper begins with some background of the legal environment of diplomatic assurances and their use in cases of expulsion. The paper then turns to discuss the legal nature of diplomatic assurances, proceeding to address various possibilities for legal value as treaties and as binding unilateral statements. This paper will not examine the legal value of diplomatic assurances solely under customary international law other than the way in which customary international law might further refine the treaty definition. In order to identify whether any assurances are contained in legal acts, this study identifies a pool of relevant assurances and qualitatively analyzes whether any of those are contained in treaties or binding unilateral statements. To the author’s best knowledge, this study is the first large-scale, qualitative qualitative analysis of assurances as a group of instruments that accounts for their heterogenous nature. It is also the first study to identify the indicators of whether an instrument is a treaty or promise.

Keywords: diplomatic assurances, deportation, extradition, expulsion, non-refoulement, torture, persecution, death penalty, human rights, memorandum of understanding, promises, secret, monitoring, compliance, enforcement

Procedia PDF Downloads 83
229 Predictive Analytics of Bike Sharing Rider Parameters

Authors: Bongs Lainjo

Abstract:

The evolution and escalation of bike-sharing programs (BSP) continue unabated. Since the sixties, many countries have introduced different models and strategies of BSP. These include variations ranging from dockless models to electronic real-time monitoring systems. Reasons for using this BSP include recreation, errands, work, etc. And there is all indication that complex, and more innovative rider-friendly systems are yet to be introduced. The objective of this paper is to analyze current variables established by different operators and streamline them identifying the most compelling ones using analytics. Given the contents of available databases, there is a lack of uniformity and common standard on what is required and what is not. Two factors appear to be common: user type (registered and unregistered, and duration of each trip). This article uses historical data provided by one operator based in the greater Washington, District of Columbia, USA area. Several variables including categorical and continuous data types were screened. Eight out of 18 were considered acceptable and significantly contribute to determining a useful and reliable predictive model. Bike-sharing systems have become popular in recent years all around the world. Although this trend has resulted in many studies on public cycling systems, there have been few previous studies on the factors influencing public bicycle travel behavior. A bike-sharing system is a computer-controlled system in which individuals can borrow bikes for a fee or free for a limited period. This study has identified unprecedented useful, and pragmatic parameters required in improving BSP ridership dynamics.

Keywords: sharing program, historical data, parameters, ridership dynamics, trip duration

Procedia PDF Downloads 138
228 Traditional Mechanisms of Conflict Resolution in Africa: A Pathway to Sustainable Peace in Nigeria

Authors: Ejovi Eghwubare Augustine

Abstract:

This study delved into the traditional mechanisms of conflict resolution in Africa, a pathway to sustainable peace in Nigeria. It deployed the quantitative and qualitative methods of data collection and content analysis. The work adopted the Peace Process theory propounded by John Darby and Roger Macunity. It ascertained that disputes or disagreements are unarguably and necessarily an inevitable part of human existence, flowing directly from communication, interaction, and relationships which can occur at individual and national levels, even at international levels in view of the current trend of globalization. The alternative Dispute Resolution (ADR) mechanism is a basket of procedures outside the traditional process of litigation or strict determination of legal rights. It may also be elucidated as a range of procedures that serve as generally involve the intercession and assistance of a neutral and impartial third party. The traditional mechanisms of conflict resolution in Africa are alien to the Western world; this paper is of utmost importance to the Western world and also enriched their pool of literature. Nigeria is a country that is dominated by various ethnic groups anchored on diverse cultures, customs, and traditions. It is, therefore, not surprising to see conflicts arise, and despite the various attempts at resolving these conflicts through litigation, they still remained unabated. The paper investigated the lessons learned from Traditional Mechanisms of Conflict resolution; it also interrogated its impact and the way forward. In light of the lessons that were learned and the impact of the traditional mechanisms of conflict resolution, suggestions on how to attain a sustainable, peaceful society were proffered. In conclusion, the study crystallized reforms on the alternative dispute resolution introduced through the traditional mechanism, which includes, amongst others, that constitutional recognition should be given to traditional institutions of conflict resolution to enable quick dispensation of matters.

Keywords: traditional, conflict, peace, resolution

Procedia PDF Downloads 72
227 Using Deep Learning for the Detection of Faulty RJ45 Connectors on a Radio Base Station

Authors: Djamel Fawzi Hadj Sadok, Marrone Silvério Melo Dantas Pedro Henrique Dreyer, Gabriel Fonseca Reis de Souza, Daniel Bezerra, Ricardo Souza, Silvia Lins, Judith Kelner

Abstract:

A radio base station (RBS), part of the radio access network, is a particular type of equipment that supports the connection between a wide range of cellular user devices and an operator network access infrastructure. Nowadays, most of the RBS maintenance is carried out manually, resulting in a time consuming and costly task. A suitable candidate for RBS maintenance automation is repairing faulty links between devices caused by missing or unplugged connectors. A suitable candidate for RBS maintenance automation is repairing faulty links between devices caused by missing or unplugged connectors. This paper proposes and compares two deep learning solutions to identify attached RJ45 connectors on network ports. We named connector detection, the solution based on object detection, and connector classification, the one based on object classification. With the connector detection, we get an accuracy of 0:934, mean average precision 0:903. Connector classification, get a maximum accuracy of 0:981 and an AUC of 0:989. Although connector detection was outperformed in this study, this should not be viewed as an overall result as connector detection is more flexible for scenarios where there is no precise information about the environment and the possible devices. At the same time, the connector classification requires that information to be well-defined.

Keywords: radio base station, maintenance, classification, detection, deep learning, automation

Procedia PDF Downloads 199