Search results for: uncertainties of measurement
1838 Relationship of Workplace Stress and Mental Wellbeing among Health Professionals
Authors: Rabia Mushtaq, Uroosa Javaid
Abstract:
It has been observed that health professionals are at higher danger of stress in light of the fact that being a specialist is physically and emotionally demanding. The study aimed to investigate the relationship between workplace stress and mental wellbeing among health professionals. Sample of 120 male and female health professionals belonging to two age groups, i.e., early adulthood and middle adulthood, was employed through purposive sampling technique. Job stress scale, mindful attention awareness scale, and Warwick Edinburgh mental wellbeing scales were used for the measurement of study variables. Results of the study indicated that job stress has a significant negative relationship with mental wellbeing among health professionals. The current study opened the door for more exploratory work on mindfulness among health professionals. Yielding outcomes helped in consolidating adapting procedures among workers to improve their mental wellbeing and lessen the job stress.Keywords: health professionals, job stress, mental wellbeing, mindfulness
Procedia PDF Downloads 1751837 Possible Exposure of Persons with Cardiac Pacemakers to Extremely Low Frequency (ELF) Electric and Magnetic Fields
Authors: Leena Korpinen, Rauno Pääkkönen, Fabriziomaria Gobba, Vesa Virtanen
Abstract:
The number of persons with implanted cardiac pacemakers (PM) has increased in Western countries. The aim of this paper is to investigate the possible situations where persons with a PM may be exposed to extremely low frequency (ELF) electric (EF) and magnetic fields (MF) that may disturb their PM. Based on our earlier studies, it is possible to find such high public exposure to EFs only in some places near 400 kV power lines, where an EF may disturb a PM in unipolar mode. Such EFs cannot be found near 110 kV power lines. Disturbing MFs can be found near welding machines. However, we do not have measurement data from welding. Based on literature and earlier studies at Tampere University of Technology, it is difficult to find public EF or MF exposure that is high enough to interfere with PMs.Keywords: cardiac pacemaker, electric field, magnetic field, electrical engineering
Procedia PDF Downloads 4321836 An Artificial Intelligence Framework to Forecast Air Quality
Authors: Richard Ren
Abstract:
Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms
Procedia PDF Downloads 1271835 Visual Odometry and Trajectory Reconstruction for UAVs
Authors: Sandro Bartolini, Alessandro Mecocci, Alessio Medaglini
Abstract:
The growing popularity of systems based on unmanned aerial vehicles (UAVs) is highlighting their vulnerability, particularly in relation to the positioning system used. Typically, UAV architectures use the civilian GPS, which is exposed to a number of different attacks, such as jamming or spoofing. This is why it is important to develop alternative methodologies to accurately estimate the actual UAV position without relying on GPS measurements only. In this paper, we propose a position estimate method for UAVs based on monocular visual odometry. We have developed a flight control system capable of keeping track of the entire trajectory travelled, with a reduced dependency on the availability of GPS signals. Moreover, the simplicity of the developed solution makes it applicable to a wide range of commercial drones. The final goal is to allow for safer flights in all conditions, even under cyber-attacks trying to deceive the drone.Keywords: visual odometry, autonomous uav, position measurement, autonomous outdoor flight
Procedia PDF Downloads 2191834 The Impact of Online Advertising on Consumer Purchase Behaviour Based on Malaysian Organizations
Authors: Naser Zourikalatehsamad, Seyed Abdorreza Payambarpour, Ibrahim Alwashali, Zahra Abdolkarimi
Abstract:
The paper aims to evaluate the effect of online advertising on consumer purchase behavior in Malaysian organizations. The paper has potential to extend and refine theory. A survey was distributed among Students of UTM university during the winter 2014 and 160 responses were collected. Regression analysis was used to test the hypothesized relationships of the model. Result shows that the predictors (cost saving factor, convenience factor and customized product or services) have positive impact on intention to continue seeking online advertising.Keywords: consumer purchase, convenience, customized product, cost saving, customization, flow theory, mass communication, online advertising ads, online advertising measurement, online advertising mechanism, online intelligence system, self-confidence, willingness to purchase
Procedia PDF Downloads 4811833 A Comparison Study: Infant and Children’s Clothing Size Charts in South Korea and UK
Authors: Hye-Won Lim, Tom Cassidy, Tracy Cassidy
Abstract:
Infant and children’s body shapes are changing constantly while they are growing up into adults and are also distinctive physically between countries. For this reason, optimum size charts which can represent body sizes and shapes of infants and children are required. In this study, investigations of current size charts in South Korea and UK (n=50 each) were conducted for understanding and figuring out the sizing perspectives of the clothing manufacturers. The size charts of the two countries were collected randomly from online shopping websites and those size charts’ average measurements were compared with both national sizing surveys (SizeKorea and Shape GB). The size charts were also classified by age, gender, clothing type, fitting, and other factors. In addition, the key measurement body parts of size charts of each country were determined and those will be suggested for new size charts and sizing system development.Keywords: infant clothing, children’s clothing, body shapes, size charts
Procedia PDF Downloads 3171832 Measurement of VIP Edge Conduction Using Vacuum Guarded Hot Plate
Authors: Bongsu Choi, Tae-Ho Song
Abstract:
Vacuum insulation panel (VIP) is a promising thermal insulator for buildings, refrigerator, LNG carrier and so on. In general, it has the thermal conductivity of 2~4 mW/m•K. However, this thermal conductivity is that measured at the center of VIP. The total effective thermal conductivity of VIP is larger than this value due to the edge conduction through the envelope. In this paper, the edge conduction of VIP is examined theoretically, numerically and experimentally. To confirm the existence of the edge conduction, numerical analysis is performed for simple two-dimensional VIP model and a theoretical model is proposed to calculate the edge conductivity. Also, the edge conductivity is measured using the vacuum guarded hot plate and the experiment is validated against numerical analysis. The results show that the edge conductivity is dependent on the width of panel and thickness of Al-foil. To reduce the edge conduction, it is recommended that the VIP should be made as big as possible or made of thin Al film envelope.Keywords: envelope, edge conduction, thermal conductivity, vacuum insulation panel
Procedia PDF Downloads 4051831 Analysis of Transformer by Gas and Moisture Sensor during Laboratory Time Monitoring
Authors: Miroslav Gutten, Daniel Korenciak, Milan Simko, Milan Chupac
Abstract:
Ensure the reliable and correct function of transformers is the main essence of on-line non-destructive diagnostic tool, which allows the accurately track of the status parameters. Devices for on-line diagnostics are very costly. However, there are devices, whose price is relatively low and when used correctly, they can be executed a complex diagnostics. One of these devices is sensor HYDRAN M2, which is used to detect the moisture and gas content in the insulation oil. Using the sensor HYDRAN M2 in combination with temperature, load measurement, and physicochemical analysis can be made the economically inexpensive diagnostic system, which use is not restricted to distribution transformers. This system was tested in educational laboratory environment at measured oil transformer 22/0.4 kV. From the conclusions referred in article is possible to determine, which kind of fault was occurred in the transformer and how was an impact on the temperature, evolution of gases and water content.Keywords: transformer, diagnostics, gas and moisture sensor, monitoring
Procedia PDF Downloads 3851830 Guests’ Satisfaction and Intention to Revisit Smart Hotels: Qualitative Interviews Approach
Authors: Raymond Chi Fai Si Tou, Jacey Ja Young Choe, Amy Siu Ian So
Abstract:
Smart hotels can be defined as the hotel which has an intelligent system, through digitalization and networking which achieve hotel management and service information. In addition, smart hotels include high-end designs that integrate information and communication technology with hotel management fulfilling the guests’ needs and improving the quality, efficiency and satisfaction of hotel management. The purpose of this study is to identify appropriate factors that may influence guests’ satisfaction and intention to revisit Smart Hotels based on service quality measurement of lodging quality index and extended UTAUT theory. Unified Theory of Acceptance and Use of Technology (UTAUT) is adopted as a framework to explain technology acceptance and use. Since smart hotels are technology-based infrastructure hotels, UTATU theory could be as the theoretical background to examine the guests’ acceptance and use after staying in smart hotels. The UTAUT identifies four key drivers of the adoption of information systems: performance expectancy, effort expectancy, social influence, and facilitating conditions. The extended UTAUT modifies the definitions of the seven constructs for consideration; the four previously cited constructs of the UTAUT model together with three new additional constructs, which including hedonic motivation, price value and habit. Thus, the seven constructs from the extended UTAUT theory could be adopted to understand their intention to revisit smart hotels. The service quality model will also be adopted and integrated into the framework to understand the guests’ intention of smart hotels. There are rare studies to examine the service quality on guests’ satisfaction and intention to revisit in smart hotels. In this study, Lodging Quality Index (LQI) will be adopted to measure the service quality in smart hotels. Using integrated UTAUT theory and service quality model because technological applications and services require using more than one model to understand the complicated situation for customers’ acceptance of new technology. Moreover, an integrated model could provide more perspective insights to explain the relationships of the constructs that could not be obtained from only one model. For this research, ten in-depth interviews are planned to recruit this study. In order to confirm the applicability of the proposed framework and gain an overview of the guest experience of smart hotels from the hospitality industry, in-depth interviews with the hotel guests and industry practitioners will be accomplished. In terms of the theoretical contribution, it predicts that the integrated models from the UTAUT theory and the service quality will provide new insights to understand factors that influence the guests’ satisfaction and intention to revisit smart hotels. After this study identifies influential factors, smart hotel practitioners could understand which factors may significantly influence smart hotel guests’ satisfaction and intention to revisit. In addition, smart hotel practitioners could also provide outstanding guests experience by improving their service quality based on the identified dimensions from the service quality measurement. Thus, it will be beneficial to the sustainability of the smart hotels business.Keywords: intention to revisit, guest satisfaction, qualitative interviews, smart hotels
Procedia PDF Downloads 2081829 Identifying Unknown Dynamic Forces Applied on Two Dimensional Frames
Authors: H. Katkhuda
Abstract:
A time domain approach is used in this paper to identify unknown dynamic forces applied on two dimensional frames using the measured dynamic structural responses for a sub-structure in the two dimensional frame. In this paper a sub-structure finite element model with short length of measurement from only three or four accelerometers is required, and an iterative least-square algorithm is used to identify the unknown dynamic force applied on the structure. Validity of the method is demonstrated with numerical examples using noise-free and noise-contaminated structural responses. Both harmonic and impulsive forces are studied. The results show that the proposed approach can identify unknown dynamic forces within very limited iterations with high accuracy and shows its robustness even noise- polluted dynamic response measurements are utilized.Keywords: dynamic force identification, dynamic responses, sub-structure, time domain
Procedia PDF Downloads 3611828 Performance Prediction Methodology of Slow Aging Assets
Authors: M. Ben Slimene, M.-S. Ouali
Abstract:
Asset management of urban infrastructures faces a multitude of challenges that need to be overcome to obtain a reliable measurement of performances. Predicting the performance of slowly aging systems is one of those challenges, which helps the asset manager to investigate specific failure modes and to undertake the appropriate maintenance and rehabilitation interventions to avoid catastrophic failures as well as to optimize the maintenance costs. This article presents a methodology for modeling the deterioration of slowly degrading assets based on an operating history. It consists of extracting degradation profiles by grouping together assets that exhibit similar degradation sequences using an unsupervised classification technique derived from artificial intelligence. The obtained clusters are used to build the performance prediction models. This methodology is applied to a sample of a stormwater drainage culvert dataset.Keywords: artificial Intelligence, clustering, culvert, regression model, slow degradation
Procedia PDF Downloads 1121827 Influential Parameters in Estimating Soil Properties from Cone Penetrating Test: An Artificial Neural Network Study
Authors: Ahmed G. Mahgoub, Dahlia H. Hafez, Mostafa A. Abu Kiefa
Abstract:
The Cone Penetration Test (CPT) is a common in-situ test which generally investigates a much greater volume of soil more quickly than possible from sampling and laboratory tests. Therefore, it has the potential to realize both cost savings and assessment of soil properties rapidly and continuously. The principle objective of this paper is to demonstrate the feasibility and efficiency of using artificial neural networks (ANNs) to predict the soil angle of internal friction (Φ) and the soil modulus of elasticity (E) from CPT results considering the uncertainties and non-linearities of the soil. In addition, ANNs are used to study the influence of different parameters and recommend which parameters should be included as input parameters to improve the prediction. Neural networks discover relationships in the input data sets through the iterative presentation of the data and intrinsic mapping characteristics of neural topologies. General Regression Neural Network (GRNN) is one of the powerful neural network architectures which is utilized in this study. A large amount of field and experimental data including CPT results, plate load tests, direct shear box, grain size distribution and calculated data of overburden pressure was obtained from a large project in the United Arab Emirates. This data was used for the training and the validation of the neural network. A comparison was made between the obtained results from the ANN's approach, and some common traditional correlations that predict Φ and E from CPT results with respect to the actual results of the collected data. The results show that the ANN is a very powerful tool. Very good agreement was obtained between estimated results from ANN and actual measured results with comparison to other correlations available in the literature. The study recommends some easily available parameters that should be included in the estimation of the soil properties to improve the prediction models. It is shown that the use of friction ration in the estimation of Φ and the use of fines content in the estimation of E considerable improve the prediction models.Keywords: angle of internal friction, cone penetrating test, general regression neural network, soil modulus of elasticity
Procedia PDF Downloads 4151826 Internet, Fake News, and Democracy: The Case of Kosovo
Authors: Agrinë Baraku
Abstract:
This paper focuses on the convergence of the internet, fake news, and democracy. This paper will examine the convergence of these concepts, the tenets of democracy which are affected by the ever-increasing exposure to fake news, and whether the impact strengthens or can further weaken countries with fragile democracies. To demonstrate the convergence and the impact and to further the discussion about this topic, the case of Kosovo is explored. Its position in the Western Balkans makes it even more susceptible to the pressure stemming from geopolitical interests, which intersect with the generation of fake news by different international actors. Domestically, through data generated by Kantar (Index) Kosova Longitudinal Study on Media Measurement Survey (MMS), which focused on media viewership, the trend among Kosovar citizens is traced and then inserted into a bigger landscape, which is compounded by tenuous circumstances and challenges that Kosovo faces. Attention will be paid to what this can tell about where Kosovo currently is and the possibilities of what can be done regarding the phenomenon that is taking place.Keywords: democracy, disinformation, internet, social media, fake news
Procedia PDF Downloads 891825 Dynamic Reliability for a Complex System and Process: Application on Offshore Platform in Mozambique
Authors: Raed KOUTA, José-Alcebiades-Ernesto HLUNGUANE, Eric Châtele
Abstract:
The search for and exploitation of new fossil energy resources is taking place in the context of the gradual depletion of existing deposits. Despite the adoption of international targets to combat global warming, the demand for fuels continues to grow, contradicting the movement towards an energy-efficient society. The increase in the share of offshore in global hydrocarbon production tends to compensate for the depletion of terrestrial reserves, thus constituting a major challenge for the players in the sector. Through the economic potential it represents, and the energy independence it provides, offshore exploitation is also a challenge for States such as Mozambique, which have large maritime areas and whose environmental wealth must be considered. The exploitation of new reserves on economically viable terms depends on available technologies. The development of deep and ultra-deep offshore requires significant research and development efforts. Progress has also been made in managing the multiple risks inherent in this activity. Our study proposes a reliability approach to develop products and processes designed to live at sea. Indeed, the context of an offshore platform requires highly reliable solutions to overcome the difficulties of access to the system for regular maintenance and quick repairs and which must resist deterioration and degradation processes. One of the characteristics of failures that we consider is the actual conditions of use that are considered 'extreme.' These conditions depend on time and the interactions between the different causes. These are the two factors that give the degradation process its dynamic character, hence the need to develop dynamic reliability models. Our work highlights mathematical models that can explicitly manage interactions between components and process variables. These models are accompanied by numerical resolution methods that help to structure a dynamic reliability approach in a physical and probabilistic context. The application developed makes it possible to evaluate the reliability, availability, and maintainability of a floating storage and unloading platform for liquefied natural gas production.Keywords: dynamic reliability, offshore plateform, stochastic process, uncertainties
Procedia PDF Downloads 1201824 Unconventional Dating of Old Peepal Tree of Chandigarh (India) Using Optically Stimulated Luminescence
Authors: Rita Rani, Ramesh Kumar
Abstract:
The intend of the current study is to date an old grand Peepal tree that is still alive. The tree is situated in Kalibard village, Sector 9, Chandigarh (India). Due to its huge structure, it has got the status of ‘Heritage tree.’ Optically Stimulated Luminescence of sediments beneath the roots is used to determine the age of the tree. Optical dating is preferred over conventional dating methods due to more precession. The methodology includes OSL of quartz grain using SAR protocol for accumulated dose measurement. The age determination of an alive tree using sedimentary quartz is in close agreement with the approximated age provided by the related agency. This is the first attempt at using optically stimulated luminescence in the age determination of alive trees in this region. The study concludes that the Luminescence dating of alive trees is the nondestructive and more precise method.Keywords: luminescence, dose rate, optical dating, sediments
Procedia PDF Downloads 1761823 Features for Measuring Credibility on Facebook Information
Authors: Kanda Runapongsa Saikaew, Chaluemwut Noyunsan
Abstract:
Nowadays social media information, such as news, links, images, or VDOs, is shared extensively. However, the effectiveness of disseminating information through social media lacks in quality: less fact checking, more biases, and several rumors. Many researchers have investigated about credibility on Twitter, but there is no the research report about credibility information on Facebook. This paper proposes features for measuring credibility on Facebook information. We developed the system for credibility on Facebook. First, we have developed FB credibility evaluator for measuring credibility of each post by manual human’s labelling. We then collected the training data for creating a model using Support Vector Machine (SVM). Secondly, we developed a chrome extension of FB credibility for Facebook users to evaluate the credibility of each post. Based on the usage analysis of our FB credibility chrome extension, about 81% of users’ responses agree with suggested credibility automatically computed by the proposed system.Keywords: facebook, social media, credibility measurement, internet
Procedia PDF Downloads 3561822 Numerical Study of Fiber Bragg Grating Sensor: Longitudinal and Transverse Detection of Temperature and Strain
Authors: K. Khelil, H. Ammar, K. Saouchi
Abstract:
Fiber Bragg Grating (FBG) structure is an periodically modulated optical fiber. It acts as a selective filter of wavelength whose reflected peak is called Bragg wavelength and it depends on the period of the fiber and the refractive index. The simulation of FBG is based on solving the Coupled Mode Theory equation by using the Transfer Matrix Method which is carried out using MATLAB. It is found that spectral reflectivity is shifted when the change of temperature and strain is uniform. Under non-uniform temperature or strain perturbation, the spectrum is both shifted and destroyed. In case of transverse loading, reflectivity spectrum is split into two peaks, the first is specific to X axis, and the second belongs to Y axis. FBGs are used in civil engineering to detect perturbations applied to buildings.Keywords: Bragg wavelength, coupled mode theory, optical fiber, temperature measurement
Procedia PDF Downloads 4941821 Breakdown Voltage Measurement of High Voltage Transformers Oils Using an Active Microwave Resonator Sensor
Authors: Ahmed A. Al-Mudhafar, Ali A. Abduljabar, Hayder Jawad Albattat
Abstract:
This work suggests a new microwave resonator sensor (MRS) device for measuring the oil’s breakdown voltage of high voltage transformers. A precise high-sensitivity sensor is designed and manufactured based on a microstrip split ring resonator (SRR). To improve the sensor sensitivity, a RF amplifier of 30 dB gain is linked through a transmission line of 50Ω.The sensor operates at a microwave band (L) with a quality factor of 1.35x105 when it is loaded with an empty tube. In this work, the sensor has been tested with three samples of high voltage transformer oil of different ages (new, middle, and damaged) where the quality factor differs with each sample. A mathematical model was built to calculate the breakdown voltage of the transformer oils and the accuracy of the results was higher than 90%.Keywords: active resonator sensor, oil breakdown voltage, transformers oils, quality factor
Procedia PDF Downloads 2691820 Hydrothermal Synthesis of Mesoporous Carbon Nanospheres and Their Electrochemical Properties for Glucose Detection
Authors: Ali Akbar Kazemi Asl, Mansour Rahsepar
Abstract:
Mesoporous carbon nanospheres (MCNs) with uniform particle size distribution having an average of 290 nm and large specific surface area (274.4 m²/g) were synthesized by a one-step hydrothermal method followed by the calcination process and then utilized as an enzyme-free glucose biosensor. Morphology, crystal structure, and porous nature of the synthesized nanospheres were characterized by scanning electron microscopy (SEM), X-Ray diffraction (XRD), and Brunauer–Emmett–Teller (BET) analysis, respectively. Also, the electrochemical performance of the MCNs@GCE electrode for the measurement of glucose concentration in alkaline media was investigated by electrochemical impedance spectroscopy (EIS), cyclic voltammetry (CV), and chronoamperometry (CA). MCNs@GCE electrode shows good sensing performance, including a rapid glucose oxidation response within 3.1 s, a wide linear range of 0.026-12 mM, a sensitivity of 212.34 μA.mM⁻¹.cm⁻², and a detection limit of 25.7 μM with excellent selectivity.Keywords: biosensor, electrochemical, glucose, mesoporous carbon, non-enzymatic
Procedia PDF Downloads 1901819 Quality of Service of Transportation Networks: A Hybrid Measurement of Travel Time and Reliability
Authors: Chin-Chia Jane
Abstract:
In a transportation network, travel time refers to the transmission time from source node to destination node, whereas reliability refers to the probability of a successful connection from source node to destination node. With an increasing emphasis on quality of service (QoS), both performance indexes are significant in the design and analysis of transportation systems. In this work, we extend the well-known flow network model for transportation networks so that travel time and reliability are integrated into the QoS measurement simultaneously. In the extended model, in addition to the general arc capacities, each intermediate node has a time weight which is the travel time for per unit of commodity going through the node. Meanwhile, arcs and nodes are treated as binary random variables that switch between operation and failure with associated probabilities. For pre-specified travel time limitation and demand requirement, the QoS of a transportation network is the probability that source can successfully transport the demand requirement to destination while the total transmission time is under the travel time limitation. This work is pioneering, since existing literatures that evaluate travel time reliability via a single optimization path, the proposed QoS focuses the performance of the whole network system. To compute the QoS of transportation networks, we first transfer the extended network model into an equivalent min-cost max-flow network model. In the transferred network, each arc has a new travel time weight which takes value 0. Each intermediate node is replaced by two nodes u and v, and an arc directed from u to v. The newly generated nodes u and v are perfect nodes. The new direct arc has three weights: travel time, capacity, and operation probability. Then the universal set of state vectors is recursively decomposed into disjoint subsets of reliable, unreliable, and stochastic vectors until no stochastic vector is left. The decomposition is made possible by applying existing efficient min-cost max-flow algorithm. Because the reliable subsets are disjoint, QoS can be obtained directly by summing the probabilities of these reliable subsets. Computational experiments are conducted on a benchmark network which has 11 nodes and 21 arcs. Five travel time limitations and five demand requirements are set to compute the QoS value. To make a comparison, we test the exhaustive complete enumeration method. Computational results reveal the proposed algorithm is much more efficient than the complete enumeration method. In this work, a transportation network is analyzed by an extended flow network model where each arc has a fixed capacity, each intermediate node has a time weight, and both arcs and nodes are independent binary random variables. The quality of service of the transportation network is an integration of customer demands, travel time, and the probability of connection. We present a decomposition algorithm to compute the QoS efficiently. Computational experiments conducted on a prototype network show that the proposed algorithm is superior to existing complete enumeration methods.Keywords: quality of service, reliability, transportation network, travel time
Procedia PDF Downloads 2211818 Stress Intensity Factor for Dynamic Cracking of Composite Material by X-FEM Method
Authors: S. Lecheb, A. Nour, A. Chellil, H. Mechakra, N. Hamad, H. Kebir
Abstract:
The work involves develops attended by a numerical execution of the eXtend Finite Element Method premises a measurement by the fracture process cracked so many cracked plates an application will be processed for the calculation of the stress intensity factor SIF. In the first we give in statically part the distribution of stress, displacement field and strain of composite plate in two cases uncrack/edge crack, also in dynamical part the first six modes shape. Secondly, we calculate Stress Intensity Factor SIF for different orientation angle θ of central crack with length (2a=0.4mm) in plan strain condition, KI and KII are obtained for mode I and mode II respectively using X-FEM method. Finally from crack inclined involving mixed modes results, the comparison we chose dangerous inclination and the best crack angle when K is minimal.Keywords: stress intensity factor (SIF), crack orientation, glass/epoxy, natural frequencies, X-FEM
Procedia PDF Downloads 5151817 Characterization of current–voltage (I–V) and capacitance–voltage–frequency (C–V–f) features of Au/GaN Schottky diodes
Authors: Abdelaziz Rabehi
Abstract:
The current–voltage (I–V) characteristics of Au/GaN Schottky diodes were measured at room temperature. In addition, capacitance–voltage–frequency (C–V–f) characteristics are investigated by considering the interface states (Nss) at frequency range 100 kHz to 1 MHz. From the I–V characteristics of the Schottky diode, ideality factor (n) and barrier height (Φb) values of 1.22 and 0.56 eV, respectively, were obtained from a forward bias I–V plot. In addition, the interface states distribution profile as a function of (Ess − Ev) was extracted from the forward bias I–V measurements by taking into account the bias dependence of the effective barrier height (Φe) for the Schottky diode. The C–V curves gave a barrier height value higher than those obtained from I–V measurements. This discrepancy is due to the different nature of the I–V and C–V measurement techniques.Keywords: Schottky diodes, frequency dependence, barrier height, interface states
Procedia PDF Downloads 3021816 Efficiency Measurement of Indian Sugar Manufacturing Firms - a DEA Approach
Authors: Amit Kumar Dwivedi, Priyanko Ghosh
Abstract:
Data Envelopment analysis (DEA) has been used to calculate the technical and scale efficiency measures of the public and private sugar manufacturing firms of the Indian Sugar Industry (2006 to 2010). Within DEA framework, the input & Output oriented Variable Returns to Scale (VRS) & Constant Return to Scale (CRS) model is employed for the study of Decision making units (DMUs). A representative sample of 43 firms which account for major portion of the total market share is studied. The selection criterion for the inclusion of a firm in the analysis was the total sales of INR 5,000 million or more in the year 2010. After reviewing the literature it is found that no study has been conducted in the context of Indian sugar manufacturing firms in the Post-liberalization era which motivates us to initiate the study.Keywords: technical efficiency, Indian sugar manufacturing units, DEA, input output oriented
Procedia PDF Downloads 5421815 Identifying Organizational Culture to Implement Knowledge Management: Case Study of BKN, Indonesia
Authors: Maria Margaretha, Elin Cahyaningsih, Dana Indra Sensuse Lukman
Abstract:
One of key success an organization can be seen from its culture. Employee, environment, and so on are factors for organization to achieve goals and build a competitive advantage. Type of organizational culture can be a guide to implementing Knowledge Management (KM) in organization especially in BKN. Culture will determine behavior of employees or environment to support KM. This paper describes the process to decide which culture does organization belong and suggestion and creating strategic moves in the future to implement KM. OCAI (Organizational Culture Assessment Instrument) and its framework (Competing Value Framework) were used to decide the type of organizational culture. To implement KM in organization, clan is an appropriate culture, because clan culture represent cultural values and leader type to implement a successful KM. Result of the measurement will be references for BKN to improve organization culture to achieve its goals and organization effectiveness.Keywords: organizational culture, government, knowledge management, OCAI
Procedia PDF Downloads 6211814 Estimation of Natural Convection Heat Transfer from Plate-Fin Heat Sinks in a Closed Enclosure
Authors: Han-Taw Chen, Chung-Hou Lai, Tzu-Hsiang Lin, Ge-Jang He
Abstract:
This study applies the inverse method and three-dimensional CFD commercial software in conjunction with the experimental temperature data to investigate the heat transfer and fluid flow characteristics of the plate-fin heat sink in a closed rectangular enclosure for various values of fin height. The inverse method with the finite difference method and the experimental temperature data is applied to determine the heat transfer coefficient. The k-ε turbulence model is used to obtain the heat transfer and fluid flow characteristics within the fins. To validate the accuracy of the results obtained, the comparison of the average heat transfer coefficient is made. The calculated temperature at selected measurement locations on the plate-fin is also compared with experimental data.Keywords: inverse method, FLUENT, k-ε model, heat transfer characteristics, plate-fin heat sink
Procedia PDF Downloads 4601813 Recombination Center Levels in Gold and Platinum Doped N-Type Silicon
Authors: Nam Chol Yu, Kyong Il Chu
Abstract:
Using DLTS measurement techniques, we determined the dominant recombination center levels (defects of both A and B) in gold and platinum doped n-type silicon. Also, the injection and temperature dependence of the Shockley-Read-Hall (SRH) carrier lifetime was studied under low-level injection and high-level injection. Here measurements show that the dominant level under low-level injection located at EC-0.25eV(A) correlated to the Pt+G1 and the dominant level under high-level injection located at EC-0.54eV(B) correlated to the Au+G4. Finally, A and B are the same dominant levels for controlling the lifetime in gold-platinum doped n-silicon.Keywords: recombination center level, lifetime, carrier lifetime control, gold, platinum, silicon
Procedia PDF Downloads 1561812 Harmonic Distortion Caused by Electric Bus Battery Charger in Alexandria Distribution System
Authors: Mohamed Elhosieny Aly Ismail
Abstract:
The paper illustrates the total voltage and current harmonic distortion impact caused by fast-charging an electric bus and maintaining standard limit compliance. Measuring the current harmonic level in the range of 2 kHz-9 kHz. Also, the impact of the total demand distortions current caused by fast charger electric bus on the utility by measuring at the point of common coupling and comparing the measurement with IEEE519 -2014 standard. The results show that the total harmonic current distortion for the charger is within the limits of IEC 61000-3-12 and the fifth harmonic current was the most dominant frequency then the seventh harmonic current. The harmonic current in the range of 2 kHz- 9 kHz shows the frequency 5.1kHz is the most dominant frequency.Keywords: electric vehicle, total harmonic distortion, IEEE519-2014, IEC 61000-3-12, super harmonic distortion
Procedia PDF Downloads 1011811 A Fabrication Method for PEDOT: PSS Based Humidity Sensor
Authors: Nazia Tarannum, M. Ayaz Ahmad
Abstract:
The main goal of this article is to report some interesting features for the fabrication/design of PEDOT:PSS based humidity sensor. Here first we fabricated humidity sensor and then studied its electro-mechanical characteristics. In general the humidity plays an important role in various private and government sectors all over the world. Monitoring and controlling the humidity is a great task for the reliable operation of various systems. The PEDOT:PSS is very much promising humidity sensor and also is fabricated by performing various analyses. The interdigited electrode (IDE) has channel length 200 microns prepared by lithography. Lithography of IDE was done on PPR coated glass substrate using negative mask and exposing it with UV light for 10 secs via DSA. During the above said fabrication, we have taken account for the following steps: •Plasma ashing of IDE •Spincoating of PEDOT:PSS was done @3000 rpm on IDE substrace •Baked the substrace at 130 °C up to time limit 15 mins. •Resistance measurement using Labtracer 2.9 software via Keithley 2400source meter.Keywords: fabrication method, PEDOT:PSS material, humidity sensor, electro-mechanical
Procedia PDF Downloads 3501810 Designing Automated Embedded Assessment to Assess Student Learning in a 3D Educational Video Game
Authors: Mehmet Oren, Susan Pedersen, Sevket C. Cetin
Abstract:
Despite the frequently criticized disadvantages of the traditional used paper and pencil assessment, it is the most frequently used method in our schools. Although assessments do an acceptable measurement, they are not capable of measuring all the aspects and the richness of learning and knowledge. Also, many assessments used in schools decontextualize the assessment from the learning, and they focus on learners’ standing on a particular topic but do not concentrate on how student learning changes over time. For these reasons, many scholars advocate that using simulations and games (S&G) as a tool for assessment has significant potentials to overcome the problems in traditionally used methods. S&G can benefit from the change in technology and provide a contextualized medium for assessment and teaching. Furthermore, S&G can serve as an instructional tool rather than a method to test students’ learning at a particular time point. To investigate the potentials of using educational games as an assessment and teaching tool, this study presents the implementation and the validation of an automated embedded assessment (AEA), which can constantly monitor student learning in the game and assess their performance without intervening their learning. The experiment was conducted on an undergraduate level engineering course (Digital Circuit Design) with 99 participant students over a period of five weeks in Spring 2016 school semester. The purpose of this research study is to examine if the proposed method of AEA is valid to assess student learning in a 3D Educational game and present the implementation steps. To address this question, this study inspects three aspects of the AEA for the validation. First, the evidence-centered design model was used to lay out the design and measurement steps of the assessment. Then, a confirmatory factor analysis was conducted to test if the assessment can measure the targeted latent constructs. Finally, the scores of the assessment were compared with an external measure (a validated test measuring student learning on digital circuit design) to evaluate the convergent validity of the assessment. The results of the confirmatory factor analysis showed that the fit of the model with three latent factors with one higher order factor was acceptable (RMSEA < 0.00, CFI =1, TLI=1.013, WRMR=0.390). All of the observed variables significantly loaded to the latent factors in the latent factor model. In the second analysis, a multiple regression analysis was used to test if the external measure significantly predicts students’ performance in the game. The results of the regression indicated the two predictors explained 36.3% of the variance (R2=.36, F(2,96)=27.42.56, p<.00). It was found that students’ posttest scores significantly predicted game performance (β = .60, p < .000). The statistical results of the analyses show that the AEA can distinctly measure three major components of the digital circuit design course. It was aimed that this study can help researchers understand how to design an AEA, and showcase an implementation by providing an example methodology to validate this type of assessment.Keywords: educational video games, automated embedded assessment, assessment validation, game-based assessment, assessment design
Procedia PDF Downloads 4211809 Assessing the Structure of Non-Verbal Semantic Knowledge: The Evaluation and First Results of the Hungarian Semantic Association Test
Authors: Alinka Molnár-Tóth, Tímea Tánczos, Regina Barna, Katalin Jakab, Péter Klivényi
Abstract:
Supported by neuroscientific findings, the so-called Hub-and-Spoke model of the human semantic system is based on two subcomponents of semantic cognition, namely the semantic control process and semantic representation. Our semantic knowledge is multimodal in nature, as the knowledge system stored in relation to a conception is extensive and broad, while different aspects of the conception may be relevant depending on the purpose. The motivation of our research is to develop a new diagnostic measurement procedure based on the preservation of semantic representation, which is appropriate to the specificities of the Hungarian language and which can be used to compare the non-verbal semantic knowledge of healthy and aphasic persons. The development of the test will broaden the Hungarian clinical diagnostic toolkit, which will allow for more specific therapy planning. The sample of healthy persons (n=480) was determined by the last census data for the representativeness of the sample. Based on the concept of the Pyramids and Palm Tree Test, and according to the characteristics of the Hungarian language, we have elaborated a test based on different types of semantic information, in which the subjects are presented with three pictures: they have to choose the one that best fits the target word above from the two lower options, based on the semantic relation defined. We have measured 5 types of semantic knowledge representations: associative relations, taxonomy, motional representations, concrete as well as abstract verbs. As the first step in our data analysis, we examined the normal distribution of our results, and since it was not normally distributed (p < 0.05), we used nonparametric statistics further into the analysis. Using descriptive statistics, we could determine the frequency of the correct and incorrect responses, and with this knowledge, we could later adjust and remove the items of questionable reliability. The reliability was tested using Cronbach’s α, and it can be safely said that all the results were in an acceptable range of reliability (α = 0.6-0.8). We then tested for the potential gender differences using the Mann Whitney-U test, however, we found no difference between the two (p < 0.05). Likewise, we didn’t see that the age had any effect on the results using one-way ANOVA (p < 0.05), however, the level of education did influence the results (p > 0.05). The relationships between the subtests were observed by the nonparametric Spearman’s rho correlation matrix, showing statistically significant correlation between the subtests (p > 0.05), signifying a linear relationship between the measured semantic functions. A margin of error of 5% was used in all cases. The research will contribute to the expansion of the clinical diagnostic toolkit and will be relevant for the individualised therapeutic design of treatment procedures. The use of a non-verbal test procedure will allow an early assessment of the most severe language conditions, which is a priority in the differential diagnosis. The measurement of reaction time is expected to advance prodrome research, as the tests can be easily conducted in the subclinical phase.Keywords: communication disorders, diagnostic toolkit, neurorehabilitation, semantic knowlegde
Procedia PDF Downloads 103