Search results for: Secure Real Time Protocol
6406 Curing Time Effect on Behavior of Cement Treated Marine Clay
Authors: H. W. Xiao, F. H. Lee
Abstract:
Cement stabilization has been widely used for improving the strength and stiffness of soft clayey soils. Cement treated soil specimens used to investigate the stress-strain behaviour in the laboratory study are usually cured for 7 days. This paper examines the effects of curing time on the strength and stress strain behaviour of cement treated marine clay under triaxial loading condition. Laboratory-prepared cement treated Singapore marine clay with different mix proportion S-C-W (soil solid-cement solid-water) and curing time (7 days to 180 days) was investigated through conducting unconfined compressive strength test and triaxial test. The results show that the curing time has a significant effect on the unconfined compressive strength u q , isotropic compression behaviour and stress strain behaviour. Although the primary yield loci of the cement treated soil specimens with the same mix proportion expand with curing time, they are very narrowly banded and have nearly the same shape after being normalized by isotropic compression primary stress ' py p . The isotropic compression primary yield stress ' py p was shown to be linearly related to unconfined compressive strength u q for specimens with different curing time and mix proportion. The effect of curing time on the hardening behaviour will diminish with consolidation stress higher than isotropic compression primary yield stress but its damping rate is dependent on the cement content.Keywords: Cement treated soil, curing time effect, hardening behaviour, isotropic compression primary yield stress, unconfined compressive strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39166405 Artificial Neural Network Model Based Setup Period Estimation for Polymer Cutting
Authors: Zsolt János Viharos, Krisztián Balázs Kis, Imre Paniti, Gábor Belső, Péter Németh, János Farkas
Abstract:
The paper presents the results and industrial applications in the production setup period estimation based on industrial data inherited from the field of polymer cutting. The literature of polymer cutting is very limited considering the number of publications. The first polymer cutting machine is known since the second half of the 20th century; however, the production of polymer parts with this kind of technology is still a challenging research topic. The products of the applying industrial partner must met high technical requirements, as they are used in medical, measurement instrumentation and painting industry branches. Typically, 20% of these parts are new work, which means every five years almost the entire product portfolio is replaced in their low series manufacturing environment. Consequently, it requires a flexible production system, where the estimation of the frequent setup periods' lengths is one of the key success factors. In the investigation, several (input) parameters have been studied and grouped to create an adequate training information set for an artificial neural network as a base for the estimation of the individual setup periods. In the first group, product information is collected such as the product name and number of items. The second group contains material data like material type and colour. In the third group, surface quality and tolerance information are collected including the finest surface and tightest (or narrowest) tolerance. The fourth group contains the setup data like machine type and work shift. One source of these parameters is the Manufacturing Execution System (MES) but some data were also collected from Computer Aided Design (CAD) drawings. The number of the applied tools is one of the key factors on which the industrial partners’ estimations were based previously. The artificial neural network model was trained on several thousands of real industrial data. The mean estimation accuracy of the setup periods' lengths was improved by 30%, and in the same time the deviation of the prognosis was also improved by 50%. Furthermore, an investigation on the mentioned parameter groups considering the manufacturing order was also researched. The paper also highlights the manufacturing introduction experiences and further improvements of the proposed methods, both on the shop floor and on the quotation preparation fields. Every week more than 100 real industrial setup events are given and the related data are collected.Keywords: Artificial neural network, low series manufacturing, polymer cutting, setup period estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9796404 Optimization of CO2 Emissions and Cost for Composite Building Design with NSGA-II
Authors: Ji Hyeong Park, Ji Hye Jeon, Hyo Seon Park
Abstract:
Environmental pollution problems have been globally main concern in all fields including economy, society and culture into the 21st century. Beginning with the Kyoto Protocol, the reduction on the emissions of greenhouse gas such as CO2 and SOX has been a principal challenge of our day. As most buildings unlike durable goods in other industries have a characteristic and long life cycle, they consume energy in quantity and emit much CO2. Thus, for green building construction, more research is needed to reduce the CO2 emissions at each stage in the life cycle. However, recent studies are focused on the use and maintenance phase. Also, there is a lack of research on the initial design stage, especially the structure design. Therefore, in this study, we propose an optimal design plan considering CO2 emissions and cost in composite buildings simultaneously by applying to the structural design of actual building.Keywords: Multi-objective optimization, CO2 emissions, structural cost, encased composite structure
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21716403 Experimental Evaluation of Mobility Anchor Point Selection Scheme in Hierarchical Mobile IPv6
Authors: Zulkeflee Kusin, Mohamad Shanudin Zakaria
Abstract:
Hierarchical Mobile IPv6 (HMIPv6) was designed to support IP micro-mobility management in the Next Generation Networks (NGN) framework. The main design behind this protocol is the usage of Mobility Anchor Point (MAP) located at any level router of network to support hierarchical mobility management. However, the distance MAP selection in HMIPv6 causes MAP overloaded and increase frequent binding update as the network grows. Therefore, to address the issue in designing MAP selection scheme, we propose a dynamic load control mechanism integrates with a speed detection mechanism (DMS-DLC). From the experimental results we obtain that the proposed scheme gives better distribution in MAP load and increase handover speed.Keywords: Dynamic load control, HMIPv6, Mobility AnchorPoint, MAP selection scheme
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18036402 Designing of Multi-Agent Rescue Robot: Development and Basic Experiments of Master-Slave Type Rescue Robots
Authors: J. Lin, T. C. Kuo, C. -Y. Gau, K. C. Liu, Y. J. Huang, J. D. Yu, Y. W. Lin
Abstract:
A multi-agent type robot for disaster response in calamity scene is proposed in this paper. The proposed grouped rescue robots can perform cooperative reconnaissance and surveillance to achieve a given rescue mission. The multi-agent rescue of dual set robot consists of one master set and three slave units. The research for this rescue robot system is going to detect at harmful environment where human is unreachable, such as the building is infected with virus or the factory has hazardous liquid in effluent. As a dual set robot, with Bluetooth and communication network, the master set can connect with slave units and send information back to computer by wireless and monitor. Therefore, rescuer can be informed the real-time information in a calamity area. Furthermore, each slave robot is able to obstacle avoidance by ultrasonic sensors, and encodes distance and location by compass. The master robot can integrate every devices information to increase the efficiency of prospected and research unknown area.
Keywords: Designing of multi-agent rescue robot, development and basic experiments of master-slave type rescue robots.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15556401 Basic Calibration and Normalization Techniques for Time Domain Reflectometry Measurements
Authors: Shagufta Tabassum
Abstract:
The study of dielectric properties in a binary mixture of liquids is very useful to understand the liquid structure, molecular interaction, dynamics, and kinematics of the mixture. Time-domain reflectometry (TDR) is a powerful tool for studying the cooperation and molecular dynamics of the H-bonded system. Here we discuss the basic calibration and normalization procedure for TDR measurements. Our aim is to explain different types of error occur during TDR measurements and how to minimize it.
Keywords: time domain reflectometry measurement technique, cable and connector loss, oscilloscope loss, normalization technique
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5126400 Parallelization of Ensemble Kalman Filter (EnKF) for Oil Reservoirs with Time-lapse Seismic Data
Authors: Md Khairullah, Hai-Xiang Lin, Remus G. Hanea, Arnold W. Heemink
Abstract:
In this paper we describe the design and implementation of a parallel algorithm for data assimilation with ensemble Kalman filter (EnKF) for oil reservoir history matching problem. The use of large number of observations from time-lapse seismic leads to a large turnaround time for the analysis step, in addition to the time consuming simulations of the realizations. For efficient parallelization it is important to consider parallel computation at the analysis step. Our experiments show that parallelization of the analysis step in addition to the forecast step has good scalability, exploiting the same set of resources with some additional efforts.
Keywords: EnKF, Data assimilation, Parallel computing, Parallel efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22856399 Travel Time Evaluation of an Innovative U-Turn Facility on Urban Arterial Roadways
Authors: Ali Pirdavani, Tom Brijs, Tom Bellemans, Geert Wets, Koen Vanhoof
Abstract:
Signalized intersections on high-volume arterials are often congested during peak hours, causing a decrease in through movement efficiency on the arterial. Much of the vehicle delay incurred at conventional intersections is caused by high left-turn demand. Unconventional intersection designs attempt to reduce intersection delay and travel time by rerouting left-turns away from the main intersection and replacing it with right-turn followed by Uturn. The proposed new type of U-turn intersection is geometrically designed with a raised island which provides a protected U-turn movement. In this study several scenarios based on different distances between U-turn and main intersection, traffic volume of major/minor approaches and percentage of left-turn volumes were simulated by use of AIMSUN, a type of traffic microsimulation software. Subsequently some models are proposed in order to compute travel time of each movement. Eventually by correlating these equations to some in-field collected data of some implemented U-turn facilities, the reliability of the proposed models are approved. With these models it would be possible to calculate travel time of each movement under any kind of geometric and traffic condition. By comparing travel time of a conventional signalized intersection with U-turn intersection travel time, it would be possible to decide on converting signalized intersections into this new kind of U-turn facility or not. However comparison of travel time is not part of the scope of this research. In this paper only travel time of this innovative U-turn facility would be predicted. According to some before and after study about the traffic performance of some executed U-turn facilities, it is found that commonly, this new type of U-turn facility produces lower travel time. Thus, evaluation of using this type of unconventional intersection should be seriously considered.Keywords: Innovative U-turn facility, Microsimulation, Traveltime, Unconventional intersection design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13456398 Super Resolution Blind Reconstruction of Low Resolution Images using Wavelets based Fusion
Authors: Liyakathunisa, V. K. Ananthashayana
Abstract:
Crucial information barely visible to the human eye is often embedded in a series of low resolution images taken of the same scene. Super resolution reconstruction is the process of combining several low resolution images into a single higher resolution image. The ideal algorithm should be fast, and should add sharpness and details, both at edges and in regions without adding artifacts. In this paper we propose a super resolution blind reconstruction technique for linearly degraded images. In our proposed technique the algorithm is divided into three parts an image registration, wavelets based fusion and an image restoration. In this paper three low resolution images are considered which may sub pixels shifted, rotated, blurred or noisy, the sub pixel shifted images are registered using affine transformation model; A wavelet based fusion is performed and the noise is removed using soft thresolding. Our proposed technique reduces blocking artifacts and also smoothens the edges and it is also able to restore high frequency details in an image. Our technique is efficient and computationally fast having clear perspective of real time implementation.Keywords: Affine Transforms, Denoiseing, DWT, Fusion, Image registration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26736397 Voltage Problem Location Classification Using Performance of Least Squares Support Vector Machine LS-SVM and Learning Vector Quantization LVQ
Authors: Khaled Abduesslam. M, Mohammed Ali, Basher H Alsdai, Muhammad Nizam, Inayati
Abstract:
This paper presents the voltage problem location classification using performance of Least Squares Support Vector Machine (LS-SVM) and Learning Vector Quantization (LVQ) in electrical power system for proper voltage problem location implemented by IEEE 39 bus New- England. The data was collected from the time domain simulation by using Power System Analysis Toolbox (PSAT). Outputs from simulation data such as voltage, phase angle, real power and reactive power were taken as input to estimate voltage stability at particular buses based on Power Transfer Stability Index (PTSI).The simulation data was carried out on the IEEE 39 bus test system by considering load bus increased on the system. To verify of the proposed LS-SVM its performance was compared to Learning Vector Quantization (LVQ). The results showed that LS-SVM is faster and better as compared to LVQ. The results also demonstrated that the LS-SVM was estimated by 0% misclassification whereas LVQ had 7.69% misclassification.
Keywords: IEEE 39 bus, Least Squares Support Vector Machine, Learning Vector Quantization, Voltage Collapse.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24096396 Application of Model Free Adaptive Control in Main Steam Temperature System of Thermal Power Plant
Authors: Khaing Yadana Swe, Lillie Dewan
Abstract:
At present, the cascade PID control is widely used to control the superheating temperature (main steam temperature). As Main Steam Temperature has the characteristics of large inertia, large time-delay and time varying, etc., conventional PID control strategy cannot achieve good control performance. In order to overcome the bad performance and deficiencies of main steam temperature control system, Model Free Adaptive Control (MFAC) - P cascade control system is proposed in this paper. By substituting MFAC in PID of the main control loop of the main steam temperature control, it can overcome time delays, non-linearity, disturbance and time variation.
Keywords: Model free Adaptive Control, Cascade Control, Adaptive Control, PID.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28066395 Humanoid Personalized Avatar Through Multiple Natural Language Processing
Authors: Jin Hou, Xia Wang, Fang Xu, Viet Dung Nguyen, Ling Wu
Abstract:
There has been a growing interest in implementing humanoid avatars in networked virtual environment. However, most existing avatar communication systems do not take avatars- social backgrounds into consideration. This paper proposes a novel humanoid avatar animation system to represent personalities and facial emotions of avatars based on culture, profession, mood, age, taste, and so forth. We extract semantic keywords from the input text through natural language processing, and then the animations of personalized avatars are retrieved and displayed according to the order of the keywords. Our primary work is focused on giving avatars runtime instruction from multiple natural languages. Experiments with Chinese, Japanese and English input based on the prototype show that interactive avatar animations can be displayed in real time and be made available online. This system provides a more natural and interesting means of human communication, and therefore is expected to be used for cross-cultural communication, multiuser online games, and other entertainment applications.
Keywords: personalized avatar, mutiple natural luanguage processing, social backgrounds, anmimation, human computer interaction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19756394 Interdisciplinarity: A Pedagogical Practice in the Classrooms
Abstract:
The world is changing and, consequently, the young people need to acquire more sophisticated tools and skills to lead with the new societies’ challenges. In the curriculum of the Portuguese education system, in the profile of students leaving compulsory education, the critical thinking and creative thinking are pointed out as skills to be developed, as well as the capacity of interconnect different knowledge and applicate them in different contexts and learning areas. Unlike primary school teachers, teachers specialized in a specific area sometimes reveal more difficulties in developing interdisciplinary approaches in the classrooms and, despite the effort, the interdisciplinarity is not a common practice in schools. Statements like "Mathematics is everywhere" are unquestionable, however, some math teachers continue to develop an abstract teaching of mathematics devoid of any connection with reality. Good mathematical problems in real contexts are promising in the development of interdisciplinary pedagogical practices. However, these problems are often addressed by teachers in multidisciplinary rather than interdisciplinary contexts or are not addressed at all due several reasons, which range from insecurity in working on disciplinary domains with which they are not comfortable to a lack of pedagogical resources. In this study this issue is approached through a case study involving Mathematics teachers, which, in their professional development scope, attended a training aimed at stimulating interdisciplinary practices in real contexts, namely related to the COVID-19 pandemic.
Keywords: Interdisciplinarity, Mathematics, professional development, teacher training.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1816393 An Optimized Multi-block Method for Turbulent Flows
Authors: M. Goodarzi, P. Lashgari
Abstract:
A major part of the flow field involves no complicated turbulent behavior in many turbulent flows. In this research work, in order to reduce required memory and CPU time, the flow field was decomposed into several blocks, each block including its special turbulence. A two dimensional backward facing step was considered here. Four combinations of the Prandtl mixing length and standard k- E models were implemented as well. Computer memory and CPU time consumption in addition to numerical convergence and accuracy of the obtained results were mainly investigated. Observations showed that, a suitable combination of turbulence models in different blocks led to the results with the same accuracy as the high order turbulence model for all of the blocks, in addition to the reductions in memory and CPU time consumption.Keywords: Computer memory, CPU time, Multi-block method, Turbulence modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15676392 Stochastic Model Predictive Control for Linear Discrete-Time Systems with Random Dither Quantization
Authors: Tomoaki Hashimoto
Abstract:
Recently, feedback control systems using random dither quantizers have been proposed for linear discrete-time systems. However, the constraints imposed on state and control variables have not yet been taken into account for the design of feedback control systems with random dither quantization. Model predictive control is a kind of optimal feedback control in which control performance over a finite future is optimized with a performance index that has a moving initial and terminal time. An important advantage of model predictive control is its ability to handle constraints imposed on state and control variables. Based on the model predictive control approach, the objective of this paper is to present a control method that satisfies probabilistic state constraints for linear discrete-time feedback control systems with random dither quantization. In other words, this paper provides a method for solving the optimal control problems subject to probabilistic state constraints for linear discrete-time feedback control systems with random dither quantization.Keywords: Optimal control, stochastic systems, discrete-time systems, probabilistic constraints, random dither quantization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11596391 Time Series Forecasting Using Various Deep Learning Models
Authors: Jimeng Shi, Mahek Jain, Giri Narasimhan
Abstract:
Time Series Forecasting (TSF) is used to predict the target variables at a future time point based on the learning from previous time points. To keep the problem tractable, learning methods use data from a fixed length window in the past as an explicit input. In this paper, we study how the performance of predictive models change as a function of different look-back window sizes and different amounts of time to predict into the future. We also consider the performance of the recent attention-based transformer models, which had good success in the image processing and natural language processing domains. In all, we compare four different deep learning methods (Recurrent Neural Network (RNN), Long Short-term Memory (LSTM), Gated Recurrent Units (GRU), and Transformer) along with a baseline method. The dataset (hourly) we used is the Beijing Air Quality Dataset from the website of University of California, Irvine (UCI), which includes a multivariate time series of many factors measured on an hourly basis for a period of 5 years (2010-14). For each model, we also report on the relationship between the performance and the look-back window sizes and the number of predicted time points into the future. Our experiments suggest that Transformer models have the best performance with the lowest Mean Absolute Errors (MAE = 14.599, 23.273) and Root Mean Square Errors (RSME = 23.573, 38.131) for most of our single-step and multi-steps predictions. The best size for the look-back window to predict 1 hour into the future appears to be one day, while 2 or 4 days perform the best to predict 3 hours into the future.
Keywords: Air quality prediction, deep learning algorithms, time series forecasting, look-back window.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11836390 Topology Influence on TCP Congestion Control Performance in Multi-hop Ad Hoc Wireless
Authors: Haniza N., Md Khambari, M. N, Shahrin S., Adib M.Monzer Habbal, Suhaidi Hassan
Abstract:
Wireless ad hoc nodes are freely and dynamically self-organize in communicating with others. Each node can act as host or router. However it actually depends on the capability of nodes in terms of its current power level, signal strength, number of hops, routing protocol, interference and others. In this research, a study was conducted to observe the effect of hops count over different network topologies that contribute to TCP Congestion Control performance degradation. To achieve this objective, a simulation using NS-2 with different topologies have been evaluated. The comparative analysis has been discussed based on standard observation metrics: throughput, delay and packet loss ratio. As a result, there is a relationship between types of topology and hops counts towards the performance of ad hoc network. In future, the extension study will be carried out to investigate the effect of different error rate and background traffic over same topologies.Keywords: NS-2, network topology, network performance, multi-hops
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15776389 Application of a SubIval Numerical Solver for Fractional Circuits
Authors: Marcin Sowa
Abstract:
The paper discusses the subinterval-based numerical method for fractional derivative computations. It is now referred to by its acronym – SubIval. The basis of the method is briefly recalled. The ability of the method to be applied in time stepping solvers is discussed. The possibility of implementing a time step size adaptive solver is also mentioned. The solver is tested on a transient circuit example. In order to display the accuracy of the solver – the results have been compared with those obtained by means of a semi-analytical method called gcdAlpha. The time step size adaptive solver applying SubIval has been proven to be very accurate as the results are very close to the referential solution. The solver is currently able to solve FDE (fractional differential equations) with various derivative orders for each equation and any type of source time functions.Keywords: Numerical method, SubIval, fractional calculus, numerical solver, circuit analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6826388 Performance Analysis of Multiuser Diversity in Multiuser Two-Hop Decode-and-Forward Cooperative Multi-Relay Wireless Networks
Authors: Mamoun F. Al-Mistarihi, Rami Mohaisen
Abstract:
Cooperative diversity (CD) has been adopted in many communication systems because it helps in improving performance of the wireless communication systems with the help of the relays that emulate the multiple antenna terminals. This work aims to provide the derivation of the performance analysis expressions of the multiuser diversity (MUD) in the two-hop cooperative multi-relay wireless networks (TCMRNs). Considering the work analysis, we provide analytically the derivation of a closed form expression of the two most commonly used performance metrics namely, the outage probability and the symbol error probability (SEP) for the fixed decode-and-forward (FDF) protocol with MUD.
Keywords: Cooperative diversity (CD), fixed decode-andforward(FDF), multiuser diversity (MUD) , two - hop cooperative multi-relay wireless networks (TCMRN).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15436387 Removing Ocular Artifacts from EEG Signals using Adaptive Filtering and ARMAX Modeling
Authors: Parisa Shooshtari, Gelareh Mohamadi, Behnam Molaee Ardekani, Mohammad Bagher Shamsollahi
Abstract:
EEG signal is one of the oldest measures of brain activity that has been used vastly for clinical diagnoses and biomedical researches. However, EEG signals are highly contaminated with various artifacts, both from the subject and from equipment interferences. Among these various kinds of artifacts, ocular noise is the most important one. Since many applications such as BCI require online and real-time processing of EEG signal, it is ideal if the removal of artifacts is performed in an online fashion. Recently, some methods for online ocular artifact removing have been proposed. One of these methods is ARMAX modeling of EEG signal. This method assumes that the recorded EEG signal is a combination of EOG artifacts and the background EEG. Then the background EEG is estimated via estimation of ARMAX parameters. The other recently proposed method is based on adaptive filtering. This method uses EOG signal as the reference input and subtracts EOG artifacts from recorded EEG signals. In this paper we investigate the efficiency of each method for removing of EOG artifacts. A comparison is made between these two methods. Our undertaken conclusion from this comparison is that adaptive filtering method has better results compared with the results achieved by ARMAX modeling.Keywords: Ocular Artifacts, EEG, Adaptive Filtering, ARMAX
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19096386 Risk Assessment in Durations and Costs for Construction of Industrial Facilities in Egypt Using Equations and Computer
Authors: M. Kamal Elbokl, Negadi Kheira
Abstract:
Risk Evaluation is an important step in protecting your workers and your business, as well as complying with the law. It helps you focus on the risks that really matter in your workplace – the ones with the potential to cause real harm. We are in this paper introduce basics of risk assessment then we mention some of ways to risk evaluation by computer especially Monte Carlo simulation and Microsoft project.
We use Program Evaluation and Review Technique (PERT) to deal with Risks in Industrial Facilities in Evaluation and Assessment for this risk. Using PERT Technique in Microsoft Project by the PERT toolbar and using PERTMASTER Program with Primavera Program we evaluate many hazards and make calculations for that by mathematical equation to make right decisions. We define and calculate risk factor and risk severity to ranking the type of the risk then dealing with it using in that many ways like probability computation, curves, and tables. By introducing variables in the equation of functions in computer programs we calculate the risk in the time and the cost in general case and then mention some examples in industrial facilities field.
Keywords: Risk, Industrial Facilities, PERT, Monte Carlo Simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19576385 An Efficient Digital Baseband ASIC for Wireless Biomedical Signals Monitoring
Authors: Kah-Hyong Chang, Xin Liu, Jia Hao Cheong, Saisundar Sankaranarayanan, Dexing Pang, Hongzhao Zheng
Abstract:
A digital baseband Application-Specific Integrated Circuit (ASIC) (yclic Redundancy Checkis developed for a microchip transponder to transmit signals and temperature levels from biomedical monitoring devices. The transmission protocol is adapted from the ISO/IEC 11784/85 standard. The module has a decimation filter that employs only a single adder-subtractor in its datapath. The filtered output is coded with cyclic redundancy check and transmitted through backscattering Load Shift Keying (LSK) modulation to a reader. Fabricated using the 0.18-μm CMOS technology, the module occupies 0.116 mm2 in chip area (digital baseband: 0.060 mm2, decimation filter: 0.056 mm2), and consumes a total of less than 0.9 μW of power (digital baseband: 0.75 μW, decimation filter: 0.14 μW).Keywords: Biomedical sensor, decimation filter, Radio Frequency Integrated Circuit (RFIC) baseband, temperature sensor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16226384 Energy Fields as Alternative Cures for Viral Diseases
Authors: S. Amirhassan Monadjemi, Narges Zarrabi, Naser Neamatbakhsh
Abstract:
As days go by, we hear more and more about HIV, Ebola, Bird Flu and other dreadful viruses which were unknown a few decades ago. In both detecting and fighting viral diseases ordinary methods have come across some basic and important difficulties. Vaccination is by a sense introduction of the virus to the immune system before the occurrence of the real case infection. It is very successful against some viruses (e.g. Poliomyelitis), while totally ineffective against some others (e.g. HIV or Hepatitis-C). On the other hand, Anti-virus drugs are mostly some tools to control and not to cure a viral disease. This could be a good motivation to try alternative treatments. In this study, some key features of possible physical-based alternative treatments for viral diseases are presented. Electrification of body parts or fluids (especially blood) with micro electric signals with adjusted current or frequency is also studied. The main approach of this study is to find a suitable energy field, with appropriate parameters that are able to kill or deactivate viruses. This would be a lengthy, multi-disciplinary research which needs the contribution of virology, physics, and signal processing experts. It should be mentioned that all the claims made by alternative cures researchers must be tested carefully and are not advisable at the time being.Keywords: Alternative Cure, Viral disease, HIV, signals, energy filed.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19556383 Advanced Travel Information System in Heterogeneous Networks
Authors: Hsu-Yung Cheng, Victor Gau, Chih-Wei Huang, Jenq-Neng Hwang, Chih-Chang Yu
Abstract:
In order to achieve better road utilization and traffic efficiency, there is an urgent need for a travel information delivery mechanism to assist the drivers in making better decisions in the emerging intelligent transportation system applications. In this paper, we propose a relayed multicast scheme under heterogeneous networks for this purpose. In the proposed system, travel information consisting of summarized traffic conditions, important events, real-time traffic videos, and local information service contents is formed into layers and multicasted through an integration of WiMAX infrastructure and Vehicular Ad hoc Networks (VANET). By the support of adaptive modulation and coding in WiMAX, the radio resources can be optimally allocated when performing multicast so as to dynamically adjust the number of data layers received by the users. In addition to multicast supported by WiMAX, a knowledge propagation and information relay scheme by VANET is designed. The experimental results validate the feasibility and effectiveness of the proposed scheme.Keywords: Intelligent Transportation Systems, RelayedMulticast, WiMAX, Vehicular Ad hoc Networks (VANET).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17216382 A Hybrid Multi Objective Algorithm for Flexible Job Shop Scheduling
Authors: Parviz Fattahi
Abstract:
Scheduling for the flexible job shop is very important in both fields of production management and combinatorial optimization. However, it quit difficult to achieve an optimal solution to this problem with traditional optimization approaches owing to the high computational complexity. The combining of several optimization criteria induces additional complexity and new problems. In this paper, a Pareto approach to solve the multi objective flexible job shop scheduling problems is proposed. The objectives considered are to minimize the overall completion time (makespan) and total weighted tardiness (TWT). An effective simulated annealing algorithm based on the proposed approach is presented to solve multi objective flexible job shop scheduling problem. An external memory of non-dominated solutions is considered to save and update the non-dominated solutions during the solution process. Numerical examples are used to evaluate and study the performance of the proposed algorithm. The proposed algorithm can be applied easily in real factory conditions and for large size problems. It should thus be useful to both practitioners and researchers.Keywords: Flexible job shop, Scheduling, Hierarchical approach, simulated annealing, tabu search, multi objective.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20136381 Scale Time Offset Robust Modulation (STORM) in a Code Division Multiaccess Environment
Authors: David M. Jenkins Jr.
Abstract:
Scale Time Offset Robust Modulation (STORM) [1]– [3] is a high bandwidth waveform design that adds time-scale to embedded reference modulations using only time-delay [4]. In an environment where each user has a specific delay and scale, identification of the user with the highest signal power and that user-s phase is facilitated by the STORM processor. Both of these parameters are required in an efficient multiuser detection algorithm. In this paper, the STORM modulation approach is evaluated with a direct sequence spread quadrature phase shift keying (DS-QPSK) system. A misconception of the STORM time scale modulation is that a fine temporal resolution is required at the receiver. STORM will be applied to a QPSK code division multiaccess (CDMA) system by modifying the spreading codes. Specifically, the in-phase code will use a typical spreading code, and the quadrature code will use a time-delayed and time-scaled version of the in-phase code. Subsequently, the same temporal resolution in the receiver is required before and after the application of STORM. In this paper, the bit error performance of STORM in a synchronous CDMA system is evaluated and compared to theory, and the bit error performance of STORM incorporated in a single user WCDMA downlink is presented to demonstrate the applicability of STORM in a modern communication system.Keywords: Pseudonoise coded communication, Cyclic codes, Code division multiaccess
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16356380 Service Blueprint for Improving Clinical Guideline Adherence via Mobile Health Technology
Authors: Y. O’Connor, C. Heavin, S. O’ Connor, J. Gallagher, J. Wu, J. O’Donoghue
Abstract:
Background: To improve the delivery of paediatric healthcare in low resource settings, Community Health Workers (CHW) have been provided with a paper-based set of protocols known as Community Case Management (CCM). Yet research has shown that CHW adherence to CCM guidelines is poor, ultimately impacting health service delivery. Digitising the CCM guidelines via mobile technology is argued in extant literature to improve CHW adherence. However, little research exist which outlines how (a) this process can be digitised and (b) adherence could be improved as a result. Aim: To explore how an electronic mobile version of CCM (eCCM) can overcome issues associated with the paper-based CCM protocol (inadequate adherence to guidelines) vis-à-vis service blueprinting. This service blueprint will outline how (a) the CCM process can be digitised using mobile Clinical Decision Support Systems software to support clinical decision-making and (b) adherence can be improved as a result. Method: Development of a single service blueprint for a standalone application which visually depicts the service processes (eCCM) when supporting the CHWs, using an application known as Supporting LIFE (SL eCCM app) as an exemplar. Results: A service blueprint is developed which illustrates how the SL eCCM app can be utilised by CHWs to assist with the delivery of healthcare services to children. Leveraging smartphone technologies can (a) provide CHWs with just-in-time data to assist with their decision making at the point-of-care and (b) improve CHW adherence to CCM guidelines. Conclusions: The development of the eCCM opens up opportunities for the CHWs to leverage the inherent benefit of mobile devices to assist them with health service delivery in rural settings. To ensure that benefits are achieved, it is imperative to comprehend the functionality and form of the eCCM service process. By creating such a service blueprint for an eCCM approach, CHWs are provided with a clear picture regarding the role of the eCCM solution, often resulting in buy-in from the end-users.Keywords: Adherence, community health workers, developing countries, mobile clinical decision support systems, CDSS, service blueprint.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27246379 Comparison between Different Classifications of Periodontal Diseases and Their Advantages
Authors: Ilma Robo, Saimir Heta, Merilda Tarja, Sonila Kapaj, Eduart Kapaj, Geriona Lasku
Abstract:
The classification of periodontal diseases has changed significantly in favor of simplifying the protocol of diagnosis and periodontal treatment. This review study aims to highlight the latest publications in the new periodontal disease classification, talking about the most significant differences versus the old classification with the tendency to express the advantages or disadvantages of clinical application. The aim of the study also includes the growing tendency to link the way of classification of periodontal diseases with predetermined protocols of periodontal treatment of the diagnoses included in the classification. The new classification of periodontal diseases is rather comprehensive in its subdivisions, as the disease is viewed in its entirety, with the biological dimensions of the disease, the degree of aggravation and progression of the disease, in relation to risk factors, predisposition to patient susceptibility and impact of periodontal disease to the general health status of the patient.
Keywords: Periodontal diseases, clinical application, periodontal treatment, oral diagnosis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6066378 Novel Rao-Blackwellized Particle Filter for Mobile Robot SLAM Using Monocular Vision
Authors: Maohai Li, Bingrong Hong, Zesu Cai, Ronghua Luo
Abstract:
This paper presents the novel Rao-Blackwellised particle filter (RBPF) for mobile robot simultaneous localization and mapping (SLAM) using monocular vision. The particle filter is combined with unscented Kalman filter (UKF) to extending the path posterior by sampling new poses that integrate the current observation which drastically reduces the uncertainty about the robot pose. The landmark position estimation and update is also implemented through UKF. Furthermore, the number of resampling steps is determined adaptively, which seriously reduces the particle depletion problem, and introducing the evolution strategies (ES) for avoiding particle impoverishment. The 3D natural point landmarks are structured with matching Scale Invariant Feature Transform (SIFT) feature pairs. The matching for multi-dimension SIFT features is implemented with a KD-Tree in the time cost of O(log2 N). Experiment results on real robot in our indoor environment show the advantages of our methods over previous approaches.Keywords: Mobile robot, simultaneous localization and mapping, Rao-Blackwellised particle filter, evolution strategies, scale invariant feature transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21506377 Performance Analysis of Flooding Attack Prevention Algorithm in MANETs
Authors: Revathi Venkataraman, M. Pushpalatha, T. Rama Rao
Abstract:
The lack of any centralized infrastructure in mobile ad hoc networks (MANET) is one of the greatest security concerns in the deployment of wireless networks. Thus communication in MANET functions properly only if the participating nodes cooperate in routing without any malicious intention. However, some of the nodes may be malicious in their behavior, by indulging in flooding attacks on their neighbors. Some others may act malicious by launching active security attacks like denial of service. This paper addresses few related works done on trust evaluation and establishment in ad hoc networks. Related works on flooding attack prevention are reviewed. A new trust approach based on the extent of friendship between the nodes is proposed which makes the nodes to co-operate and prevent flooding attacks in an ad hoc environment. The performance of the trust algorithm is tested in an ad hoc network implementing the Ad hoc On-demand Distance Vector (AODV) protocol.Keywords: AODV, Flooding, MANETs, trust estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2392