Search results for: event monitoring
4270 Event Monitoring Based On Web Services for Heterogeneous Event Sources
Authors: Arne Koschel
Abstract:
This article discusses event monitoring options for heterogeneous event sources as they are given in nowadays heterogeneous distributed information systems. It follows the central assumption, that a fully generic event monitoring solution cannot provide complete support for event monitoring; instead, event source specific semantics such as certain event types or support for certain event monitoring techniques have to be taken into account. Following from this, the core result of the work presented here is the extension of a configurable event monitoring (Web) service for a variety of event sources. A service approach allows us to trade genericity for the exploitation of source specific characteristics. It thus delivers results for the areas of SOA, Web services, CEP and EDA.Keywords: event monitoring, ECA, CEP, SOA, web services
Procedia PDF Downloads 7444269 A Survey on a Critical Infrastructure Monitoring Using Wireless Sensor Networks
Authors: Khelifa Benahmed, Tarek Benahmed
Abstract:
There are diverse applications of wireless sensor networks (WSNs) in the real world, typically invoking some kind of monitoring, tracking, or controlling activities. In an application, a WSN is deployed over the area of interest to sense and detect the events and collect data through their sensors in a geographical area and transmit the collected data to a Base Station (BS). This paper presents an overview of the research solutions available in the field of environmental monitoring applications, more precisely the problems of critical area monitoring using wireless sensor networks.Keywords: critical infrastructure monitoring, environment monitoring, event region detection, wireless sensor networks
Procedia PDF Downloads 3514268 Complex Event Processing System Based on the Extended ECA Rule
Authors: Kwan Hee Han, Jun Woo Lee, Sung Moon Bae, Twae Kyung Park
Abstract:
ECA (Event-Condition-Action) languages are largely adopted for event processing since they are an intuitive and powerful paradigm for programming reactive systems. However, there are some limitations about ECA rules for processing of complex events such as coupling of event producer and consumer. The objective of this paper is to propose an ECA rule pattern to improve the current limitations of ECA rule, and to develop a prototype system. In this paper, conventional ECA rule is separated into 3 parts and each part is extended to meet the requirements of CEP. Finally, event processing logic is established by combining the relevant elements of 3 parts. The usability of proposed extended ECA rule is validated by a test scenario in this study.Keywords: complex event processing, ECA rule, Event processing system, event-driven architecture, internet of things
Procedia PDF Downloads 5304267 Event Extraction, Analysis, and Event Linking
Authors: Anam Alam, Rahim Jamaluddin Kanji
Abstract:
With the rapid growth of event in everywhere, event extraction has now become an important matter to retrieve the information from the unstructured data. One of the challenging problems is to extract the event from it. An event is an observable occurrence of interaction among entities. The paper investigates the effectiveness of event extraction capabilities of three software tools that are Wandora, Nitro and SPSS. We performed standard text mining techniques of these tools on the data sets of (i) Afghan War Diaries (AWD collection), (ii) MUC4 and (iii) WebKB. Information retrieval measures such as precision and recall which are computed under extensive set of experiments for Event Extraction. The experimental study analyzes the difference between events extracted by the software and human. This approach helps to construct an algorithm that will be applied for different machine learning methods.Keywords: event extraction, Wandora, nitro, SPSS, event analysis, extraction method, AFG, Afghan War Diaries, MUC4, 4 universities, dataset, algorithm, precision, recall, evaluation
Procedia PDF Downloads 5964266 Hedging and Corporate Governance: Lessons from the Financial Crisis
Authors: Rodrigo Zeidan
Abstract:
The paper identifies failures of decision making and corporate governance that allow non-financial companies around the world to develop hedging strategies that lead to hefty losses in the aftermath of the financial crisis. The sample is comprised of 346 companies from 10 international markets, of which 49 companies (and a subsample of 13 distressed companies) lose a combined US$18.9 billion. An event study shows that most companies that present losses in derivatives experience negative abnormal returns, including a number of companies in which the effect is persistent after a year. The results of a probit model indicate that the lack of a formal hedging policy, no monitoring to the CFOs, and considerations of hubris and remuneration contribute to the mismanagement of hedging policies.Keywords: risk management, hedging, derivatives, monitoring, corporate governance structure, event study, hubris
Procedia PDF Downloads 4424265 A Framework for Event-Based Monitoring of Business Processes in the Supply Chain Management of Industry 4.0
Authors: Johannes Atug, Andreas Radke, Mitchell Tseng, Gunther Reinhart
Abstract:
In modern supply chains, large numbers of SKU (Stock-Keeping-Unit) need to be timely managed, and any delays in noticing disruptions of items often limit the ability to defer the impact on customer order fulfillment. However, in supply chains of IoT-connected enterprises, the ERP (Enterprise-Resource-Planning), the MES (Manufacturing-Execution-System) and the SCADA (Supervisory-Control-and-Data-Acquisition) systems generate large amounts of data, which generally glean much earlier notice of deviations in the business process steps. That is, analyzing these streams of data with process mining techniques allows the monitoring of the supply chain business processes and thus identification of items that deviate from the standard order fulfillment process. In this paper, a framework to enable event-based SCM (Supply-Chain-Management) processes including an overview of core enabling technologies are presented, which is based on the RAMI (Reference-Architecture-Model for Industrie 4.0) architecture. The application of this framework in the industry is presented, and implications for SCM in industry 4.0 and further research are outlined.Keywords: cyber-physical production systems, event-based monitoring, supply chain management, RAMI (Reference-Architecture-Model for Industrie 4.0)
Procedia PDF Downloads 2364264 Optimizing Stormwater Sampling Design for Estimation of Pollutant Loads
Authors: Raja Umer Sajjad, Chang Hee Lee
Abstract:
Stormwater runoff is the leading contributor to pollution of receiving waters. In response, an efficient stormwater monitoring program is required to quantify and eventually reduce stormwater pollution. The overall goals of stormwater monitoring programs primarily include the identification of high-risk dischargers and the development of total maximum daily loads (TMDLs). The challenge in developing better monitoring program is to reduce the variability in flux estimates due to sampling errors; however, the success of monitoring program mainly depends on the accuracy of the estimates. Apart from sampling errors, manpower and budgetary constraints also influence the quality of the estimates. This study attempted to develop optimum stormwater monitoring design considering both cost and the quality of the estimated pollutants flux. Three years stormwater monitoring data (2012 – 2014) from a mix land use located within Geumhak watershed South Korea was evaluated. The regional climate is humid and precipitation is usually well distributed through the year. The investigation of a large number of water quality parameters is time-consuming and resource intensive. In order to identify a suite of easy-to-measure parameters to act as a surrogate, Principal Component Analysis (PCA) was applied. Means, standard deviations, coefficient of variation (CV) and other simple statistics were performed using multivariate statistical analysis software SPSS 22.0. The implication of sampling time on monitoring results, number of samples required during the storm event and impact of seasonal first flush were also identified. Based on the observations derived from the PCA biplot and the correlation matrix, total suspended solids (TSS) was identified as a potential surrogate for turbidity, total phosphorus and for heavy metals like lead, chromium, and copper whereas, Chemical Oxygen Demand (COD) was identified as surrogate for organic matter. The CV among different monitored water quality parameters were found higher (ranged from 3.8 to 15.5). It suggests that use of grab sampling design to estimate the mass emission rates in the study area can lead to errors due to large variability. TSS discharge load calculation error was found only 2 % with two different sample size approaches; i.e. 17 samples per storm event and equally distributed 6 samples per storm event. Both seasonal first flush and event first flush phenomena for most water quality parameters were observed in the study area. Samples taken at the initial stage of storm event generally overestimate the mass emissions; however, it was found that collecting a grab sample after initial hour of storm event more closely approximates the mean concentration of the event. It was concluded that site and regional climate specific interventions can be made to optimize the stormwater monitoring program in order to make it more effective and economical.Keywords: first flush, pollutant load, stormwater monitoring, surrogate parameters
Procedia PDF Downloads 2404263 Bivariate Time-to-Event Analysis with Copula-Based Cox Regression
Authors: Duhania O. Mahara, Santi W. Purnami, Aulia N. Fitria, Merissa N. Z. Wirontono, Revina Musfiroh, Shofi Andari, Sagiran Sagiran, Estiana Khoirunnisa, Wahyudi Widada
Abstract:
For assessing interventions in numerous disease areas, the use of multiple time-to-event outcomes is common. An individual might experience two different events called bivariate time-to-event data, the events may be correlated because it come from the same subject and also influenced by individual characteristics. The bivariate time-to-event case can be applied by copula-based bivariate Cox survival model, using the Clayton and Frank copulas to analyze the dependence structure of each event and also the covariates effect. By applying this method to modeling the recurrent event infection of hemodialysis insertion on chronic kidney disease (CKD) patients, from the AIC and BIC values we find that the Clayton copula model was the best model with Kendall’s Tau is (τ=0,02).Keywords: bivariate cox, bivariate event, copula function, survival copula
Procedia PDF Downloads 824262 Analyzing the Evolution of Adverse Events in Pharmacovigilance: A Data-Driven Approach
Authors: Kwaku Damoah
Abstract:
This study presents a comprehensive data-driven analysis to understand the evolution of adverse events (AEs) in pharmacovigilance. Utilizing data from the FDA Adverse Event Reporting System (FAERS), we employed three analytical methods: rank-based, frequency-based, and percentage change analyses. These methods assessed temporal trends and patterns in AE reporting, focusing on various drug-active ingredients and patient demographics. Our findings reveal significant trends in AE occurrences, with both increasing and decreasing patterns from 2000 to 2023. This research highlights the importance of continuous monitoring and advanced analysis in pharmacovigilance, offering valuable insights for healthcare professionals and policymakers to enhance drug safety.Keywords: event analysis, FDA adverse event reporting system, pharmacovigilance, temporal trend analysis
Procedia PDF Downloads 484261 The Event of Extreme Precipitation Occurred in the Metropolitan Mesoregion of the Capital of Para
Authors: Natasha Correa Vitória Bandeira, Lais Cordeiro Soares, Claudineia Brazil, Luciane Teresa Salvi
Abstract:
The intense rain event that occurred between February 16 and 18, 2018, in the city of Barcarena in Pará, located in the North region of Brazil, demonstrates the importance of analyzing this type of event. The metropolitan mesoregion of Belem was severely punished by rains much above the averages normally expected for that time of year; this phenomenon affected, in addition to the capital, the municipalities of Barcarena, Murucupi and Muruçambá. Resulting in a great flood in the rivers of the region, whose basins were affected with great intensity of precipitation, causing concern for the local population because in this region, there are located companies that accumulate ore tailings, and in this specific case, the dam of any of these companies, leaching the ore to the water bodies of the Murucupi River Basin. This article aims to characterize this phenomenon through a special analysis of the distribution of rainfall, using data from atmospheric soundings, satellite images, radar images and data from the GPCP (Global Precipitation Climatology Project), in addition to rainfall stations located in the study region. The results of the work demonstrated a dissociation between the data measured in the meteorological stations and the other forms of analysis of this extreme event. Monitoring carried out solely on the basis of data from pluviometric stations is not sufficient for monitoring and/or diagnosing extreme weather events, and investment by the competent bodies is important to install a larger network of pluviometric stations sufficient to meet the demand in a given region.Keywords: extreme precipitation, great flood, GPCP, ore dam
Procedia PDF Downloads 1084260 Structural Health Monitoring-Integrated Structural Reliability Based Decision Making
Authors: Caglayan Hizal, Kutay Yuceturk, Ertugrul Turker Uzun, Hasan Ceylan, Engin Aktas, Gursoy Turan
Abstract:
Monitoring concepts for structural systems have been investigated by researchers for decades since such tools are quite convenient to determine intervention planning of structures. Despite the considerable development in this regard, the efficient use of monitoring data in reliability assessment, and prediction models are still in need of improvement in their efficiency. More specifically, reliability-based seismic risk assessment of engineering structures may play a crucial role in the post-earthquake decision-making process for the structures. After an earthquake, professionals could identify heavily damaged structures based on visual observations. Among these, it is hard to identify the ones with minimum signs of damages, even if they would experience considerable structural degradation. Besides, visual observations are open to human interpretations, which make the decision process controversial, and thus, less reliable. In this context, when a continuous monitoring system has been previously installed on the corresponding structure, this decision process might be completed rapidly and with higher confidence by means of the observed data. At this stage, the Structural Health Monitoring (SHM) procedure has an important role since it can make it possible to estimate the system reliability based on a recursively updated mathematical model. Therefore, integrating an SHM procedure into the reliability assessment process comes forward as an important challenge due to the arising uncertainties for the updated model in case of the environmental, material and earthquake induced changes. In this context, this study presents a case study on SHM-integrated reliability assessment of the continuously monitored progressively damaged systems. The objective of this study is to get instant feedback on the current state of the structure after an extreme event, such as earthquakes, by involving the observed data rather than the visual inspections. Thus, the decision-making process after such an event can be carried out on a rational basis. In the near future, this can give wing to the design of self-reported structures which can warn about its current situation after an extreme event.Keywords: condition assessment, vibration-based SHM, reliability analysis, seismic risk assessment
Procedia PDF Downloads 1444259 Role of Discrete Event Simulation in the Assessment and Selection of the Potential Reconfigurable Manufacturing Solutions
Authors: Mohsin Raza, Arne Bilberg, Thomas Ditlev Brunø, Ann-Louise Andersen, Filip SKärin
Abstract:
Shifting from a dedicated or flexible manufacturing system to a reconfigurable manufacturing system (RMS) requires a significant amount of time, money, and effort. Therefore, it is vital to verify beforehand that the potential reconfigurable solution will be able to achieve the organizational objectives. Discrete event simulation offers the opportunity of assessing several reconfigurable alternatives against the set objectives. This study signifies the importance of using discrete-event simulation as a tool to verify several reconfiguration options. Two different industrial cases have been presented in the study to elaborate on the role of discrete event simulation in the implementation methodology of RMSs. The study concluded that discrete event simulation is one of the important tools to consider in the RMS implementation methodology.Keywords: reconfigurable manufacturing system, discrete event simulation, Tecnomatix plant simulation, RMS
Procedia PDF Downloads 1244258 Modeling and Analyzing the WAP Class 2 Wireless Transaction Protocol Using Event-B
Authors: Rajaa Filali, Mohamed Bouhdadi
Abstract:
This paper presents an incremental formal development of the Wireless Transaction Protocol (WTP) in Event-B. WTP is part of the Wireless Application Protocol (WAP) architectures and provides a reliable request-response service. To model and verify the protocol, we use the formal technique Event-B which provides an accessible and rigorous development method. This interaction between modelling and proving reduces the complexity and helps to eliminate misunderstandings, inconsistencies, and specification gaps. As result, verification of WTP allows us to find some deficiencies in the current specification.Keywords: event-B, wireless transaction protocol, proof obligation, refinement, Rodin, ProB
Procedia PDF Downloads 3174257 An Incremental Refinement Approach to a Development of Dynamic Host Configuration Protocol (DHCP) Using Event-B
Authors: Rajaa Filali, Mohamed Bouhdadi
Abstract:
This paper presents an incremental development of the Dynamic Host Configuration Protocol (DHCP) in Event-B. DHCP is widely used communication protocol, which provides a standard mechanism to obtain configuration parameters. The specification is performed in a stepwise manner and verified through a series of refinements. The Event-B formal method uses the Rodin platform to modeling and verifying some properties of the protocol such as safety, liveness and deadlock freedom. To model and verify the protocol, we use the formal technique Event-B which provides an accessible and rigorous development method. This interaction between modelling and proving reduces the complexity and helps to eliminate misunderstandings, inconsistencies, and specification gaps.Keywords: DHCP protocol, Event-B, refinement, proof obligation, Rodin
Procedia PDF Downloads 2284256 A Refinement Strategy Coupling Event-B and Planning Domain Definition Language (PDDL) for Planning Problems
Authors: Sabrine Ammar, Mohamed Tahar Bhiri
Abstract:
Automatic planning has a de facto standard language called Planning Domain Definition Language (PDDL) for describing planning problems. It aims to formalize the planning problems described by the concept of state space. PDDL-related dynamic analysis tools, namely planners and validators, are insufficient for verifying and validating PDDL descriptions. Indeed, these tools made it possible to detect errors a posteriori by means of test activity. In this paper, we recommend a formal approach coupling the two languages Event-B and PDDL, for automatic planning. Event-B is used for formal modeling by stepwise refinement with mathematical proofs of planning problems. Thus, this paper proposes a refinement strategy allowing to obtain reliable PDDL descriptions from an ultimate Event-B model correct by construction. The ultimate Event-B model, correct by construction which is supposed to be translatable into PDDL, is automatically translated into PDDL using our MDE Event-B2PDDL tool.Keywords: code generation, event-b, PDDL, refinement strategy, translation rules
Procedia PDF Downloads 1964255 Topic-to-Essay Generation with Event Element Constraints
Authors: Yufen Qin
Abstract:
Topic-to-Essay generation is a challenging task in Natural language processing, which aims to generate novel, diverse, and topic-related text based on user input. Previous research has overlooked the generation of articles under the constraints of event elements, resulting in issues such as incomplete event elements and logical inconsistencies in the generated results. To fill this gap, this paper proposes an event-constrained approach for a topic-to-essay generation that enforces the completeness of event elements during the generation process. Additionally, a language model is employed to verify the logical consistency of the generated results. Experimental results demonstrate that the proposed model achieves a better BLEU-2 score and performs better than the baseline in terms of subjective evaluation on a real dataset, indicating its capability to generate higher-quality topic-related text.Keywords: event element, language model, natural language processing, topic-to-essay generation.
Procedia PDF Downloads 2364254 Event-Led Strategy for Cultural Tourism Development: The Case of Liverpool as the 2008 European Capital of Culture
Authors: Yi-De Liu
Abstract:
Cultural tourism is one of the largest and fastest growing global tourism markets and the cultures are increasingly being used to promote cities and to increase their competitiveness and attractiveness. One of the major forms of cultural tourism development undertaken throughout Europe has been the staging of a growing number of cultural events. The event of European Capitals of Culture (ECOC) is probably the best example of the new trends of cultural tourism in Europe, which is therefore used in this article to demonstrate some of the key issues surrounding the event-led strategy for cultural tourism development. Based on the experience of the 2008 ECOC Liverpool, UK, the study’s findings point to a number of ways in which the ECOC constitutes a boost for the development of cultural tourism in terms of realising experience economy, enhancing city image, facilitating urban regeneration, promoting cultural production and consumption, as well as establishing partnerships. This study is concluded by drawing some critical factors that event and tourism organisers should consider.Keywords: cultural tourism, event tourism, cultural event, European capital of culture, Liverpool
Procedia PDF Downloads 7064253 A Fuzzy TOPSIS Based Model for Safety Risk Assessment of Operational Flight Data
Authors: N. Borjalilu, P. Rabiei, A. Enjoo
Abstract:
Flight Data Monitoring (FDM) program assists an operator in aviation industries to identify, quantify, assess and address operational safety risks, in order to improve safety of flight operations. FDM is a powerful tool for an aircraft operator integrated into the operator’s Safety Management System (SMS), allowing to detect, confirm, and assess safety issues and to check the effectiveness of corrective actions, associated with human errors. This article proposes a model for safety risk assessment level of flight data in a different aspect of event focus based on fuzzy set values. It permits to evaluate the operational safety level from the point of view of flight activities. The main advantages of this method are proposed qualitative safety analysis of flight data. This research applies the opinions of the aviation experts through a number of questionnaires Related to flight data in four categories of occurrence that can take place during an accident or an incident such as: Runway Excursions (RE), Controlled Flight Into Terrain (CFIT), Mid-Air Collision (MAC), Loss of Control in Flight (LOC-I). By weighting each one (by F-TOPSIS) and applying it to the number of risks of the event, the safety risk of each related events can be obtained.Keywords: F-topsis, fuzzy set, flight data monitoring (FDM), flight safety
Procedia PDF Downloads 1684252 Pharmacovigilance: An Empowerment in Safe Utilization of Pharmaceuticals
Authors: Pankaj Prashar, Bimlesh Kumar, Ankita Sood, Anamika Gautam
Abstract:
Pharmacovigilance (PV) is a rapidly growing discipline in pharmaceutical industries as an integral part of clinical research and drug development over the past few decades. PV carries a breadth of scope from drug manufacturing to its regulation with safer utilization. The fundamental steps of PV not only includes data collection and verification, coding of drugs with adverse drug reactions, causality assessment and timely reporting to the authorities but also monitoring drug manufacturing, safety issues, product quality and conduction of due diligence. Standardization of adverse event information, collaboration of multiple departments in different companies, preparation of documents in accordance to both governmental as well as non-governmental organizations (FDA, EMA, GVP, ICH) are the advancements in discipline of PV. De-harmonization, lack of predictive drug safety models, improper funding by government, non-reporting, and non-acceptability of ADRs by developing countries and reports directly from patients to the monitoring centres respectively are the major road backs of PV. Mandatory pharmacovigilance reporting, frequent inspections, funding by government, educating and training medical students, pharmacists and nurses in this segment can bring about empowerment in PV. This area needs to be addressed with a sense of urgency for the safe utilization of pharmaceuticals.Keywords: pharmacovigilance, regulatory, adverse event, drug safety
Procedia PDF Downloads 1244251 Health Status Monitoring of COVID-19 Patient's through Blood Tests and Naïve-Bayes
Authors: Carlos Arias-Alcaide, Cristina Soguero-Ruiz, Paloma Santos-Álvarez, Adrián García-Romero, Inmaculada Mora-Jiménez
Abstract:
Analysing clinical data with computers in such a way that have an impact on the practitioners’ workflow is a challenge nowadays. This paper provides a first approach for monitoring the health status of COVID-19 patients through the use of some biomarkers (blood tests) and the simplest Naïve Bayes classifier. Data of two Spanish hospitals were considered, showing the potential of our approach to estimate reasonable posterior probabilities even some days before the event.Keywords: Bayesian model, blood biomarkers, classification, health tracing, machine learning, posterior probability
Procedia PDF Downloads 2334250 The Impact of Reshuffle in Indonesian Working Cabinet Volume II to Abnormal Return and Abnormal Trading Activity of Companies Listed in the Jakarta Islamic Index
Authors: Fatin Fadhilah Hasib, Dewi Nuraini, Nisful Laila, Muhammad Madyan
Abstract:
A big political event such as Cabinet reshuffle mostly can affect the stock price positively or negatively, depend on the perception of each investor and potential investor. This study aims to analyze the movement of the market and trading activities which respect to an event using event study method. This method is used to measure the movement of the stock exchange in which abnormal return can be obtained by investor related to the event. This study examines the differences of reaction on abnormal return and trading volume activity from the companies listed in the Jakarta Islamic Index (JII), before and after the announcement of the Cabinet Work Volume II on 27 July 2016. The study was conducted in observation of 21 days in total which consists of 10 days before the event and 10 days after the event. The method used in this study is event study with market adjusted model method that observes market reaction to the information of an announcement or publicity events. The Results from the study showed that there is no significant negative nor positive reaction at the abnormal return and abnormal trading before and after the announcement of the cabinet reshuffle. It is indicated by the results of statistical tests whose value not exceeds the level of significance. Stock exchange of the JII just reflects from the previous stock prices without reflecting the information regarding to the Cabinet reshuffle event. It can be concluded that the capital market is efficient with a weak form.Keywords: abnormal return, abnormal trading volume activity, event study, political event
Procedia PDF Downloads 2934249 Modelling of Relocation and Battery Autonomy Problem on Electric Cars Sharing Dynamic by Using Discrete Event Simulation and Petri Net
Authors: Taha Benarbia, Kay W. Axhausen, Anugrah Ilahi
Abstract:
Electric car sharing system as ecologic transportation increasing in the world. The complexity of managing electric car sharing systems, especially one-way trips and battery autonomy have direct influence to on supply and demand of system. One must be able to precisely model the demand and supply of these systems to better operate electric car sharing and estimate its effect on mobility management and the accessibility that it provides in urban areas. In this context, our work focus to develop performances optimization model of the system based on discrete event simulation and stochastic Petri net. The objective is to search optimal decisions and management parameters of the system in order to fulfil at best demand while minimizing undesirable situations. In this paper, we present new model of electric cars sharing with relocation based on monitoring system. The proposed approach also help to precise the influence of battery charging level on the behaviour of system as important decision parameter of this complex and dynamical system.Keywords: electric car-sharing systems, smart mobility, Petri nets modelling, discrete event simulation
Procedia PDF Downloads 1834248 Risk Assessments of Longest Dry Spells Phenomenon in Northern Tunisia
Authors: Majid Mathlouthi, Fethi Lebdi
Abstract:
Throughout the world, the extent and magnitude of droughts have economic, social and environmental consequences. Today climate change has become more and more felt; most likely they increase the frequency and duration of droughts. An analysis by event of dry event, from series of observations of the daily rainfall is carried out. A daily precipitation threshold value has been set. A catchment localized in Northern Tunisia where the average rainfall is about 600 mm has been studied. Rainfall events are defined as an uninterrupted series of rainfall days understanding at least a day having received a precipitation superior or equal to a fixed threshold. The dry events are constituted of a series of dry days framed by two successive rainfall events. A rainfall event is a vector of coordinates the duration, the rainfall depth per event and the duration of the dry event. The depth and duration are found to be correlated. So we use conditional probabilities to analyse the depth per event. The negative binomial distribution fits well the dry event. The duration of the rainfall event follows a geometric distribution. The length of the climatically cycle adjusts to the Incomplete Gamma. Results of this analysis was used to study of the effects of climate change on water resources and crops and to calibrate precipitation models with little rainfall records. In response to long droughts in the basin, the drought management system is based on three phases during each of the three phases; different measurements are applied and executed. The first is before drought, preparedness and early warning; the second is drought management, mitigation in the event of drought; and the last subsequent drought, when the drought is over.Keywords: dry spell, precipitation threshold, climate vulnerability, adaptation measures
Procedia PDF Downloads 844247 Single Event Transient Tolerance Analysis in 8051 Microprocessor Using Scan Chain
Authors: Jun Sung Go, Jong Kang Park, Jong Tae Kim
Abstract:
As semi-conductor manufacturing technology evolves; the single event transient problem becomes more significant issue. Single event transient has a critical impact on both combinational and sequential logic circuits, so it is important to evaluate the soft error tolerance of the circuits at the design stage. In this paper, we present a soft error detecting simulation using scan chain. The simulation model generates a single event transient randomly in the circuit, and detects the soft error during the execution of the test patterns. We verified this model by inserting a scan chain in an 8051 microprocessor using 65 nm CMOS technology. While the test patterns generated by ATPG program are passing through the scan chain, we insert a single event transient and detect the number of soft errors per sub-module. The experiments show that the soft error rates per cell area of the SFR module is 277% larger than other modules.Keywords: scan chain, single event transient, soft error, 8051 processor
Procedia PDF Downloads 3474246 Cross-Layer Design of Event-Triggered Adaptive OFDMA Resource Allocation Protocols with Application to Vehicle Clusters
Authors: Shaban Guma, Naim Bajcinca
Abstract:
We propose an event-triggered algorithm for the solution of a distributed optimization problem by means of the projected subgradient method. Thereby, we invoke an OFDMA resource allocation scheme by applying an event-triggered sensitivity analysis at the access point. The optimal resource assignment of the subcarriers to the involved wireless nodes is carried out by considering the sensitivity analysis of the overall objective function as defined by the control of vehicle clusters with respect to the information exchange between the nodes.Keywords: consensus, cross-layer, distributed, event-triggered, multi-vehicle, protocol, resource, OFDMA, wireless
Procedia PDF Downloads 3314245 Empirical Decomposition of Time Series of Power Consumption
Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats
Abstract:
Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).Keywords: general appliance model, non intrusive load monitoring, events detection, unsupervised techniques;
Procedia PDF Downloads 824244 Sensing to Respond & Recover in Emergency
Authors: Alok Kumar, Raviraj Patil
Abstract:
The ability to respond to an incident of a disastrous event in a vulnerable area is very crucial an aspect of emergency management. The ability to constantly predict the likelihood of an event along with its severity in an area and react to those significant events which are likely to have a high impact allows the authorities to respond by allocating resources optimally in a timely manner. It provides for measuring, monitoring, and modeling facilities that integrate underlying systems into one solution to improve operational efficiency, planning, and coordination. We were particularly involved in this innovative incubation work on the current state of research and development in collaboration. technologies & systems for a disaster.Keywords: predictive analytics, advanced analytics, area flood likelihood model, area flood severity model, level of impact model, mortality score, economic loss score, resource allocation, crew allocation
Procedia PDF Downloads 3214243 Electrical Decomposition of Time Series of Power Consumption
Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats
Abstract:
Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).Keywords: electrical disaggregation, DTW, general appliance modeling, event detection
Procedia PDF Downloads 784242 Condition Monitoring System of Mine Air Compressors Based on Wireless Sensor Network
Authors: Sheng Fu, Yinbo Gao, Hao Lin
Abstract:
In the current mine air compressors monitoring system, there are some difficulties in the installation and maintenance because of the wired connection. To solve the problem, this paper introduces a new air compressors monitoring system based on ZigBee in which the monitoring parameters are transmitted wirelessly. The collecting devices are designed to form a cluster network to collect vibration, temperature, and pressure of air cylinders and other parameters. All these devices are battery-powered. Besides, the monitoring software in PC is developed using MFC. Experiments show that the designed wireless sensor network works well in the site environmental condition and the system is very convenient to be installed since the wireless connection. This monitoring system will have a wide application prospect in the upgrade of the old monitoring system of the air compressors.Keywords: condition monitoring, wireless sensor network, air compressor, zigbee, data collecting
Procedia PDF Downloads 5084241 Chinese Event Detection Technique Based on Dependency Parsing and Rule Matching
Authors: Weitao Lin
Abstract:
To quickly extract adequate information from large-scale unstructured text data, this paper studies the representation of events in Chinese scenarios and performs the regularized abstraction. It proposes a Chinese event detection technique based on dependency parsing and rule matching. The method first performs dependency parsing on the original utterance, then performs pattern matching at the word or phrase granularity based on the results of dependent syntactic analysis, filters out the utterances with prominent non-event characteristics, and obtains the final results. The experimental results show the effectiveness of the method.Keywords: natural language processing, Chinese event detection, rules matching, dependency parsing
Procedia PDF Downloads 141