Search results for: discrete events
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2717

Search results for: discrete events

2537 Development of Earthquake and Typhoon Loss Models for Japan, Specifically Designed for Underwriting and Enterprise Risk Management Cycles

Authors: Nozar Kishi, Babak Kamrani, Filmon Habte

Abstract:

Natural hazards such as earthquakes and tropical storms, are very frequent and highly destructive in Japan. Japan experiences, every year on average, more than 10 tropical cyclones that come within damaging reach, and earthquakes of moment magnitude 6 or greater. We have developed stochastic catastrophe models to address the risk associated with the entire suite of damaging events in Japan, for use by insurance, reinsurance, NGOs and governmental institutions. KCC’s (Karen Clark and Company) catastrophe models are procedures constituted of four modular segments: 1) stochastic events sets that would represent the statistics of the past events, hazard attenuation functions that could model the local intensity, vulnerability functions that would address the repair need for local buildings exposed to the hazard, and financial module addressing policy conditions that could estimates the losses incurring as result of. The events module is comprised of events (faults or tracks) with different intensities with corresponding probabilities. They are based on the same statistics as observed through the historical catalog. The hazard module delivers the hazard intensity (ground motion or wind speed) at location of each building. The vulnerability module provides library of damage functions that would relate the hazard intensity to repair need as percentage of the replacement value. The financial module reports the expected loss, given the payoff policies and regulations. We have divided Japan into regions with similar typhoon climatology, and earthquake micro-zones, within each the characteristics of events are similar enough for stochastic modeling. For each region, then, a set of stochastic events is developed that results in events with intensities corresponding to annual occurrence probabilities that are of interest to financial communities; such as 0.01, 0.004, etc. The intensities, corresponding to these probabilities (called CE, Characteristics Events) are selected through a superstratified sampling approach that is based on the primary uncertainty. Region specific hazard intensity attenuation functions followed by vulnerability models leads to estimation of repair costs. Extensive economic exposure model addresses all local construction and occupancy types, such as post-linter Shinand Okabe wood, as well as concrete confined in steel, SRC (Steel-Reinforced Concrete), high-rise.

Keywords: typhoon, earthquake, Japan, catastrophe modelling, stochastic modeling, stratified sampling, loss model, ERM

Procedia PDF Downloads 239
2536 Statistical Wavelet Features, PCA, and SVM-Based Approach for EEG Signals Classification

Authors: R. K. Chaurasiya, N. D. Londhe, S. Ghosh

Abstract:

The study of the electrical signals produced by neural activities of human brain is called Electroencephalography. In this paper, we propose an automatic and efficient EEG signal classification approach. The proposed approach is used to classify the EEG signal into two classes: epileptic seizure or not. In the proposed approach, we start with extracting the features by applying Discrete Wavelet Transform (DWT) in order to decompose the EEG signals into sub-bands. These features, extracted from details and approximation coefficients of DWT sub-bands, are used as input to Principal Component Analysis (PCA). The classification is based on reducing the feature dimension using PCA and deriving the support-vectors using Support Vector Machine (SVM). The experimental are performed on real and standard dataset. A very high level of classification accuracy is obtained in the result of classification.

Keywords: discrete wavelet transform, electroencephalogram, pattern recognition, principal component analysis, support vector machine

Procedia PDF Downloads 609
2535 Numerical Investigation of Embankments for Protecting Rock Fall

Authors: Gökhan Altay, Cafer Kayadelen

Abstract:

Rock fall is a movement of huge rock blocks from dip slopes due to physical effects. It generally occurs where loose tuffs lying under basalt flow or stringcourse is being constituted by limestone layers which stand on clay. By corrosion of some parts, big cracks occur on layers and these cracks continue to grow with the effect of freezing-thawing. In this way, the breaking rocks fall down from these dip slopes. Earthquakes which can induce lots of rock movements is another reason for rock fall events. In Turkey, we have a large number of regions prone to the earthquake as in the World so this increases the possibility of rock fall events. A great number of rock fall events take place in Turkey as in the World every year. The rock fall events occurring in urban areas cause serious damages in houses, roads and workplaces. Sometimes it also hinders transportation and furthermore it maybe kills people. In Turkey, rock fall events happen mostly in Spring and Winter because of freezing- thawing of water in rock cracks frequently. In mountain and inclined areas, rock fall is risky for engineering construction and environment. Some countries can invest significant money for these risky areas. For instance, in Switzerland, approximately 6.7 million dollars is spent annually for a distance of 4 km, to the systems to prevent rock fall events. In Turkey, we have lots of urban areas and engineering structure that have the rock fall risk. The embankments are preferable for rock fall events because of its low maintenance and repair costs. Also, embankments are able to absorb much more energy according to other protection systems. The current design method of embankments is only depended on field tests results so there are inadequate studies about this design method. In this paper, the field test modeled in three dimensions and analysis are carried out with the help of ANSYS programme. By the help of field test from literature the numerical model validated. After the validity of numerical models additional parametric studies performed. Changes in deformation of embankments are investigated by the changes in, geometry, velocity and impact height of falling rocks.

Keywords: ANSYS, embankment, impact height, numerical analysis, rock fall

Procedia PDF Downloads 485
2534 Analysis of Nonlinear Bertrand Duopoly Game with Heterogeneous Players

Authors: Jixiang Zhang

Abstract:

A dynamic of Bertrand duopoly game is analyzed, where players use different production methods and choose their prices with bounded rationality. The equilibriums of the corresponding discrete dynamical systems are investigated. The stability conditions of Nash equilibrium under a local adjustment process are studied. The stability conditions of Nash equilibrium under a local adjustment process are studied. The stability of Nash equilibrium, as some parameters of the model are varied, gives rise to complex dynamics such as cycles of higher order and chaos. On this basis, we discover that an increase of adjustment speed of bounded rational player can make Bertrand market sink into the chaotic state. Finally, the complex dynamics, bifurcations and chaos are displayed by numerical simulation.

Keywords: Bertrand duopoly model, discrete dynamical system, heterogeneous expectations, nash equilibrium

Procedia PDF Downloads 372
2533 Bayesian Meta-Analysis to Account for Heterogeneity in Studies Relating Life Events to Disease

Authors: Elizabeth Stojanovski

Abstract:

Associations between life events and various forms of cancers have been identified. The purpose of a recent random-effects meta-analysis was to identify studies that examined the association between adverse events associated with changes to financial status including decreased income and breast cancer risk. The same association was studied in four separate studies which displayed traits that were not consistent between studies such as the study design, location and time frame. It was of interest to pool information from various studies to help identify characteristics that differentiated study results. Two random-effects Bayesian meta-analysis models are proposed to combine the reported estimates of the described studies. The proposed models allow major sources of variation to be taken into account, including study level characteristics, between study variance, and within study variance and illustrate the ease with which uncertainty can be incorporated using a hierarchical Bayesian modelling approach.

Keywords: random-effects, meta-analysis, Bayesian, variation

Procedia PDF Downloads 133
2532 Longitudinal Assessment on the Economic Impacts of Hosting Major Sports Events

Authors: Huei-Fu Lu

Abstract:

Hosting international major sports events (MSEs) has become a globalized strategy for many countries. Most modern countries believe that MSEs can bring the hosting countries with substantial and considerable economic and non-economic benefits; so many cities also input a huge of resources to bid for hosting MSEs. Despite the growing importance of MSEs, limited longitudinal analysis has been carried out to understand and explain the long term economic effects of such events. This paper is to continue the focus of previous literature on the economic effects of hosting MSEs. The study periods are from 1950 to 2014 and the secondary macro-economic data are selected from the countries that have hosted the Asian Games and the Olympic Games (including summer and winter) to precede a longitudinal analysis. A comparison of the real economic growth rate, investment, employment and international trade of hosting countries and the duration of these economic effects are also explored and discussed. Based on the countries’ attributes and locating area, aiming to ascertain whether hosting MSEs is economically worthwhile and whether the economic effects from MSEs are realized as anticipated. The results indicate that hosting MSEs to create positive economic effects like GDP growth or long-term employment may be a myth even for developing countries. However, the empirical findings can provide the sport management or authority with longitudinal and comprehensive elaboration for biding or hosting MSEs in the future.

Keywords: Asian Games, economic effects, major sports events (MSEs), olympic games

Procedia PDF Downloads 304
2531 Seismic Assessment of a Pre-Cast Recycled Concrete Block Arch System

Authors: Amaia Martinez Martinez, Martin Turek, Carlos Ventura, Jay Drew

Abstract:

This study aims to assess the seismic performance of arch and dome structural systems made from easy to assemble precast blocks of recycled concrete. These systems have been developed by Lock Block Ltd. Company from Vancouver, Canada, as an extension of their currently used retaining wall system. The characterization of the seismic behavior of these structures is performed by a combination of experimental static and dynamic testing, and analytical modeling. For the experimental testing, several tilt tests, as well as a program of shake table testing were undertaken using small scale arch models. A suite of earthquakes with different characteristics from important past events are chosen and scaled properly for the dynamic testing. Shake table testing applying the ground motions in just one direction (in the weak direction of the arch) and in the three directions were conducted and compared. The models were tested with increasing intensity until collapse occurred; which determines the failure level for each earthquake. Since the failure intensity varied with type of earthquake, a sensitivity analysis of the different parameters was performed, being impulses the dominant factor. For all cases, the arches exhibited the typical four-hinge failure mechanism, which was also shown in the analytical model. Experimental testing was also performed reinforcing the arches using a steel band over the structures anchored at both ends of the arch. The models were tested with different pretension levels. The bands were instrumented with strain gauges to measure the force produced by the shaking. These forces were used to develop engineering guidelines for the design of the reinforcement needed for these systems. In addition, an analytical discrete element model was created using 3DEC software. The blocks were designed as rigid blocks, assigning all the properties to the joints including also the contribution of the interlocking shear key between blocks. The model is calibrated to the experimental static tests and validated with the obtained results from the dynamic tests. Then the model can be used to scale up the results to the full scale structure and expanding it to different configurations and boundary conditions.

Keywords: arch, discrete element model, seismic assessment, shake-table testing

Procedia PDF Downloads 188
2530 Dynamic of Nonlinear Duopoly Game with Heterogeneous Players

Authors: Jixiang Zhang, Yanhua Wang

Abstract:

A dynamic of Bertrand duopoly game is analyzed, where players use different production methods and choose their prices with bounded rationality. The equilibriums of the corresponding discrete dynamical systems are investigated. The stability conditions of Nash equilibrium under a local adjustment process are studied. The stability conditions of Nash equilibrium under a local adjustment process are studied. The stability of Nash equilibrium, as some parameters of the model are varied, gives rise to complex dynamics such as cycles of higher order and chaos. On this basis, we discover that an increase of adjustment speed of bounded rational player can make Bertrand market sink into the chaotic state. Finally, the complex dynamics, bifurcations and chaos are displayed by numerical simulation.

Keywords: Bertrand duopoly model, discrete dynamical system, heterogeneous expectations, nash equilibrium

Procedia PDF Downloads 386
2529 A Discrete Event Simulation Model to Manage Bed Usage for Non-Elective Admissions in a Geriatric Medicine Speciality

Authors: Muhammed Ordu, Eren Demir, Chris Tofallis

Abstract:

Over the past decade, the non-elective admissions in the UK have increased significantly. Taking into account limited resources (i.e. beds), the related service managers are obliged to manage their resources effectively due to the non-elective admissions which are mostly admitted to inpatient specialities via A&E departments. Geriatric medicine is one of specialities that have long length of stay for the non-elective admissions. This study aims to develop a discrete event simulation model to understand how possible increases on non-elective demand over the next 12 months affect the bed occupancy rate and to determine required number of beds in a geriatric medicine speciality in a UK hospital. In our validated simulation model, we take into account observed frequency distributions which are derived from a big data covering the period April, 2009 to January, 2013, for the non-elective admission and the length of stay. An experimental analysis, which consists of 16 experiments, is carried out to better understand possible effects of case studies and scenarios related to increase on demand and number of bed. As a result, the speciality does not achieve the target level in the base model although the bed occupancy rate decreases from 125.94% to 96.41% by increasing the number of beds by 30%. In addition, the number of required beds is more than the number of beds considered in the scenario analysis in order to meet the bed requirement. This paper sheds light on bed management for service managers in geriatric medicine specialities.

Keywords: bed management, bed occupancy rate, discrete event simulation, geriatric medicine, non-elective admission

Procedia PDF Downloads 202
2528 Real-Time Image Encryption Using a 3D Discrete Dual Chaotic Cipher

Authors: M. F. Haroun, T. A. Gulliver

Abstract:

In this paper, an encryption algorithm is proposed for real-time image encryption. The scheme employs a dual chaotic generator based on a three dimensional (3D) discrete Lorenz attractor. Encryption is achieved using non-autonomous modulation where the data is injected into the dynamics of the master chaotic generator. The second generator is used to permute the dynamics of the master generator using the same approach. Since the data stream can be regarded as a random source, the resulting permutations of the generator dynamics greatly increase the security of the transmitted signal. In addition, a technique is proposed to mitigate the error propagation due to the finite precision arithmetic of digital hardware. In particular, truncation and rounding errors are eliminated by employing an integer representation of the data which can easily be implemented. The simple hardware architecture of the algorithm makes it suitable for secure real-time applications.

Keywords: chaotic systems, image encryption, non-autonomous modulation, FPGA

Procedia PDF Downloads 485
2527 Discrete PID and Discrete State Feedback Control of a Brushed DC Motor

Authors: I. Valdez, J. Perdomo, M. Colindres, N. Castro

Abstract:

Today, digital servo systems are extensively used in industrial manufacturing processes, robotic applications, vehicles and other areas. In such control systems, control action is provided by digital controllers with different compensation algorithms, which are designed to meet specific requirements for a given application. Due to the constant search for optimization in industrial processes, it is of interest to design digital controllers that offer ease of realization, improved computational efficiency, affordable return rates, and ease of tuning that ultimately improve the performance of the controlled actuators. There is a vast range of options of compensation algorithms that could be used, although in the industry, most controllers used are based on a PID structure. This research article compares different types of digital compensators implemented in a servo system for DC motor position control. PID compensation is evaluated on its two most common architectures: PID position form (1 DOF), and PID speed form (2 DOF). State feedback algorithms are also evaluated, testing two modern control theory techniques: discrete state observer for non-measurable variables tracking, and a linear quadratic method which allows a compromise between the theoretical optimal control and the realization that most closely matches it. The compared control systems’ performance is evaluated through simulations in the Simulink platform, in which it is attempted to model accurately each of the system’s hardware components. The criteria by which the control systems are compared are reference tracking and disturbance rejection. In this investigation, it is considered that the accurate tracking of the reference signal for a position control system is particularly important because of the frequency and the suddenness in which the control signal could change in position control applications, while disturbance rejection is considered essential because the torque applied to the motor shaft due to sudden load changes can be modeled as a disturbance that must be rejected, ensuring reference tracking. Results show that 2 DOF PID controllers exhibit high performance in terms of the benchmarks mentioned, as long as they are properly tuned. As for controllers based on state feedback, due to the nature and the advantage which state space provides for modelling MIMO, it is expected that such controllers evince ease of tuning for disturbance rejection, assuming that the designer of such controllers is experienced. An in-depth multi-dimensional analysis of preliminary research results indicate that state feedback control method is more satisfactory, but PID control method exhibits easier implementation in most control applications.

Keywords: control, DC motor, discrete PID, discrete state feedback

Procedia PDF Downloads 235
2526 A Co-Relational Descriptive Study to Assess the Impact of Cancer Event on Self, Family, Coping Level of Cancer Clients and Quality of Life among Them

Authors: Padma Sree Potru

Abstract:

Abstract: A co-relational descriptive study was conducted to assess the impact of cancer event on self, on family, coping strategies of cancer clients and quality of life among them in G.G.H., Guntur, Andhra Pradesh, India. Aim: The aim of the study was to investigate the impact of cancer events on self, on family, coping of clients and quality of life among cancer patients. Methods: 50 cancer patients were selected through random sampling technique. The data were obtained by using impact of events scale, impact on family scale, coping health inventory and WHOQOL-BREF scale. Results: The results revealed that majority (32%) of them were in the age group of 36-45 years, 72% were females, 44% were having the income of Rs. 5001-10000/- per month, 40% were working for daily wage, and 15% were newly diagnosed of cancer. Among 50 cancer patients, 65% had extreme impact of events, 61% shows extreme impact on family, 46% possess minimal coping strategies and 68% had poor quality of life. This study focuses on that there is a strong positive correlation between quality of life and coping behavior r=0.603 and also between impact of event and impact on family r=0.610, but a negative correlation existed between quality of life and impact of events r= -0.201. ANOVA test reveals that there is a significant difference between subscales of impact on family and coping behavior with f values = 3.893, 3.957 respectively. Chi-square highlights that there is a significant association between impact of events with age, occupation and impact on family with duration of illness. Conclusion: Even though cancer is a dreadful disease still there are many emerging treatment modalities and innovative procedures which are focusing on improving the standards of life among cancer clients. But all this can happen only when the clients accepts the reality, increase their willpower and confidence, desire to live, focusing on coping mechanisms and good ongoing support from the family members.

Keywords: impact of event, impact on family, coping, quality of event

Procedia PDF Downloads 422
2525 1D Velocity Model for the Gobi-Altai Region from Local Earthquakes

Authors: Dolgormaa Munkhbaatar, Munkhsaikhan Adiya, Tseedulam Khuut

Abstract:

We performed an inversion method to determine the 1D-velocity model with station corrections of the Gobi-Altai area in the southern part of Mongolia using earthquake data collected in the National Data Center during the last 10 years. In this study, the concept of the new 1D model has been employed to minimize the average RMS of a set of well-located earthquakes, recorded at permanent (between 2006 and 2016) and temporary seismic stations (between 2014 and 2016), compute solutions for the coupled hypocenter and 1D velocity model. We selected 4800 events with RMS less than 0.5 seconds and with a maximum GAP of 170 degrees and determined velocity structures. Also, we relocated all possible events located in the Gobi-Altai area using the new 1D velocity model and achieved constrained hypocentral determinations for events within this area. We concluded that the estimated new 1D velocity model is a relatively low range compared to the previous velocity model in a significant improvement intend to, and the quality of the information basis for future research center locations to determine the earthquake epicenter area with this new transmission model.

Keywords: 1D velocity model, earthquake, relocation, Velest

Procedia PDF Downloads 136
2524 Multi-Objective Discrete Optimization of External Thermal Insulation Composite Systems in Terms of Thermal and Embodied Energy Performance

Authors: Berfin Yildiz

Abstract:

These days, increasing global warming effects, limited amount of energy resources, etc., necessitates the awareness that must be present in every profession group. The architecture and construction sectors are responsible for both the embodied and operational energy of the materials. This responsibility has led designers to seek alternative solutions for energy-efficient material selection. The choice of energy-efficient material requires consideration of the entire life cycle, including the building's production, use, and disposal energy. The aim of this study is to investigate the method of material selection of external thermal insulation composite systems (ETICS). Embodied and in-use energy values of material alternatives were used for the evaluation in this study. The operational energy is calculated according to the u-value calculation method defined in the TS 825 (Thermal Insulation Requirements) standard for Turkey, and the embodied energy is calculated based on the manufacturer's Energy Performance Declaration (EPD). ETICS consists of a wall, adhesive, insulation, lining, mechanical, mesh, and exterior finishing materials. In this study, lining, mechanical, and mesh materials were ignored because EPD documents could not be obtained. The material selection problem is designed as a hypothetical volume area (5x5x3m) and defined as a multi-objective discrete optimization problem for external thermal insulation composite systems. Defining the problem as a discrete optimization problem is important in order to choose between materials of various thicknesses and sizes. Since production and use energy values, which are determined as optimization objectives in the study, are often conflicting values, material selection is defined as a multi-objective optimization problem, and it is aimed to obtain many solution alternatives by using Hypervolume (HypE) algorithm. The enrollment process started with 100 individuals and continued for 50 generations. According to the obtained results, it was observed that autoclaved aerated concrete and Ponce block as wall material, glass wool, as insulation material gave better results.

Keywords: embodied energy, multi-objective discrete optimization, performative design, thermal insulation

Procedia PDF Downloads 103
2523 Forklift Allocation in Warehouse Operations with Restricted Halls

Authors: Mauricio Becerra Fernández, Olga Rosana Romero Quiroga, Elsa Cristina González La Rotta

Abstract:

The logistics facilities design and construction is one of the strategic decisions that critically affects the performance of the company, from the economic perspective and relationship with customers. The case study company is the Colombian logistic sector leader, with over 60 years of experience, with sales of about one hundred twenty million dollars at the end of 2014. The preliminary design for the warehouse layout and operation includes a customer that provides approximately 17% of the profits of the company, considering the possibility of moving two forklifts in the warehouse halls. Some changes were not consider in previous stages of design, operations required forklift with different characteristics, whose size, do not allow the circulation of more than a forklift at a time. Therefore, it is necessary to assess the impact of this restriction on the warehouse operation, so decision makers implement actions to achieve efficient operation. The problem is addressed by recognizing logistics processes, which develop in a warehouse, collection of processes information behavior, the simulation of the current situation using ProModel software, model validation, making adjustments required, experiments design, conclusions and recommendations for the company.

Keywords: design, discrete events simulation, forklift allocation, logistics facilities, warehouse

Procedia PDF Downloads 275
2522 A Ground Structure Method to Minimize the Total Installed Cost of Steel Frame Structures

Authors: Filippo Ranalli, Forest Flager, Martin Fischer

Abstract:

This paper presents a ground structure method to optimize the topology and discrete member sizing of steel frame structures in order to minimize total installed cost, including material, fabrication and erection components. The proposed method improves upon existing cost-based ground structure methods by incorporating constructability considerations well as satisfying both strength and serviceability constraints. The architecture for the method is a bi-level Multidisciplinary Feasible (MDF) architecture in which the discrete member sizing optimization is nested within the topology optimization process. For each structural topology generated, the sizing optimization process seek to find a set of discrete member sizes that result in the lowest total installed cost while satisfying strength (member utilization) and serviceability (node deflection and story drift) criteria. To accurately assess cost, the connection details for the structure are generated automatically using accurate site-specific cost information obtained directly from fabricators and erectors. Member continuity rules are also applied to each node in the structure to improve constructability. The proposed optimization method is benchmarked against conventional weight-based ground structure optimization methods resulting in an average cost savings of up to 30% with comparable computational efficiency.

Keywords: cost-based structural optimization, cost-based topology and sizing, optimization, steel frame ground structure optimization, multidisciplinary optimization of steel structures

Procedia PDF Downloads 319
2521 Analysis of Rainfall Hazard in North East of Algeria

Authors: Imene Skhakhfa, Lahbaci Ouerdachi

Abstract:

The design of sewerage systems is directly related to rainfall, which has a highly random character. Showers are usually described by three characteristics: intensity, volume and duration. Several studies considered only in two of the three models. The objective of our work is to perform an analysis of the impact of three variables on put in charge of sewerage system, responsible for misbehavior, origin of urban flooding. 30 events were considered events for the longest, most rushed and most intense period which runs from 1986 -2001. We built the IDF curves and heavy projects double symmetrical triangles associated with this selection. A simulation of the operation, with the model canoe, sewage from the city of Annaba (Algeria) in the three rain solicitation project, double triangles associated with events considered. It appears that the sewage of the city of Annaba, in terms of charging, is much more sensitive to rain most precipitous, and the more intense causing loadings and last the longest. Further analysis of all the rain and the field measurements are underway to confirm the test simulations.

Keywords: intensity, volume, duration, sewerage, design, simulation

Procedia PDF Downloads 420
2520 Stator Short-Circuits Fault Diagnosis in Induction Motors

Authors: K. Yahia, M. Sahraoui, A. Guettaf

Abstract:

This paper deals with the problem of stator faults diagnosis in induction motors. Using the discrete wavelet transform (DWT) for the current Park’s vector modulus (CPVM) analysis, the inter-turn short-circuit faults diagnosis can be achieved. This method is based on the decomposition of the CPVM signal, where wavelet approximation and detail coefficients of this signal have been extracted. The energy evaluation of a known bandwidth detail permits to define a fault severity factor (FSF). This method has been tested through the simulation of an induction motor using a mathematical model based on the winding-function approach. Simulation, as well as experimental results, show the effectiveness of the used method.

Keywords: induction motors (IMs), inter-turn short-circuits diagnosis, discrete wavelet transform (DWT), Current Park’s Vector Modulus (CPVM)

Procedia PDF Downloads 433
2519 The Stock Price Effect of Apple Keynotes

Authors: Ethan Petersen

Abstract:

In this paper, we analyze the volatility of Apple’s stock beginning January 3, 2005 up to October 9, 2014, then focus on a range from 30 days prior to each product announcement until 30 days after. Product announcements are filtered; announcements whose 60 day range is devoid of other events are separated. This filtration is chosen to isolate, and study, a potential cross-effect. Concerning Apple keynotes, there are two significant dates: the day the invitations to the event are received and the day of the event itself. As such, the statistical analysis is conducted for both invite-centered and event-centered time frames. A comparison to the VIX is made to determine if the trend is simply following the market or deviating. Regardless of the filtration, we find that there is a clear deviation from the market. Comparing these data sets, there are significantly different trends: isolated events have a constantly decreasing, erratic trend in volatility but an increasing, linear trend is observed for clustered events. According to the Efficient Market Hypothesis, we would expect a change when new information is publicly known and the results of this study support this claim.

Keywords: efficient market hypothesis, event study, volatility, VIX

Procedia PDF Downloads 259
2518 Particle Swarm Optimization Algorithm vs. Genetic Algorithm for Image Watermarking Based Discrete Wavelet Transform

Authors: Omaima N. Ahmad AL-Allaf

Abstract:

Over communication networks, images can be easily copied and distributed in an illegal way. The copyright protection for authors and owners is necessary. Therefore, the digital watermarking techniques play an important role as a valid solution for authority problems. Digital image watermarking techniques are used to hide watermarks into images to achieve copyright protection and prevent its illegal copy. Watermarks need to be robust to attacks and maintain data quality. Therefore, we discussed in this paper two approaches for image watermarking, first is based on Particle Swarm Optimization (PSO) and the second approach is based on Genetic Algorithm (GA). Discrete wavelet transformation (DWT) is used with the two approaches separately for embedding process to cover image transformation. Each of PSO and GA is based on co-relation coefficient to detect the high energy coefficient watermark bit in the original image and then hide the watermark in original image. Many experiments were conducted for the two approaches with different values of PSO and GA parameters. From experiments, PSO approach got better results with PSNR equal 53, MSE equal 0.0039. Whereas GA approach got PSNR equal 50.5 and MSE equal 0.0048 when using population size equal to 100, number of iterations equal to 150 and 3×3 block. According to the results, we can note that small block size can affect the quality of image watermarking based PSO/GA because small block size can increase the search area of the watermarking image. Better PSO results were obtained when using swarm size equal to 100.

Keywords: image watermarking, genetic algorithm, particle swarm optimization, discrete wavelet transform

Procedia PDF Downloads 195
2517 Discrete Crack Modeling of Side Face FRP-Strengthened Concrete Beam

Authors: Shahriar Shahbazpanahi, Mohammad Hemen Jannaty, Alaleh Kamgar

Abstract:

Shear strengthening can be carried out in concrete structures by external fibre reinforced polymer (FRP). In the present investigation, a new fracture mechanics model is developed to model side face of strengthened concrete beam by external FRP. Discrete crack is simulated by a spring element with softening behavior ahead of the crack tip to model the cohesive zone in concrete. A truss element is used, parallel to the spring element, to simulate the energy dissipation rate by the FRP. The strain energy release rate is calculated directly by using a virtual crack closure technique and then, the crack propagation criterion is presented. The results are found acceptable when compared to previous experimental results and ABAQUS software data. It is observed that the length of the fracture process zone (FPZ) increases with the application of FRP in side face at the same load in comparison with that of the control beam.

Keywords: FPZ, fracture, FRP, shear

Procedia PDF Downloads 510
2516 Identifying Critical Links of a Transport Network When Affected by a Climatological Hazard

Authors: Beatriz Martinez-Pastor, Maria Nogal, Alan O'Connor

Abstract:

During the last years, the number of extreme weather events has increased. A variety of extreme weather events, including river floods, rain-induced landslides, droughts, winter storms, wildfire, and hurricanes, have threatened and damaged many different regions worldwide. These events have a devastating impact on critical infrastructure systems resulting in high social, economical and environmental costs. These events have a huge impact in transport systems. Since, transport networks are completely exposed to every kind of climatological perturbations, and its performance is closely related with these events. When a traffic network is affected by a climatological hazard, the quality of its service is threatened, and the level of the traffic conditions usually decreases. With the aim of understanding this process, the concept of resilience has become most popular in the area of transport. Transport resilience analyses the behavior of a traffic network when a perturbation takes place. This holistic concept studies the complete process, from the beginning of the perturbation until the total recovery of the system, when the perturbation has finished. Many concepts are included in the definition of resilience, such as vulnerability, redundancy, adaptability, and safety. Once the resilience of a transport network can be evaluated, in this case, the methodology used is a dynamic equilibrium-restricted assignment model that allows the quantification of the concept, the next step is its improvement. Through the improvement of this concept, it will be possible to create transport networks that are able to withstand and have a better performance under the presence of climatological hazards. Analyzing the impact of a perturbation in a traffic network, it is observed that the response of the different links, which are part of the network, can be completely different from one to another. Consequently and due to this effect, many questions arise, as what makes a link more critical before an extreme weather event? or how is it possible to identify these critical links? With this aim, and knowing that most of the times the owners or managers of the transport systems have limited resources, the identification of the critical links of a transport network before extreme weather events, becomes a crucial objective. For that reason, using the available resources in the areas that will generate a higher improvement of the resilience, will contribute to the global development of the network. Therefore, this paper wants to analyze what kind of characteristic makes a link a critical one when an extreme weather event damages a transport network and finally identify them.

Keywords: critical links, extreme weather events, hazard, resilience, transport network

Procedia PDF Downloads 262
2515 Blockchain’s Feasibility in Military Data Networks

Authors: Brenden M. Shutt, Lubjana Beshaj, Paul L. Goethals, Ambrose Kam

Abstract:

Communication security is of particular interest to military data networks. A relatively novel approach to network security is blockchain, a cryptographically secured distribution ledger with a decentralized consensus mechanism for data transaction processing. Recent advances in blockchain technology have proposed new techniques for both data validation and trust management, as well as different frameworks for managing dataflow. The purpose of this work is to test the feasibility of different blockchain architectures as applied to military command and control networks. Various architectures are tested through discrete-event simulation and the feasibility is determined based upon a blockchain design’s ability to maintain long-term stable performance at industry standards of throughput, network latency, and security. This work proposes a consortium blockchain architecture with a computationally inexpensive consensus mechanism, one that leverages a Proof-of-Identity (PoI) concept and a reputation management mechanism.

Keywords: blockchain, consensus mechanism, discrete-event simulation, fog computing

Procedia PDF Downloads 109
2514 Daily Probability Model of Storm Events in Peninsular Malaysia

Authors: Mohd Aftar Abu Bakar, Noratiqah Mohd Ariff, Abdul Aziz Jemain

Abstract:

Storm Event Analysis (SEA) provides a method to define rainfalls events as storms where each storm has its own amount and duration. By modelling daily probability of different types of storms, the onset, offset and cycle of rainfall seasons can be determined and investigated. Furthermore, researchers from the field of meteorology will be able to study the dynamical characteristics of rainfalls and make predictions for future reference. In this study, four categories of storms; short, intermediate, long and very long storms; are introduced based on the length of storm duration. Daily probability models of storms are built for these four categories of storms in Peninsular Malaysia. The models are constructed by using Bernoulli distribution and by applying linear regression on the first Fourier harmonic equation. From the models obtained, it is found that daily probability of storms at the Eastern part of Peninsular Malaysia shows a unimodal pattern with high probability of rain beginning at the end of the year and lasting until early the next year. This is very likely due to the Northeast monsoon season which occurs from November to March every year. Meanwhile, short and intermediate storms at other regions of Peninsular Malaysia experience a bimodal cycle due to the two inter-monsoon seasons. Overall, these models indicate that Peninsular Malaysia can be divided into four distinct regions based on the daily pattern for the probability of various storm events.

Keywords: daily probability model, monsoon seasons, regions, storm events

Procedia PDF Downloads 321
2513 Petri Net Modeling and Simulation of a Call-Taxi System

Authors: T. Godwin

Abstract:

A call-taxi system is a type of taxi service where a taxi could be requested through a phone call or mobile app. A schematic functioning of a call-taxi system is modeled using Petri net, which provides the necessary conditions for a taxi to be assigned by a dispatcher to pick a customer as well as the conditions for the taxi to be released by the customer. A Petri net is a graphical modeling tool used to understand sequences, concurrences, and confluences of activities in the working of discrete event systems. It uses tokens on a directed bipartite multi-graph to simulate the activities of a system. The Petri net model is translated into a simulation model and a call-taxi system is simulated. The simulation model helps in evaluating the operation of a call-taxi system based on the fleet size as well as the operating policies for call-taxi assignment and empty call-taxi repositioning. The developed Petri net based simulation model can be used to decide the fleet size as well as the call-taxi assignment policies for a call-taxi system.

Keywords: call-taxi, discrete event system, petri net, simulation modeling

Procedia PDF Downloads 401
2512 Land Use Sensitivity Map for the Extreme Flood Events in the Kelantan River Basin

Authors: Nader Saadatkhah, Jafar Rahnamarad, Shattri Mansor, Zailani Khuzaimah, Arnis Asmat, Nor Aizam Adnan, Siti Noradzah Adam

Abstract:

Kelantan river basin as a flood prone area at the east coast of the peninsular Malaysia has suffered several flood and mudflow events in the recent years. The current research attempted to assess the land cover changes impact in the Kelantan river basin focused on the runoff contributions from different land cover classes and the potential impact of land cover changes on runoff generation. In this regards, the hydrological regional modeling of rainfall induced runoff event as the improved transient rainfall infiltration and grid based regional model (Improved-TRIGRS) was employed to compute rate of infiltration, and subsequently changes in the discharge volume in this study. The effects of land use changes on peak flow and runoff volume was investigated using storm rainfall events during the last three decades.

Keywords: improved-TRIGRS model, land cover changes, Kelantan river basin, flood event

Procedia PDF Downloads 384
2511 Opacity Synthesis with Orwellian Observers

Authors: Moez Yeddes

Abstract:

The property of opacity is widely used in the formal verification of security in computer systems and protocols. Opacity is a general language-theoretic scheme of many security properties of systems. Opacity is parametrized with framework in which several security properties of a system can be expressed. A secret behaviour of a system is opaque if a passive attacker can never deduce its occurrence from the system observation. Instead of considering the case of static observability where the set of observable events is fixed off-line or dynamic observability where the set of observable events changes over time depending on the history of the trace, we introduce Orwellian partial observability where unobservable events are not revealed provided that downgrading events never occurs in the future of the trace. Orwellian partial observability is needed to model intransitive information flow. This Orwellian observability is knwon as ipurge function. We show in previous work how to verify opacity for regular secret is opaque for a regular language L w.r.t. an Orwellian projection is PSPACE-complete while it has been proved undecidable even for a regular language L w.r.t. a general Orwellian observation function. In this paper, we address two problems of opacification of a regular secret ϕ for a regular language L w.r.t. an Orwellian projection: Given L and a secret ϕ ∈ L, the first problem consist to compute some minimal regular super-language M of L, if it exists, such that ϕ is opaque for M and the second consists to compute the supremal sub-language M′ of L such that ϕ is opaque for M′. We derive both language-theoretic characterizations and algorithms to solve these two dual problems.

Keywords: security policies, opacity, formal verification, orwellian observation

Procedia PDF Downloads 202
2510 Stressful Life Events and Their Influence on Childhood Obesity and Emotional Well-Being: Cross-Sectional Study

Authors: M. Rojo, M. Blanco, T. Lacruz, S. Solano, L. Beltran, M. Graell, A. R. Sepulveda.

Abstract:

There is an association between an early accumulation of Stressful Life Events (SLE) during childhood and various physical and psychological health complications. However, there are only a few studies on this topic in children and adolescents with overweight or obesity. The general aim of the study was to evaluate the accumulation and type of SLE in 200 children from 8 to 12 years old and analyze the relationship with their emotional well-being and weight status (obesity, overweight and normal weight). The children and their families completed an interview. The evaluated variables that are included in this study are sociodemographic measures, medical/psychological history, anthropometric measures (BMI, z-BMI), and psychological variables (children's clinical interview K-SADS-PL(Schedule for Affective Disorders and Schizophrenia for School-Age Children Present and Lifetime Version) and battery of questionnaires). Results: Children with overweight and obesity accumulate more stressful events from an early age and have a significantly higher percentage of psychiatric diagnoses, compared to their peers with normal weight. Presenting a child psychiatric disorder is related to greater z-BMI and the total number of SLE (p < 0.001). A higher z-BMI is also related to a greater number of stressful events during childhood. There is also a positive and significant relationship between the total number of SLE and worse emotional well-being (higher levels of anxious and depressive symptoms and low self-esteem of children) (p < 0.01). Conclusion: Children with overweight and obesity grow up in a family, school, and social context where more stressors are accumulated. This is also directly associated with worse emotional well-being. It is necessary to implement multidisciplinary prevention and intervention strategies in different changes (school, family, and health). This study is included in a project funded by the Ministry of Innovation and Science (PSI2011-23127).

Keywords: childhood obesity, emotional well-being, psychopathology, stressful life events

Procedia PDF Downloads 100
2509 A Hybrid Watermarking Model Based on Frequency of Occurrence

Authors: Hamza A. A. Al-Sewadi, Adnan H. M. Al-Helali, Samaa A. K. Khamis

Abstract:

Ownership proofs of multimedia such as text, image, audio or video files can be achieved by the burial of watermark is them. It is achieved by introducing modifications into these files that are imperceptible to the human senses but easily recoverable by a computer program. These modifications would be in the time domain or frequency domain or both. This paper presents a procedure for watermarking by mixing amplitude modulation with frequency transformation histogram; namely a specific value is used to modulate the intensity component Y of the YIQ components of the carrier image. This scheme is referred to as histogram embedding technique (HET). Results comparison with those of other techniques such as discrete wavelet transform (DWT), discrete cosine transform (DCT) and singular value decomposition (SVD) have shown an enhance efficiency in terms of ease and performance. It has manifested a good degree of robustness against various environment effects such as resizing, rotation and different kinds of noise. This method would prove very useful technique for copyright protection and ownership judgment.

Keywords: authentication, copyright protection, information hiding, ownership, watermarking

Procedia PDF Downloads 544
2508 Measuring Banks’ Antifragility via Fuzzy Logic

Authors: Danielle Sandler dos Passos, Helder Coelho, Flávia Mori Sarti

Abstract:

Analysing the world banking sector, we realize that traditional risk measurement methodologies no longer reflect the actual scenario with uncertainty and leave out events that can change the dynamics of markets. Considering this, regulators and financial institutions began to search more realistic models. The aim is to include external influences and interdependencies between agents, to describe and measure the operationalization of these complex systems and their risks in a more coherent and credible way. Within this context, X-Events are more frequent than assumed and, with uncertainties and constant changes, the concept of antifragility starts to gain great prominence in comparison to others methodologies of risk management. It is very useful to analyse whether a system succumbs (fragile), resists (robust) or gets benefits (antifragile) from disorder and stress. Thus, this work proposes the creation of the Banking Antifragility Index (BAI), which is based on the calculation of a triangular fuzzy number – to "quantify" qualitative criteria linked to antifragility.

Keywords: adaptive complex systems, X-Events, risk management, antifragility, banking antifragility index, triangular fuzzy number

Procedia PDF Downloads 153