Search results for: calculating method of cracking probability
19875 A Phase Field Approach to Model Crack Interface Interaction in Ceramic Matrix Composites
Authors: Dhaladhuli Pranavi, Amirtham Rajagopal
Abstract:
There are various failure modes in ceramic matrix composites; notable ones are fiber breakage, matrix cracking and fiber matrix debonding. Crack nucleation and propagation in microstructure of such composites requires an understanding of interaction of crack with the multiple inclusion heterogeneous system and interfaces. In order to assess structural integrity, the material parameters especially of the interface that governs the crack growth should be determined. In the present work, a nonlocal phase field approach is proposed to model the crack interface interaction in such composites. Nonlocal approaches help in understanding the complex mechanisms of delamination growth and mitigation and operates at a material length scale. The performance of the proposed formulation is illustrated through representative numerical examples. The model proposed is implemented in the framework of the finite element method. Several parametric studies on interface crack interaction are conducted. The proposed model is easy and simple to implement and works very well in modeling fracture in composite systems.Keywords: composite, interface, nonlocal, phase field
Procedia PDF Downloads 14219874 Models of Environmental: Cracker Propagation of Some Aluminum Alloys (7xxx)
Authors: H. Jawan
Abstract:
This review describes the models of environmental-related crack propagation of aluminum alloys (7xxx) during the last few decades. Acknowledge on effects of different factors on the susceptibility to SCC permits to propose valuable mechanisms on crack advancement. The reliable mechanism of cracking give a possibility to propose the optimum chemical composition and thermal treatment conditions resulting in microstructure the most suitable for real environmental condition and stress state.Keywords: microstructure, environmental, propagation, mechanism
Procedia PDF Downloads 39019873 An Investigation of the Fracture Behavior of Model MgO-C Refractories Using the Discrete Element Method
Authors: Júlia Cristina Bonaldo, Christophe L. Martin, Martiniano Piccico, Keith Beale, Roop Kishore, Severine Romero-Baivier
Abstract:
Refractory composite materials employed in steel casting applications are prone to cracking and material damage because of the very high operating temperature (thermal shock) and mismatched properties of the constituent phases. The fracture behavior of a model MgO-C composite refractory is investigated to quantify and characterize its thermal shock resistance, employing a cold crushing test and Brazilian test with fractographic analysis. The discrete element method (DEM) is used to generate numerical refractory composites. The composite in DEM is represented by an assembly of bonded particle clusters forming perfectly spherical aggregates and single spherical particles. For the stresses to converge with a low standard deviation and a minimum number of particles to allow reasonable CPU calculation time, representative volume element (RVE) numerical packings are created with various numbers of particles. Key microscopic properties are calibrated sequentially by comparing stress-strain curves from crushing experimental data. Comparing simulations with experiments also allows for the evaluation of crack propagation, fracture energy, and strength. The crack propagation during Brazilian experimental tests is monitored with digital image correlation (DIC). Simulations and experiments reveal three distinct types of fracture. The crack may spread throughout the aggregate, at the aggregate-matrix interface, or throughout the matrix.Keywords: refractory composite, fracture mechanics, crack propagation, DEM
Procedia PDF Downloads 8119872 A Numerical Study on the Connection of an SC Wall to an RC Foundation
Authors: Siamak Epackachi, Andrew S. Whittaker, Amit H. Varma
Abstract:
There are a large number of methods to connect SC walls to RC foundations. An experimental study of the cyclic nonlinear behavior of SC walls in the NEES laboratory at the University at Buffalo used a connection detail involving the post-tensioning of a steel baseplate to the SC wall to a RC foundation. This type of connection introduces flexibility that influenced substantially the global response of the SC walls. The assumption of a rigid base, which would be commonly made by practitioners, would lead to a substantial overestimation of initial stiffness. This paper presents an analytical approach to characterize the rotational flexibility and to predict the initial stiffness of flexure-critical SC wall piers with baseplate connection. The good agreement between the analytical and test results confirmed the utility of the proposed method for calculating the initial stiffness of an SC wall with baseplate connection.Keywords: steel-plate composite shear wall, flexure-critical wall, cyclic loading, analytical model
Procedia PDF Downloads 34019871 Practical Challenges of Tunable Parameters in Matlab/Simulink Code Generation
Authors: Ebrahim Shayesteh, Nikolaos Styliaras, Alin George Raducu, Ozan Sahin, Daniel Pombo VáZquez, Jonas Funkquist, Sotirios Thanopoulos
Abstract:
One of the important requirements in many code generation projects is defining some of the model parameters tunable. This helps to update the model parameters without performing the code generation again. This paper studies the concept of embedded code generation by MATLAB/Simulink coder targeting the TwinCAT Simulink system. The generated runtime modules are then tested and deployed to the TwinCAT 3 engineering environment. However, defining the parameters tunable in MATLAB/Simulink code generation targeting TwinCAT is not very straightforward. This paper focuses on this subject and reviews some of the techniques tested here to make the parameters tunable in generated runtime modules. Three techniques are proposed for this purpose, including normal tunable parameters, callback functions, and mask subsystems. Moreover, some test Simulink models are developed and used to evaluate the results of proposed approaches. A brief summary of the study results is presented in the following. First of all, the parameters defined tunable and used in defining the values of other Simulink elements (e.g., gain value of a gain block) could be changed after the code generation and this value updating will affect the values of all elements defined based on the values of the tunable parameter. For instance, if parameter K=1 is defined as a tunable parameter in the code generation process and this parameter is used to gain a gain block in Simulink, the gain value for the gain block is equal to 1 in the gain block TwinCAT environment after the code generation. But, the value of K can be changed to a new value (e.g., K=2) in TwinCAT (without doing any new code generation in MATLAB). Then, the gain value of the gain block will change to 2. Secondly, adding a callback function in the form of “pre-load function,” “post-load function,” “start function,” and will not help to make the parameters tunable without performing a new code generation. This means that any MATLAB files should be run before performing the code generation. The parameters defined/calculated in this file will be used as fixed values in the generated code. Thus, adding these files as callback functions to the Simulink model will not make these parameters flexible since the MATLAB files will not be attached to the generated code. Therefore, to change the parameters defined/calculated in these files, the code generation should be done again. However, adding these files as callback functions forces MATLAB to run them before the code generation, and there is no need to define the parameters mentioned in these files separately. Finally, using a tunable parameter in defining/calculating the values of other parameters through the mask is an efficient method to change the value of the latter parameters after the code generation. For instance, if tunable parameter K is used in calculating the value of two other parameters K1 and K2 and, after the code generation, the value of K is updated in TwinCAT environment, the value of parameters K1 and K2 will also be updated (without any new code generation).Keywords: code generation, MATLAB, tunable parameters, TwinCAT
Procedia PDF Downloads 22719870 Ship Detection Requirements Analysis for Different Sea States: Validation on Real SAR Data
Authors: Jaime Martín-de-Nicolás, David Mata-Moya, Nerea del-Rey-Maestre, Pedro Gómez-del-Hoyo, María-Pilar Jarabo-Amores
Abstract:
Ship detection is nowadays quite an important issue in tasks related to sea traffic control, fishery management and ship search and rescue. Although it has traditionally been carried out by patrol ships or aircrafts, coverage and weather conditions and sea state can become a problem. Synthetic aperture radars can surpass these coverage limitations and work under any climatological condition. A fast CFAR ship detector based on a robust statistical modeling of sea clutter with respect to sea states in SAR images is used. In this paper, the minimum SNR required to obtain a given detection probability with a given false alarm rate for any sea state is determined. A Gaussian target model using real SAR data is considered. Results show that SNR does not depend heavily on the class considered. Provided there is some variation in the backscattering of targets in SAR imagery, the detection probability is limited and a post-processing stage based on morphology would be suitable.Keywords: SAR, generalized gamma distribution, detection curves, radar detection
Procedia PDF Downloads 45219869 Text Similarity in Vector Space Models: A Comparative Study
Authors: Omid Shahmirzadi, Adam Lugowski, Kenneth Younge
Abstract:
Automatic measurement of semantic text similarity is an important task in natural language processing. In this paper, we evaluate the performance of different vector space models to perform this task. We address the real-world problem of modeling patent-to-patent similarity and compare TFIDF (and related extensions), topic models (e.g., latent semantic indexing), and neural models (e.g., paragraph vectors). Contrary to expectations, the added computational cost of text embedding methods is justified only when: 1) the target text is condensed; and 2) the similarity comparison is trivial. Otherwise, TFIDF performs surprisingly well in other cases: in particular for longer and more technical texts or for making finer-grained distinctions between nearest neighbors. Unexpectedly, extensions to the TFIDF method, such as adding noun phrases or calculating term weights incrementally, were not helpful in our context.Keywords: big data, patent, text embedding, text similarity, vector space model
Procedia PDF Downloads 17519868 Pyramidal Lucas-Kanade Optical Flow Based Moving Object Detection in Dynamic Scenes
Authors: Hyojin Lim, Cuong Nguyen Khac, Yeongyu Choi, Ho-Youl Jung
Abstract:
In this paper, we propose a simple moving object detection, which is based on motion vectors obtained from pyramidal Lucas-Kanade optical flow. The proposed method detects moving objects such as pedestrians, the other vehicles and some obstacles at the front-side of the host vehicle, and it can provide the warning to the driver. Motion vectors are obtained by using pyramidal Lucas-Kanade optical flow, and some outliers are eliminated by comparing the amplitude of each vector with the pre-defined threshold value. The background model is obtained by calculating the mean and the variance of the amplitude of recent motion vectors in the rectangular shaped local region called the cell. The model is applied as the reference to classify motion vectors of moving objects and those of background. Motion vectors are clustered to rectangular regions by using the unsupervised clustering K-means algorithm. Labeling method is applied to label groups which is close to each other, using by distance between each center points of rectangular. Through the simulations tested on four kinds of scenarios such as approaching motorbike, vehicle, and pedestrians to host vehicle, we prove that the proposed is simple but efficient for moving object detection in parking lots.Keywords: moving object detection, dynamic scene, optical flow, pyramidal optical flow
Procedia PDF Downloads 34919867 Gearbox Defect Detection in the Semi Autogenous Mills Using the Vibration Analysis Technique
Authors: Mostafa Firoozabadi, Alireza Foroughi Nematollahi
Abstract:
Semi autogenous mills are designed for grinding or primary crushed ore, and are the most widely used in concentrators globally. Any defect occurrence in semi autogenous mills can stop the production line. A Gearbox is a significant part of a rotating machine or a mill, so, the gearbox monitoring is a necessary process to prevent the unwanted defects. When a defect happens in a gearbox bearing, this defect can be transferred to the other parts of the equipment like inner ring, outer ring, balls, and the bearing cage. Vibration analysis is one of the most effective and common ways to detect the bearing defects in the mills. Vibration signal in a mill can be made by different parts of the mill including electromotor, pinion girth gear, different rolling bearings, and tire. When a vibration signal, made by the aforementioned parts, is added to the gearbox vibration spectrum, an accurate and on time defect detection in the gearbox will be difficult. In this paper, a new method is proposed to detect the gearbox bearing defects in the semi autogenous mill on time and accurately, using the vibration signal analysis method. In this method, if the vibration values are increased in the vibration curve, the probability of defect occurrence is investigated by comparing the equipment vibration values and the standard ones. Then, all vibration frequencies are extracted from the vibration signal and the equipment defect is detected using the vibration spectrum curve. This method is implemented on the semi autogenous mills in the Golgohar mining and industrial company in Iran. The results show that the proposed method can detect the bearing looseness on time and accurately. After defect detection, the bearing is opened before the equipment failure and the predictive maintenance actions are implemented on it.Keywords: condition monitoring, gearbox defects, predictive maintenance, vibration analysis
Procedia PDF Downloads 46419866 Deformation Mechanisms of Mg-Based Composite Studied by Neutron Diffraction and Acoustic Emission
Authors: G. Farkas, K. Mathis, J. Pilch, P. Minarik
Abstract:
Deformation mechanisms in an Mg-Al-Ca alloy reinforced with short alumina fibres were studied by acoustic emission and in-situ neutron diffraction method. The fibres plane orientation with respect to the loading axis was found to be a key parameter, which influences the acting deformation processes, such as twinning or dislocation slip. In-situ neutron diffraction tests were measured at different temperatures from room temperature (RT) to 200°C. The measurement shows the lattice strain changes in the matrix and also in the reinforcement phase depending on macroscopic compressive deformation and stress. In case of parallel fibre plane orientation, the increment of compressive lattice strain is lower in the matrix and higher in the fibres in comparison to perpendicular fibre orientation. Furthermore, acoustic emission results indicate a larger twinning activity and more frequent fibre cracking in sample with perpendicular fibre plane orientation. Both types of mechanisms are more dominant at elevated temperatures.Keywords: neutron diffraction, acoustic emission, magnesium based composite, deformation mechanisms
Procedia PDF Downloads 16219865 Modeling the Risk Perception of Pedestrians Using a Nested Logit Structure
Authors: Babak Mirbaha, Mahmoud Saffarzadeh, Atieh Asgari Toorzani
Abstract:
Pedestrians are the most vulnerable road users since they do not have a protective shell. One of the most common collisions for them is pedestrian-vehicle at intersections. In order to develop appropriate countermeasures to improve safety for them, researches have to be conducted to identify the factors that affect the risk of getting involved in such collisions. More specifically, this study investigates factors such as the influence of walking alone or having a baby while crossing the street, the observable age of pedestrian, the speed of pedestrians and the speed of approaching vehicles on risk perception of pedestrians. A nested logit model was used for modeling the behavioral structure of pedestrians. The results show that the presence of more lanes at intersections and not being alone especially having a baby while crossing, decrease the probability of taking a risk among pedestrians. Also, it seems that teenagers show more risky behaviors in crossing the street in comparison to other age groups. Also, the speed of approaching vehicles was considered significant. The probability of risk taking among pedestrians decreases by increasing the speed of approaching vehicle in both the first and the second lanes of crossings.Keywords: pedestrians, intersection, nested logit, risk
Procedia PDF Downloads 18619864 Disaster Probability Analysis of Banghabandhu Multipurpose Bridge for Train Accidents and Its Socio-Economic Impact on Bangladesh
Authors: Shahab Uddin, Kazi M. Uddin, Hamamah Sadiqa
Abstract:
The paper deals with the Banghabandhu Multipurpose Bridge (BMB), the 11th longest bridge in the world was constructed in 1998 aimed at contributing to promote economic development in Bangladesh. In recent years, however, the high incidence of traffic accidents and injuries at the bridge sites looms as a great safety concern. Investigation into the derailment of nine bogies out of thirteen of Dinajpur-bound intercity train ‘Drutajan Express ’were derailed and inclined on the Banghabandhu Multipurpose Bridge on 28 April 2014. The train accident in Bridge will be deep concern for both structural safety of bridge and people than other vehicles accident. In this study we analyzed the disaster probability of the Banghabandhu Multipurpose Bridge for accidents by checking the fitness of Bridge structure. We found that train accident impact is more risky than other vehicles accidents. We also found that socio-economic impact on Bangladesh will be deep concerned.Keywords: train accident, derailment, disaster, socio-economic
Procedia PDF Downloads 30219863 Profit-Based Artificial Neural Network (ANN) Trained by Migrating Birds Optimization: A Case Study in Credit Card Fraud Detection
Authors: Ashkan Zakaryazad, Ekrem Duman
Abstract:
A typical classification technique ranks the instances in a data set according to the likelihood of belonging to one (positive) class. A credit card (CC) fraud detection model ranks the transactions in terms of probability of being fraud. In fact, this approach is often criticized, because firms do not care about fraud probability but about the profitability or costliness of detecting a fraudulent transaction. The key contribution in this study is to focus on the profit maximization in the model building step. The artificial neural network proposed in this study works based on profit maximization instead of minimizing the error of prediction. Moreover, some studies have shown that the back propagation algorithm, similar to other gradient–based algorithms, usually gets trapped in local optima and swarm-based algorithms are more successful in this respect. In this study, we train our profit maximization ANN using the Migrating Birds optimization (MBO) which is introduced to literature recently.Keywords: neural network, profit-based neural network, sum of squared errors (SSE), MBO, gradient descent
Procedia PDF Downloads 47519862 Temperature Distribution for Asphalt Concrete-Concrete Composite Pavement
Authors: Tetsya Sok, Seong Jae Hong, Young Kyu Kim, Seung Woo Lee
Abstract:
The temperature distribution for asphalt concrete (AC)-Concrete composite pavement is one of main influencing factor that affects to performance life of pavement. The temperature gradient in concrete slab underneath the AC layer results the critical curling stress and lead to causes de-bonding of AC-Concrete interface. These stresses, when enhanced by repetitive axial loadings, also contribute to the fatigue damage and eventual crack development within the slab. Moreover, the temperature change within concrete slab extremely causes the slab contracts and expands that significantly induces reflective cracking in AC layer. In this paper, the numerical prediction of pavement temperature was investigated using one-dimensional finite different method (FDM) in fully explicit scheme. The numerical predicted model provides a fundamental and clear understanding of heat energy balance including incoming and outgoing thermal energies in addition to dissipated heat in the system. By using the reliable meteorological data for daily air temperature, solar radiation, wind speech and variable pavement surface properties, the predicted pavement temperature profile was validated with the field measured data. Additionally, the effects of AC thickness and daily air temperature on the temperature profile in underlying concrete were also investigated. Based on obtained results, the numerical predicted temperature of AC-Concrete composite pavement using FDM provided a good accuracy compared to field measured data and thicker AC layer significantly insulates the temperature distribution in underlying concrete slab.Keywords: asphalt concrete, finite different method (FDM), curling effect, heat transfer, solar radiation
Procedia PDF Downloads 26919861 D-Care: Diabetes Care Application to Enhance Diabetic Awareness to Diabetes in Indonesia
Authors: Samara R. Dania, Maulana S. Aji, Dewi Lestari
Abstract:
Diabetes is a common disease in Indonesia. One of the risk factors of diabetes is an unhealthy diet which is consuming food that contains too much glucose, one of glucose sources presents in food containing carbohydrate. The purpose of this study is to identify the amount of glucose level in the consumed food. The authors use literature studies for this research method. For the results of this study, the authors expect diabetics to be more aware of diabetes by applying daily dietary regulation through D-Care. D-Care is an application that can enhance people awareness to diabetes in Indonesia. D-Care provides two menus; there are nutrition calculation and healthy food. Nutrition calculation menu is used for knowing estimated glucose intake level by calculating food that consumed each day. Whereas healthy food menu, it provides a combination of healthy food menu for diabetic. The conclusion is D-Care is useful to be used for reducing diabetes prevalence in Indonesia.Keywords: D-Care, diabetes, awareness, healthy food
Procedia PDF Downloads 41919860 Dynamic Reroute Modeling for Emergency Evacuation: Case Study of Brunswick City, Germany
Authors: Yun-Pang Flötteröd, Jakob Erdmann
Abstract:
The human behaviors during evacuations are quite complex. One of the critical behaviors which affect the efficiency of evacuation is route choice. Therefore, the respective simulation modeling work needs to function properly. In this paper, Simulation of Urban Mobility’s (SUMO) current dynamic route modeling during evacuation, i.e. the rerouting functions, is examined with a real case study. The result consistency of the simulation and the reality is checked as well. Four influence factors (1) time to get information, (2) probability to cancel a trip, (3) probability to use navigation equipment, and (4) rerouting and information updating period are considered to analyze possible traffic impacts during the evacuation and to examine the rerouting functions in SUMO. Furthermore, some behavioral characters of the case study are analyzed with use of the corresponding detector data and applied in the simulation. The experiment results show that the dynamic route modeling in SUMO can deal with the proposed scenarios properly. Some issues and function needs related to route choice are discussed and further improvements are suggested.Keywords: evacuation, microscopic traffic simulation, rerouting, SUMO
Procedia PDF Downloads 19419859 Modeling of Glycine Transporters in Mammalian Using the Probability Approach
Authors: K. S. Zaytsev, Y. R. Nartsissov
Abstract:
Glycine is one of the key inhibitory neurotransmitters in Central nervous system (CNS) meanwhile glycinergic transmission is highly dependable on its appropriate reuptake from synaptic cleft. Glycine transporters (GlyT) of types 1 and 2 are the enzymes providing glycine transport back to neuronal and glial cells along with Na⁺ and Cl⁻ co-transport. The distribution and stoichiometry of GlyT1 and GlyT2 differ in details, and GlyT2 is more interesting for the research as it reuptakes glycine to neuron cells, whereas GlyT1 is located in glial cells. In the process of GlyT2 activity, the translocation of the amino acid is accompanied with binding of both one chloride and three sodium ions consequently (two sodium ions for GlyT1). In the present study, we developed a computer simulator of GlyT2 and GlyT1 activity based on known experimental data for quantitative estimation of membrane glycine transport. The trait of a single protein functioning was described using the probability approach where each enzyme state was considered separately. Created scheme of transporter functioning realized as a consequence of elemental steps allowed to take into account each event of substrate association and dissociation. Computer experiments using up-to-date kinetic parameters allowed receiving the number of translocated glycine molecules, Na⁺ and Cl⁻ ions per time period. Flexibility of developed software makes it possible to evaluate glycine reuptake pattern in time under different internal characteristics of enzyme conformational transitions. We investigated the behavior of the system in a wide range of equilibrium constant (from 0.2 to 100), which is not determined experimentally. The significant influence of equilibrium constant in the range from 0.2 to 10 on the glycine transfer process is shown. The environmental conditions such as ion and glycine concentrations are decisive if the values of the constant are outside the specified range.Keywords: glycine, inhibitory neurotransmitters, probability approach, single protein functioning
Procedia PDF Downloads 11919858 Analytical Slope Stability Analysis Based on the Statistical Characterization of Soil Shear Strength
Authors: Bernardo C. P. Albuquerque, Darym J. F. Campos
Abstract:
Increasing our ability to solve complex engineering problems is directly related to the processing capacity of computers. By means of such equipments, one is able to fast and accurately run numerical algorithms. Besides the increasing interest in numerical simulations, probabilistic approaches are also of great importance. This way, statistical tools have shown their relevance to the modelling of practical engineering problems. In general, statistical approaches to such problems consider that the random variables involved follow a normal distribution. This assumption tends to provide incorrect results when skew data is present since normal distributions are symmetric about their means. Thus, in order to visualize and quantify this aspect, 9 statistical distributions (symmetric and skew) have been considered to model a hypothetical slope stability problem. The data modeled is the friction angle of a superficial soil in Brasilia, Brazil. Despite the apparent universality, the normal distribution did not qualify as the best fit. In the present effort, data obtained in consolidated-drained triaxial tests and saturated direct shear tests have been modeled and used to analytically derive the probability density function (PDF) of the safety factor of a hypothetical slope based on Mohr-Coulomb rupture criterion. Therefore, based on this analysis, it is possible to explicitly derive the failure probability considering the friction angle as a random variable. Furthermore, it is possible to compare the stability analysis when the friction angle is modelled as a Dagum distribution (distribution that presented the best fit to the histogram) and as a Normal distribution. This comparison leads to relevant differences when analyzed in light of the risk management.Keywords: statistical slope stability analysis, skew distributions, probability of failure, functions of random variables
Procedia PDF Downloads 33819857 Error Probability of Multi-User Detection Techniques
Authors: Komal Babbar
Abstract:
Multiuser Detection is the intelligent estimation/demodulation of transmitted bits in the presence of Multiple Access Interference. The authors have presented the Bit-error rate (BER) achieved by linear multi-user detectors: Matched filter (which treats the MAI as AWGN), Decorrelating and MMSE. In this work, authors investigate the bit error probability analysis for Matched filter, decorrelating, and MMSE. This problem arises in several practical CDMA applications where the receiver may not have full knowledge of the number of active users and their signature sequences. In particular, the behavior of MAI at the output of the Multi-user detectors (MUD) is examined under various asymptotic conditions including large signal to noise ratio; large near-far ratios; and a large number of users. In the last section Authors also shows Matlab Simulation results for Multiuser detection techniques i.e., Matched filter, Decorrelating, MMSE for 2 users and 10 users.Keywords: code division multiple access, decorrelating, matched filter, minimum mean square detection (MMSE) detection, multiple access interference (MAI), multiuser detection (MUD)
Procedia PDF Downloads 52819856 A Study of Predicting Judgments on Causes of Online Privacy Invasions: Based on U.S Judicial Cases
Authors: Minjung Park, Sangmi Chai, Myoung Jun Lee
Abstract:
Since there are growing concerns on online privacy, enterprises could involve various personal privacy infringements cases resulting legal causations. For companies that are involving online business, it is important for them to pay extra attentions to protect users’ privacy. If firms can aware consequences from possible online privacy invasion cases, they can more actively prevent future online privacy infringements. This study attempts to predict the probability of ruling types caused by various invasion cases under U.S Personal Privacy Act. More specifically, this research explores online privacy invasion cases which was sentenced guilty to identify types of criminal punishments such as penalty, imprisonment, probation as well as compensation in civil cases. Based on the 853 U.S judicial cases ranged from January, 2000 to May, 2016, which related on data privacy, this research examines the relationship between personal information infringements cases and adjudications. Upon analysis results of 41,724 words extracted from 853 regal cases, this study examined online users’ privacy invasion cases to predict the probability of conviction for a firm as an offender in both of criminal and civil law. This research specifically examines that a cause of privacy infringements and a judgment type, whether it leads a civil or criminal liability, from U.S court. This study applies network text analysis (NTA) for data analysis, which is regarded as a useful method to discover embedded social trends within texts. According to our research results, certain online privacy infringement cases caused by online spamming and adware have a high possibility that firms are liable in the case. Our research results provide meaningful insights to academia as well as industry. First, our study is providing a new insight by applying Big Data analytics to legal cases so that it can predict the cause of invasions and legal consequences. Since there are few researches applying big data analytics in the domain of law, specifically in online privacy, this study suggests new area that future studies can explore. Secondly, this study reflects social influences, such as a development of privacy invasion technologies and changes of users’ level of awareness of online privacy on judicial cases analysis by adopting NTA method. Our research results indicate that firms need to improve technical and managerial systems to protect users’ online privacy to avoid negative legal consequences.Keywords: network text analysis, online privacy invasions, personal information infringements, predicting judgements
Procedia PDF Downloads 22919855 Research on Pilot Sequence Design Method of Multiple Input Multiple Output Orthogonal Frequency Division Multiplexing System Based on High Power Joint Criterion
Authors: Linyu Wang, Jiahui Ma, Jianhong Xiang, Hanyu Jiang
Abstract:
For the pilot design of the sparse channel estimation model in Multiple Input Multiple Output Orthogonal Frequency Division Multiplexing (MIMO-OFDM) systems, the observation matrix constructed according to the matrix cross-correlation criterion, total correlation criterion and other optimization criteria are not optimal, resulting in inaccurate channel estimation and high bit error rate at the receiver. This paper proposes a pilot design method combining high-power sum and high-power variance criteria, which can more accurately estimate the channel. First, the pilot insertion position is designed according to the high-power variance criterion under the condition of equal power. Then, according to the high power sum criterion, the pilot power allocation is converted into a cone programming problem, and the power allocation is carried out. Finally, the optimal pilot is determined by calculating the weighted sum of the high power sum and the high power variance. Compared with the traditional pilot frequency, under the same conditions, the constructed MIMO-OFDM system uses the optimal pilot frequency for channel estimation, and the communication bit error rate performance obtains a gain of 6~7dB.Keywords: MIMO-OFDM, pilot optimization, compressed sensing, channel estimation
Procedia PDF Downloads 14919854 Detection of Resistive Faults in Medium Voltage Overhead Feeders
Authors: Mubarak Suliman, Mohamed Hassan
Abstract:
Detection of downed conductors occurring with high fault resistance (reaching kilo-ohms) has always been a challenge, especially in countries like Saudi Arabia, on which earth resistivity is very high in general (reaching more than 1000 Ω-meter). The new approaches for the detection of resistive and high impedance faults are based on the analysis of the fault current waveform. These methods are still under research and development, and they are currently lacking security and dependability. The other approach is communication-based solutions which depends on voltage measurement at the end of overhead line branches and communicate the measured signals to substation feeder relay or a central control center. However, such a detection method is costly and depends on the availability of communication medium and infrastructure. The main objective of this research is to utilize the available standard protection schemes to increase the probability of detection of downed conductors occurring with a low magnitude of fault currents and at the same time avoiding unwanted tripping in healthy conditions and feeders. By specifying the operating region of the faulty feeder, use of tripping curve for discrimination between faulty and healthy feeders, and with proper selection of core balance current transformer (CBCT) and voltage transformers with fewer measurement errors, it is possible to set the pick-up of sensitive earth fault current to minimum values of few amps (i.e., Pick-up Settings = 3 A or 4 A, …) for the detection of earth faults with fault resistance more than (1 - 2 kΩ) for 13.8kV overhead network and more than (3-4) kΩ fault resistance in 33kV overhead network. By implementation of the outcomes of this study, the probability of detection of downed conductors is increased by the utilization of existing schemes (i.e., Directional Sensitive Earth Fault Protection).Keywords: sensitive earth fault, zero sequence current, grounded system, resistive fault detection, healthy feeder
Procedia PDF Downloads 11519853 Optimized Dynamic Bayesian Networks and Neural Verifier Test Applied to On-Line Isolated Characters Recognition
Authors: Redouane Tlemsani, Redouane, Belkacem Kouninef, Abdelkader Benyettou
Abstract:
In this paper, our system is a Markovien system which we can see it like a Dynamic Bayesian Networks. One of the major interests of these systems resides in the complete training of the models (topology and parameters) starting from training data. The Bayesian Networks are representing models of dubious knowledge on complex phenomena. They are a union between the theory of probability and the graph theory in order to give effective tools to represent a joined probability distribution on a set of random variables. The representation of knowledge bases on description, by graphs, relations of causality existing between the variables defining the field of study. The theory of Dynamic Bayesian Networks is a generalization of the Bayesians networks to the dynamic processes. Our objective amounts finding the better structure which represents the relationships (dependencies) between the variables of a dynamic bayesian network. In applications in pattern recognition, one will carry out the fixing of the structure which obliges us to admit some strong assumptions (for example independence between some variables).Keywords: Arabic on line character recognition, dynamic Bayesian network, pattern recognition, networks
Procedia PDF Downloads 61719852 An Analysis of a Queueing System with Heterogeneous Servers Subject to Catastrophes
Authors: M. Reni Sagayaraj, S. Anand Gnana Selvam, R. Reynald Susainathan
Abstract:
This study analyzed a queueing system with blocking and no waiting line. The customers arrive according to a Poisson process and the service times follow exponential distribution. There are two non-identical servers in the system. The queue discipline is FCFS, and the customers select the servers on fastest server first (FSF) basis. The service times are exponentially distributed with parameters μ1 and μ2 at servers I and II, respectively. Besides, the catastrophes occur in a Poisson manner with rate γ in the system. When server I is busy or blocked, the customer who arrives in the system leaves the system without being served. Such customers are called lost customers. The probability of losing a customer was computed for the system. The explicit time dependent probabilities of system size are obtained and a numerical example is presented in order to show the managerial insights of the model. Finally, the probability that arriving customer finds system busy and average number of server busy in steady state are obtained numerically.Keywords: queueing system, blocking, poisson process, heterogeneous servers, queue discipline FCFS, busy period
Procedia PDF Downloads 50319851 A Quasi-Experimental Study of the Impact of 5Es Instructional Model on Students' Mathematics Achievement in Northern Province, Rwanda
Authors: Emmanuel Iyamuremye, Jean François Maniriho, Irenee Ndayambaje
Abstract:
Mathematics is the foundational enabling discipline that underpins science, technology, and engineering disciplines. Science, technology, engineering, and mathematics (STEM) subjects are foreseen as the engine for socio-economic transformation. Rwanda has done reforms in education aiming at empowering and preparing students for the real world job by providing career pathways in science, technology, engineering, and mathematics related fields. While that considered so, the performance in mathematics has remained deplorable in both formative and national examinations. Therefore, this paper aims at exploring the extent to which the engage, explore, explain, elaborate and evaluate (5Es) instructional model contributing towards students’ achievement in mathematics. The present study adopted the pre-test, post-test non-equivalent control group quasi-experimental design. The 5Es instructional model was applied to the experimental group while the control group received instruction with the conventional teaching method for eight weeks. One research-made instrument, mathematics achievement test (MAT), was used for data collection. A pre-test was given to students before the intervention to make sure that both groups have equivalent characteristics. At the end of the experimental period, the two groups have undergone a post-test to ascertain the contribution of the 5Es instructional model. Descriptive statistics and analysis of covariance (ANCOVA) were used for the analysis of the study. For determining the improvement in mathematics, Hakes methods of calculating gain were used to analyze the pre-test and post-test scores. Results showed that students exposed to 5Es instructional model achieved significantly better performance in mathematics than students instructed using the conventional teaching method. It was also found that 5Es instructional model made lessons more interesting, easy and created friendship among students. Thus, 5Es instructional model was recommended to be adopted as a close substitute to the conventional teaching method in teaching mathematics in lower secondary schools in Rwanda.Keywords: 5Es instructional model, achievement, conventional teaching method, mathematics
Procedia PDF Downloads 10319850 Exergy and Energy Analysis of Pre-Heating Unit of Fluid Catalytic Cracking Unit in Kaduna Refining and Petrochemical Company
Authors: M. Nuhu, S. Bilal, A. A. Hamisu, J. A. Abbas, Y. Z. Aminu, P. O. Helen
Abstract:
Exergy and energy analysis of preheating unit of FCCU of KRPC has been calculated and presented in this study. From the design, the efficiency of each heat exchanger was 86%. However, on completion of this work the efficiencies was calculated to be 39.90%, 55.66%, 56.22%, and 57.14% for 16E02, 16E03, 16E04, and 16E05 respectively. 16E04 has the minimum energy loss of 0.86%. The calculated second law and exergy efficiencies of the system were 43.01 and 56.99% respectively.Keywords: exergy analysis, ideal work, efficiency, exergy destruction, temperature
Procedia PDF Downloads 43619849 Improving Fluid Catalytic Cracking Unit Performance through Low Cost Debottlenecking
Authors: Saidulu Gadari, Manoj Kumar Yadav, V. K. Satheesh, Debasis Bhattacharyya, S. S. V. Ramakumar, Subhajit Sarkar
Abstract:
Most Fluid Catalytic Cracking Units (FCCUs) are big profit makers and hence, always operated with several constraints. It is the primary source for production of gasoline, light olefins as petrochemical feedstocks, feedstock for alkylate & oxygenates, LPG, etc. in a refinery. Increasing unit capacity and improving product yields as well as qualities such as gasoline RON have dramatic impact on the refinery economics. FCCUs are often debottlenecked significantly beyond their original design capacities. Depending upon the unit configuration, operating conditions, and feedstock quality, the FCC unit can have a variety of bottlenecks. While some of these are aimed to increase the feed rate, improve the conversion, etc., the others are aimed to improve the reliability of the equipment or overall unit. Apart from investment cost, the other factors considered generally while evaluating the debottlenecking options are shutdown days, faster payback, risk on investment, etc. A low-cost solution such as replacement of feed injectors, air distributor, steam distributors, spent catalyst distributor, efficient cyclone system, etc. are the preferred way of upgrading FCCU. It also has lower lead time from idea inception to implementation. This paper discusses various bottlenecks generally encountered in FCCU and presents a case study on improvement of performance of one of the FCCUs in IndianOil through implementation of cost-effective technical solution including use of improved internals in Reactor-Regeneration (R-R) section. After implementation reduction in regenerator air, gas superficial velocity in regenerator and cyclone velocities by about 10% and improvement of CLO yield from 10 to 6 wt% have been achieved. By ensuring proper pressure balance and optimum immersion of cyclone dipleg in the standpipe, frequent formation of perforations in regenerator cyclones could be addressed which in turn improved the unit on-stream factor.Keywords: FCC, low-cost, revamp, debottleneck, internals, distributors, cyclone, dipleg
Procedia PDF Downloads 21519848 Residual Life Estimation Based on Multi-Phase Nonlinear Wiener Process
Authors: Hao Chen, Bo Guo, Ping Jiang
Abstract:
Residual life (RL) estimation based on multi-phase nonlinear Wiener process was studied in this paper, which is significant for complicated products with small samples. Firstly, nonlinear Wiener model with random parameter was introduced and multi-phase nonlinear Wiener model was proposed to model degradation process of products that were nonlinear and separated into different phases. Then the multi-phase RL probability density function based on the presented model was derived approximately in a closed form and parameters estimation was achieved with the method of maximum likelihood estimation (MLE). Finally, the method was applied to estimate the RL of high voltage plus capacitor. Compared with the other three different models by log-likelihood function (Log-LF) and Akaike information criterion (AIC), the results show that the proposed degradation model can capture degradation process of high voltage plus capacitors in a better way and provide a more reliable result.Keywords: multi-phase nonlinear wiener process, residual life estimation, maximum likelihood estimation, high voltage plus capacitor
Procedia PDF Downloads 45319847 Sizing Residential Solar Power Systems Based on Site-Specific Energy Statistics
Authors: Maria Arechavaleta, Mark Halpin
Abstract:
In the United States, costs of solar energy systems have declined to the point that they are viable options for most consumers. However, there are no consistent procedures for specifying sufficient systems. The factors that must be considered are energy consumption, potential solar energy production, and cost. The traditional method of specifying solar energy systems is based on assumed daily levels of available solar energy and average amounts of daily energy consumption. The mismatches between energy production and consumption are usually mitigated using battery energy storage systems, and energy use is curtailed when necessary. The main consumer decision question that drives the total system cost is how much unserved (or curtailed) energy is acceptable? Of course additional solar conversion equipment can be installed to provide greater peak energy production and extra energy storage capability can be added to mitigate longer lasting low solar energy production periods. Each option increases total cost and provides a benefit which is difficult to quantify accurately. An approach to quantify the cost-benefit of adding additional resources, either production or storage or both, based on the statistical concepts of loss-of-energy probability and expected unserved energy, is presented in this paper. Relatively simple calculations, based on site-specific energy availability and consumption data, can be used to show the value of each additional increment of production or storage. With this incremental benefit-cost information, consumers can select the best overall performance combination for their application at a cost they are comfortable paying. The approach is based on a statistical analysis of energy consumption and production characteristics over time. The characteristics are in the forms of curves with each point on the curve representing an energy consumption or production value over a period of time; a one-minute period is used for the work in this paper. These curves are measured at the consumer location under the conditions that exist at the site and the duration of the measurements is a minimum of one week. While greater accuracy could be obtained with longer recording periods, the examples in this paper are based on a single week for demonstration purposes. The weekly consumption and production curves are overlaid on each other and the mismatches are used to size the battery energy storage system. Loss-of-energy probability and expected unserved energy indices are calculated in addition to the total system cost. These indices allow the consumer to recognize and quantify the benefit (probably a reduction in energy consumption curtailment) available for a given increase in cost. Consumers can then make informed decisions that are accurate for their location and conditions and which are consistent with their available funds.Keywords: battery energy storage systems, loss of load probability, residential renewable energy, solar energy systems
Procedia PDF Downloads 23419846 Automated Method Time Measurement System for Redesigning Dynamic Facility Layout
Authors: Salam Alzubaidi, G. Fantoni, F. Failli, M. Frosolini
Abstract:
The dynamic facility layout problem is a really critical issue in the competitive industrial market; thus, solving this problem requires robust design and effective simulation systems. The sustainable simulation requires inputting reliable and accurate data into the system. So this paper describes an automated system integrated into the real environment to measure the duration of the material handling operations, collect the data in real-time, and determine the variances between the actual and estimated time schedule of the operations in order to update the simulation software and redesign the facility layout periodically. The automated method- time measurement system collects the real data through using Radio Frequency-Identification (RFID) and Internet of Things (IoT) technologies. Hence, attaching RFID- antenna reader and RFID tags enables the system to identify the location of the objects and gathering the time data. The real duration gathered will be manipulated by calculating the moving average duration of the material handling operations, choosing the shortest material handling path, and then updating the simulation software to redesign the facility layout accommodating with the shortest/real operation schedule. The periodic simulation in real-time is more sustainable and reliable than the simulation system relying on an analysis of historical data. The case study of this methodology is in cooperation with a workshop team for producing mechanical parts. Although there are some technical limitations, this methodology is promising, and it can be significantly useful in the redesigning of the manufacturing layout.Keywords: dynamic facility layout problem, internet of things, method time measurement, radio frequency identification, simulation
Procedia PDF Downloads 120