Search results for: probability to breach the SCR
824 Attention Multiple Instance Learning for Cancer Tissue Classification in Digital Histopathology Images
Authors: Afaf Alharbi, Qianni Zhang
Abstract:
The identification of malignant tissue in histopathological slides holds significant importance in both clinical settings and pathology research. This paper introduces a methodology aimed at automatically categorizing cancerous tissue through the utilization of a multiple-instance learning framework. This framework is specifically developed to acquire knowledge of the Bernoulli distribution of the bag label probability by employing neural networks. Furthermore, we put forward a neural network based permutation-invariant aggregation operator, equivalent to attention mechanisms, which is applied to the multi-instance learning network. Through empirical evaluation of an openly available colon cancer histopathology dataset, we provide evidence that our approach surpasses various conventional deep learning methods.Keywords: attention multiple instance learning, MIL and transfer learning, histopathological slides, cancer tissue classification
Procedia PDF Downloads 111823 Mapping Tunnelling Parameters for Global Optimization in Big Data via Dye Laser Simulation
Authors: Sahil Imtiyaz
Abstract:
One of the biggest challenges has emerged from the ever-expanding, dynamic, and instantaneously changing space-Big Data; and to find a data point and inherit wisdom to this space is a hard task. In this paper, we reduce the space of big data in Hamiltonian formalism that is in concordance with Ising Model. For this formulation, we simulate the system using dye laser in FORTRAN and analyse the dynamics of the data point in energy well of rhodium atom. After mapping the photon intensity and pulse width with energy and potential we concluded that as we increase the energy there is also increase in probability of tunnelling up to some point and then it starts decreasing and then shows a randomizing behaviour. It is due to decoherence with the environment and hence there is a loss of ‘quantumness’. This interprets the efficiency parameter and the extent of quantum evolution. The results are strongly encouraging in favour of the use of ‘Topological Property’ as a source of information instead of the qubit.Keywords: big data, optimization, quantum evolution, hamiltonian, dye laser, fermionic computations
Procedia PDF Downloads 195822 On a Single Server Queue with Arrivals in Batches of Variable Size, Generalized Coxian-2 Service and Compulsory Server Vacations
Authors: Kailash C. Madan
Abstract:
We study the steady state behaviour of a batch arrival single server queue in which the first service with general service times is compulsory and the second service with general service times is optional. We term such a two phase service as generalized Coxian-2 service. Just after completion of a service the server must take a vacation of random length of time with general vacation times. We obtain steady state probability generating functions for the queue size as well as the steady state mean queue size at a random epoch of time in explicit and closed forms. Some particular cases of interest including some known results have been derived.Keywords: batch arrivals, compound Poisson process, generalized Coxian-2 service, steady state
Procedia PDF Downloads 457821 Fiber Orientation Measurements in Reinforced Thermoplastics
Authors: Ihsane Modhaffar
Abstract:
Fiber orientation is essential for the physical properties of composite materials. The theoretical parameters of a given reinforcement are usually known and widely used to predict the behavior of the material. In this work, we propose an image processing approach to estimate true principal directions and fiber orientation during injection molding processes of short fiber reinforced thermoplastics. Generally, a group of fibers are described in terms of probability distribution function or orientation tensor. Numerical techniques for the prediction of fiber orientation are also considered for concentrated situations. The flow was considered to be incompressible, and behave as Newtonian fluid containing suspensions of short-fibers. The governing equations, of this problem are: the continuity, the momentum and the energy. The obtained results were compared to available experimental findings. A good agreement between the numerical results and the experimental data was achieved.Keywords: injection, composites, short-fiber reinforced thermoplastics, fiber orientation, incompressible fluid, numerical simulation
Procedia PDF Downloads 533820 [Keynote Talk]: Evidence Fusion in Decision Making
Authors: Mohammad Abdullah-Al-Wadud
Abstract:
In the current era of automation and artificial intelligence, different systems have been increasingly keeping on depending on decision-making capabilities of machines. Such systems/applications may range from simple classifiers to sophisticated surveillance systems based on traditional sensors and related equipment which are becoming more common in the internet of things (IoT) paradigm. However, the available data for such problems are usually imprecise and incomplete, which leads to uncertainty in decisions made based on traditional probability-based classifiers. This requires a robust fusion framework to combine the available information sources with some degree of certainty. The theory of evidence can provide with such a method for combining evidence from different (may be unreliable) sources/observers. This talk will address the employment of the Dempster-Shafer Theory of evidence in some practical applications.Keywords: decision making, dempster-shafer theory, evidence fusion, incomplete data, uncertainty
Procedia PDF Downloads 426819 3D Object Model Reconstruction Based on Polywogs Wavelet Network Parametrization
Authors: Mohamed Othmani, Yassine Khlifi
Abstract:
This paper presents a technique for compact three dimensional (3D) object model reconstruction using wavelet networks. It consists to transform an input surface vertices into signals,and uses wavelet network parameters for signal approximations. To prove this, we use a wavelet network architecture founded on several mother wavelet families. POLYnomials WindOwed with Gaussians (POLYWOG) wavelet families are used to maximize the probability to select the best wavelets which ensure the good generalization of the network. To achieve a better reconstruction, the network is trained several iterations to optimize the wavelet network parameters until the error criterion is small enough. Experimental results will shown that our proposed technique can effectively reconstruct an irregular 3D object models when using the optimized wavelet network parameters. We will prove that an accurateness reconstruction depends on the best choice of the mother wavelets.Keywords: 3d object, optimization, parametrization, polywog wavelets, reconstruction, wavelet networks
Procedia PDF Downloads 285818 Reliability Based Optimal Design of Laterally Loaded Pile with Limited Residual Strain Energy Capacity
Authors: M. Movahedi Rad
Abstract:
In this study, a general approach to the reliability based limit analysis of laterally loaded piles is presented. In engineering practice, the uncertainties play a very important role. The aim of this study is to evaluate the lateral load capacity of free head and fixed-head long pile when the plastic limit analysis is considered. In addition to the plastic limit analysis to control the plastic behaviour of the structure, uncertain bound on the complementary strain energy of the residual forces is also applied. This bound has a significant effect for the load parameter. The solution to reliability-based problems is obtained by a computer program which is governed by the reliability index calculation.Keywords: reliability, laterally loaded pile, residual strain energy, probability, limit analysis
Procedia PDF Downloads 349817 The Relationship of Brand Value and Perceived Brand Quality in the Television Business: A Case Study of Television Viewers in Bangkok
Authors: Natnicha Hasoontree
Abstract:
The purpose of this paper was to study the relationship between brand value and perceived brand quality of television viewers in Bangkok towards the television business in Thailand. The population included television viewers in Bangkok, Thailand. A probability sampling technique was performed to get a sample group that included 500 respondents. Taro Yamane technique was utilized to get a proper sample size. A five Likert scale questionnaire was designed specifically to investigate brand value and perceived brand quality from the perspectives of television viewers in Bangkok. The findings implied that consumers in Bangkok attached a high importance towards the brand equity of television companies that comprised brand ability, brand reputation, brand credibility, and business ethics. Perceived brand quality received high rank in all aspects.Keywords: brand value, perceived brand quality, television business, television viewers
Procedia PDF Downloads 438816 Reliability Analysis of Dam under Quicksand Condition
Authors: Manthan Patel, Vinit Ahlawat, Anshh Singh Claire, Pijush Samui
Abstract:
This paper focuses on the analysis of quicksand condition for a dam foundation. The quicksand condition occurs in cohesion less soil when effective stress of soil becomes zero. In a dam, the saturated sediment may appear quite solid until a sudden change in pressure or shock initiates liquefaction. This causes the sand to form a suspension and lose strength hence resulting in failure of dam. A soil profile shows different properties at different points and the values obtained are uncertain thus reliability analysis is performed. The reliability is defined as probability of safety of a system in a given environment and loading condition and it is assessed as Reliability Index. The reliability analysis of dams under quicksand condition is carried by Gaussian Process Regression (GPR). Reliability index and factor of safety relating to liquefaction of soil is analysed using GPR. The results of reliability analysis by GPR is compared to that of conventional method and it is demonstrated that on applying GPR the probabilistic analysis reduces the computational time and efforts.Keywords: factor of safety, GPR, reliability index, quicksand
Procedia PDF Downloads 483815 Fuzzy Availability Analysis of a Battery Production System
Authors: Merve Uzuner Sahin, Kumru D. Atalay, Berna Dengiz
Abstract:
In today’s competitive market, there are many alternative products that can be used in similar manner and purpose. Therefore, the utility of the product is an important issue for the preferability of the brand. This utility could be measured in terms of its functionality, durability, reliability. These all are affected by the system capabilities. Reliability is an important system design criteria for the manufacturers to be able to have high availability. Availability is the probability that a system (or a component) is operating properly to its function at a specific point in time or a specific period of times. System availability provides valuable input to estimate the production rate for the company to realize the production plan. When considering only the corrective maintenance downtime of the system, mean time between failure (MTBF) and mean time to repair (MTTR) are used to obtain system availability. Also, the MTBF and MTTR values are important measures to improve system performance by adopting suitable maintenance strategies for reliability engineers and practitioners working in a system. Failure and repair time probability distributions of each component in the system should be known for the conventional availability analysis. However, generally, companies do not have statistics or quality control departments to store such a large amount of data. Real events or situations are defined deterministically instead of using stochastic data for the complete description of real systems. A fuzzy set is an alternative theory which is used to analyze the uncertainty and vagueness in real systems. The aim of this study is to present a novel approach to compute system availability using representation of MTBF and MTTR in fuzzy numbers. Based on the experience in the system, it is decided to choose 3 different spread of MTBF and MTTR such as 15%, 20% and 25% to obtain lower and upper limits of the fuzzy numbers. To the best of our knowledge, the proposed method is the first application that is used fuzzy MTBF and fuzzy MTTR for fuzzy system availability estimation. This method is easy to apply in any repairable production system by practitioners working in industry. It is provided that the reliability engineers/managers/practitioners could analyze the system performance in a more consistent and logical manner based on fuzzy availability. This paper presents a real case study of a repairable multi-stage production line in lead-acid battery production factory in Turkey. The following is focusing on the considered wet-charging battery process which has a higher production level than the other types of battery. In this system, system components could exist only in two states, working or failed, and it is assumed that when a component in the system fails, it becomes as good as new after repair. Instead of classical methods, using fuzzy set theory and obtaining intervals for these measures would be very useful for system managers, practitioners to analyze system qualifications to find better results for their working conditions. Thus, much more detailed information about system characteristics is obtained.Keywords: availability analysis, battery production system, fuzzy sets, triangular fuzzy numbers (TFNs)
Procedia PDF Downloads 225814 Modeling Intelligent Threats: Case of Continuous Attacks on a Specific Target
Authors: Asma Ben Yaghlane, Mohamed Naceur Azaiez
Abstract:
In this paper, we treat a model that falls in the area of protecting targeted systems from intelligent threats including terrorism. We introduce the concept of system survivability, in the context of continuous attacks, as the probability that a system under attack will continue operation up to some fixed time t. We define a constant attack rate (CAR) process as an attack on a targeted system that follows an exponential distribution. We consider the superposition of several CAR processes. From the attacker side, we determine the optimal attack strategy that minimizes the system survivability. We also determine the optimal strengthening strategy that maximizes the system survivability under limited defensive resources. We use operations research techniques to identify optimal strategies of each antagonist. Our results may be used as interesting starting points to develop realistic protection strategies against intentional attacks.Keywords: CAR processes, defense/attack strategies, exponential failure, survivability
Procedia PDF Downloads 395813 Exploratory Tests on Structures Resistance during Forest Fires
Authors: Luis M. Ribeiro, Jorge Raposo, Ricardo Oliveira, David Caballero, Domingos X. Viegas
Abstract:
Under the scope of European project WUIWATCH a set of experimental tests on house vulnerability was performed in order to assess the resistance of selected house components during the passage of a forest fire. Among the individual elements most affected by the passage of a wildfire the windows are the ones with greater exposure. In this sense, a set of exploratory experimental tests was designed to assess some particular aspects related to the vulnerability of windows and blinds. At the same time, the importance of leaving them closed (as well as the doors inside a house) during a wild fire was explored in order to give some scientific background to guidelines for homeowners. Three sets of tests were performed: 1. Windows and blinds resistance to heat. Three types of protective blinds were tested (aluminium, PVC and wood) on 2 types of windows (single and double pane). The objective was to assess the structures resistance. 2. The influence of air flow on the transport of burning embers inside a house. A room was built to scale, and placed inside a wind tunnel, with one window and one door on opposite sides. The objective was to assess the importance of leaving an inside door opened on the probability of burning embers entering the room. 3. The influence of the dimension of openings on a window or door related to the probability of ignition inside a house. The objective was to assess the influence of different window openings in relation to the amount of burning particles that can enter a house. The main results were: 1. The purely radiative heat source provides 1.5 KW/m2 of heat impact in the structure, while the real fire generates 10 Kw/m2. When protected by the blind, the single pane window reaches 30ºC on both sides, and the double pane window has a differential of 10º from the side facing the heat (30ºC) and the opposite side (40ºC). Unprotected window constantly increases temperature until the end of the test. Window blinds reach considerably higher temperatures. PVC loses its consistency above 150ºC and melts. 2. Leaving the inside door closed results in a positive pressure differential of +1Pa from the outside to the inside, inhibiting the air flow. Opening the door in half or full reverts the pressure differential to -6 and -8 times respectively, favouring the air flow from the outside to the inside. The number of particles entering the house follows the same tendency. 3. As the bottom opening in a window increases from 0,5 cm to 4 cm the number of particles that enter the house per second also increases greatly. From 5 cm until 80cm there is no substantial increase in the number of entering particles. This set of exploratory tests proved to be an added value in supporting guidelines for home owners, regarding self-protection in WUI areas.Keywords: forest fire, wildland urban interface, house vulnerability, house protective elements
Procedia PDF Downloads 285812 Hypothesis about the Origin of the Lighting
Authors: Igor Kuzminov
Abstract:
Till now, the nature of lightning is not established. A hypothesis of the origin of lightning is proposed. The lightning charge is formed by electromagnetic induction. The role of the conductor is performed by the air mass of the cloud. This conductor moves in the Earth's magnetic field. The upper and lower edges of the cloud are the plates of the capacitor. Lightning is a special case of electromagnetic processes in an atmosphere. The category of lightning occurs in the process of accumulation of a charge. The process of accumulation goes constantly, but the charge is not fixed. Naturally, the hypothesis demands the carrying out of additional experiments and official acknowledgement. As the proof of a hypothesis can serve that the maximal lighting activity in an equatorial zone where cosφ it is close to 1. An experiment conducted privately showed that there is a potential difference in the atmosphere at different levels. The probability of applied value development of power installation is great.Keywords: electromagnetic induction, Earth's magnetic field, plates of the capacitors, charge accumulation
Procedia PDF Downloads 85811 Trace Network: A Probabilistic Relevant Pattern Recognition Approach to Attribution Trace Analysis
Authors: Jian Xu, Xiaochun Yun, Yongzheng Zhang, Yafei Sang, Zhenyu Cheng
Abstract:
Network attack prevention is a critical research area of information security. Network attack would be oppressed if attribution techniques are capable to trace back to the attackers after the hacking event. Therefore attributing these attacks to a particular identification becomes one of the important tasks when analysts attempt to differentiate and profile the attacker behind a piece of attack trace. To assist analysts in expose attackers behind the scenes, this paper researches on the connections between attribution traces and proposes probabilistic relevance based attribution patterns. This method facilitates the evaluation of the plausibility relevance between different traceable identifications. Furthermore, through analyzing the connections among traces, it could confirm the existence probability of a certain organization as well as discover its affinitive partners by the means of drawing relevance matrix from attribution traces.Keywords: attribution trace, probabilistic relevance, network attack, attacker identification
Procedia PDF Downloads 368810 Comparison of Crossover Types to Obtain Optimal Queries Using Adaptive Genetic Algorithm
Authors: Wafa’ Alma'Aitah, Khaled Almakadmeh
Abstract:
this study presents an information retrieval system of using genetic algorithm to increase information retrieval efficiency. Using vector space model, information retrieval is based on the similarity measurement between query and documents. Documents with high similarity to query are judge more relevant to the query and should be retrieved first. Using genetic algorithms, each query is represented by a chromosome; these chromosomes are fed into genetic operator process: selection, crossover, and mutation until an optimized query chromosome is obtained for document retrieval. Results show that information retrieval with adaptive crossover probability and single point type crossover and roulette wheel as selection type give the highest recall. The proposed approach is verified using (242) proceedings abstracts collected from the Saudi Arabian national conference.Keywords: genetic algorithm, information retrieval, optimal queries, crossover
Procedia PDF Downloads 294809 Machine Learning for Aiding Meningitis Diagnosis in Pediatric Patients
Authors: Karina Zaccari, Ernesto Cordeiro Marujo
Abstract:
This paper presents a Machine Learning (ML) approach to support Meningitis diagnosis in patients at a children’s hospital in Sao Paulo, Brazil. The aim is to use ML techniques to reduce the use of invasive procedures, such as cerebrospinal fluid (CSF) collection, as much as possible. In this study, we focus on predicting the probability of Meningitis given the results of a blood and urine laboratory tests, together with the analysis of pain or other complaints from the patient. We tested a number of different ML algorithms, including: Adaptative Boosting (AdaBoost), Decision Tree, Gradient Boosting, K-Nearest Neighbors (KNN), Logistic Regression, Random Forest and Support Vector Machines (SVM). Decision Tree algorithm performed best, with 94.56% and 96.18% accuracy for training and testing data, respectively. These results represent a significant aid to doctors in diagnosing Meningitis as early as possible and in preventing expensive and painful procedures on some children.Keywords: machine learning, medical diagnosis, meningitis detection, pediatric research
Procedia PDF Downloads 151808 Institutional Capacity and Corruption: Evidence from Brazil
Authors: Dalson Figueiredo, Enivaldo Rocha, Ranulfo Paranhos, José Alexandre
Abstract:
This paper analyzes the effects of institutional capacity on corruption. Methodologically, the research design combines both descriptive and multivariate statistics to examine two original datasets based on secondary data. In particular, we employ a principal component model to estimate an indicator of institutional capacity for both state audit institutions and subnational judiciary courts. Then, we estimate the effect of institutional capacity on two dependent variables: (1) incidence of administrative irregularities and (2) time elapsed to judge corruption cases. The preliminary results using ordinary least squares, negative binomial and Tobit models suggest the same conclusions: higher the institutional audit capacity, higher is the probability of detecting a corruption case. On the other hand, higher the institutional capacity of state judiciary, the lower is the time to judge corruption cases.Keywords: institutional capacity, corruption, state level institutions, evidence from Brazil
Procedia PDF Downloads 373807 An Investigation of Prior Educational Achievement on Engineering Student Performance
Authors: Jovanca Smith, Derek Gay
Abstract:
All universities possess a standard by which students are assessed and administered into their programs. This paper considers the effect of the educational history of students, as measured by specific subject grades in Caribbean examinations, on overall performance in introductory engineering math and mechanics courses. Results reflect a correlation between the highest grade in the Caribbean examinations with a higher probability of successful advancement in the university courses. Alternatively, lower entrance grades are commensurate with underperformance in the university courses. Results also demonstrate that students matriculating with the Caribbean examinations will not necessarily possess a significant advantage over students entering through an alternative route, and while previous educational background of students is a significant indicator of tentative performance in the University level math and mechanics courses, it is not the sole factor.Keywords: bimodal distribution, differential learning, engineering education, entrance qualification
Procedia PDF Downloads 363806 Stochastic Variation of the Hubble's Parameter Using Ornstein-Uhlenbeck Process
Authors: Mary Chriselda A
Abstract:
This paper deals with the fact that the Hubble's parameter is not constant and tends to vary stochastically with time. This premise has been proven by converting it to a stochastic differential equation using the Ornstein-Uhlenbeck process. The formulated stochastic differential equation is further solved analytically using the Euler and the Kolmogorov Forward equations, thereby obtaining the probability density function using the Fourier transformation, thereby proving that the Hubble's parameter varies stochastically. This is further corroborated by simulating the observations using Python and R-software for validation of the premise postulated. We can further draw conclusion that the randomness in forces affecting the white noise can eventually affect the Hubble’s Parameter leading to scale invariance and thereby causing stochastic fluctuations in the density and the rate of expansion of the Universe.Keywords: Chapman Kolmogorov forward differential equations, fourier transformation, hubble's parameter, ornstein-uhlenbeck process , stochastic differential equations
Procedia PDF Downloads 202805 Nutrition Bio-Shield Superfood: Healthy and Live Herbal Supplement for Immune System Enhancement
Authors: Azam Bayat, Aref Khalkhali, Ali Reza Mahjoub
Abstract:
Healthy and viable herbal supplement were prepared from wheat by a green route. This organic biomaterial was named Nutrition Bio-shield Superfood (NBS). The NBS supplement had various vitamins, macro and micro molecules, and ingredients. In this study, 20 small Balb/C labile specimens were used in a weighing 30 ± 5 range. The samples were randomly divided into different groups, then the groups were divided into 5 groups. According to the results of this study, the mean number of white blood cells and neutrophil percentage in the experimental group receiving healthy and live dietary supplement showed a significant increase at the 5% probability level in all three groups received 50, 100 and 150 mg/ kg body weight of the mouse compared to the control group. In general, the dietary supplement increases the level of immunity.Keywords: healthy and live herbal supplement, biomaterial, immune system, enhancement
Procedia PDF Downloads 148804 Networked Implementation of Milling Stability Optimization with Bayesian Learning
Authors: Christoph Ramsauer, Jaydeep Karandikar, Tony Schmitz, Friedrich Bleicher
Abstract:
Machining stability is an important limitation to discrete part machining. In this work, a networked implementation of milling stability optimization with Bayesian learning is presented. The milling process was monitored with a wireless sensory tool holder instrumented with an accelerometer at the Vienna University of Technology, Vienna, Austria. The recorded data from a milling test cut is used to classify the cut as stable or unstable based on the frequency analysis. The test cut result is fed to a Bayesian stability learning algorithm at the University of Tennessee, Knoxville, Tennessee, USA. The algorithm calculates the probability of stability as a function of axial depth of cut and spindle speed and recommends the parameters for the next test cut. The iterative process between two transatlantic locations repeats until convergence to a stable optimal process parameter set is achieved.Keywords: machining stability, machine learning, sensor, optimization
Procedia PDF Downloads 206803 Impact of Population Size on Symmetric Travelling Salesman Problem Efficiency
Authors: Wafa' Alsharafat, Suhila Farhan Abu-Owida
Abstract:
Genetic algorithm (GA) is a powerful evolutionary searching technique that is used successfully to solve and optimize problems in different research areas. Genetic Algorithm (GA) considered as one of optimization methods used to solve Travel salesman Problem (TSP). The feasibility of GA in finding a TSP solution is dependent on GA operators; encoding method, population size, termination criteria, in general. In specific, crossover and its probability play a significant role in finding possible solutions for Symmetric TSP (STSP). In addition, the crossover should be determined and enhanced in term reaching optimal or at least near optimal. In this paper, we spot the light on using a modified crossover method called modified sequential constructive crossover and its impact on reaching optimal solution. To justify the relevance of a parameter value in solving the TSP, a set comparative analysis conducted on different crossover methods values.Keywords: genetic algorithm, crossover, mutation, TSP
Procedia PDF Downloads 229802 Lipschitz Classifiers Ensembles: Usage for Classification of Target Events in C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev
Abstract:
This paper introduces an original method for guaranteed estimation of the accuracy of an ensemble of Lipschitz classifiers. The solution was obtained as a finite closed set of alternative hypotheses, which contains an object of classification with a probability of not less than the specified value. Thus, the classification is represented by a set of hypothetical classes. In this case, the smaller the cardinality of the discrete set of hypothetical classes is, the higher is the classification accuracy. Experiments have shown that if the cardinality of the classifiers ensemble is increased then the cardinality of this set of hypothetical classes is reduced. The problem of the guaranteed estimation of the accuracy of an ensemble of Lipschitz classifiers is relevant in the multichannel classification of target events in C-OTDR monitoring systems. Results of suggested approach practical usage to accuracy control in C-OTDR monitoring systems are present.Keywords: Lipschitz classifiers, confidence set, C-OTDR monitoring, classifiers accuracy, classifiers ensemble
Procedia PDF Downloads 492801 Impact of Weather Conditions on Generalized Frequency Division Multiplexing over Gamma Gamma Channel
Authors: Muhammad Sameer Ahmed, Piotr Remlein, Tansal Gucluoglu
Abstract:
The technique called as Generalized frequency division multiplexing (GFDM) used in the free space optical channel can be a good option for implementation free space optical communication systems. This technique has several strengths e.g. good spectral efficiency, low peak-to-average power ratio (PAPR), adaptability and low co-channel interference. In this paper, the impact of weather conditions such as haze, rain and fog on GFDM over the gamma-gamma channel model is discussed. A Trade off between link distance and system performance under intense weather conditions is also analysed. The symbol error probability (SEP) of GFDM over the gamma-gamma turbulence channel is derived and verified with the computer simulations.Keywords: free space optics, generalized frequency division multiplexing, weather conditions, gamma gamma distribution
Procedia PDF Downloads 175800 Study of Factors Linked to Alcohol Consumption among Young People from the Lycée De La Convivialité De Kanyosha in Burundi
Authors: Niyiragira Sixte, Jules Verne Nakimana
Abstract:
Introduction: Alcoholism is gradually becoming a public health issue due to its frequency, which continues to increase, especially in schools and at young ages. The general objective of the study was to contribute to the determination of the factors associated with alcohol consumption among young people. Methodology: This descriptive and analytical cross-sectional study entitled “Study of factors associated with alcohol consumption among young people aged 15 to 24. The study was conducted using a non-probability method, and the sampling technique was for convenience. The data collection technique used was the survey by questionnaire and the exploitation of the documentary. Microsoft Word 2013, Microsoft Excel 2.13 and EPI INFO7 software were used for this purpose. Results: The results of in study showed that 43.36% of the students surveyed took alcohol, and the factors associated with alcohol consumption are: religion, smoking and influence from friends. Conclusion: The prevalence of alcohol consumption among young people is very high, and awareness is more than necessary to prevent alcohol-related harm among young people.Keywords: consumption, alcohol, young people, factors
Procedia PDF Downloads 84799 Unit Root Tests Based On the Robust Estimator
Authors: Wararit Panichkitkosolkul
Abstract:
The unit root tests based on the robust estimator for the first-order autoregressive process are proposed and compared with the unit root tests based on the ordinary least squares (OLS) estimator. The percentiles of the null distributions of the unit root test are also reported. The empirical probabilities of Type I error and powers of the unit root tests are estimated via Monte Carlo simulation. Simulation results show that all unit root tests can control the probability of Type I error for all situations. The empirical power of the unit root tests based on the robust estimator are higher than the unit root tests based on the OLS estimator.
Keywords: autoregressive, ordinary least squares, type i error, power of the test, Monte Carlo simulation
Procedia PDF Downloads 289798 Identification of Bayesian Network with Convolutional Neural Network
Authors: Mohamed Raouf Benmakrelouf, Wafa Karouche, Joseph Rynkiewicz
Abstract:
In this paper, we propose an alternative method to construct a Bayesian Network (BN). This method relies on a convolutional neural network (CNN classifier), which determinates the edges of the network skeleton. We train a CNN on a normalized empirical probability density distribution (NEPDF) for predicting causal interactions and relationships. We have to find the optimal Bayesian network structure for causal inference. Indeed, we are undertaking a search for pair-wise causality, depending on considered causal assumptions. In order to avoid unreasonable causal structure, we consider a blacklist and a whitelist of causality senses. We tested the method on real data to assess the influence of education on the voting intention for the extreme right-wing party. We show that, with this method, we get a safer causal structure of variables (Bayesian Network) and make to identify a variable that satisfies the backdoor criterion.Keywords: Bayesian network, structure learning, optimal search, convolutional neural network, causal inference
Procedia PDF Downloads 178797 Performance Evaluation of an Efficient Asynchronous Protocol for WDM Ring MANs
Authors: Baziana Peristera
Abstract:
The idea of the asynchronous transmission in wavelength division multiplexing (WDM) ring MANs is studied in this paper. Especially, we present an efficient access technique to coordinate the collisions-free transmission of the variable sizes of IP traffic in WDM ring core networks. Each node is equipped with a tunable transmitter and a tunable receiver. In this way, all the wavelengths are exploited for both transmission and reception. In order to evaluate the performance measures of average throughput, queuing delay and packet dropping probability at the buffers, a simulation model that assumes symmetric access rights among the nodes is developed based on Poisson statistics. Extensive numerical results show that the proposed protocol achieves apart from high bandwidth exploitation for a wide range of offered load, fairness of queuing delay and dropping events among the different packets size categories.Keywords: asynchronous transmission, collision avoidance, wavelength division multiplexing, WDM
Procedia PDF Downloads 376796 Rodriguez Diego, Del Valle Martin, Hargreaves Matias, Riveros Jose Luis
Authors: Nathainail Bashir, Neil Anderson
Abstract:
The objective of this study site was to investigate the current state of the practice with regards to karst detection methods and recommend the best method and pattern of arrays to acquire the desire results. Proper site investigation in karst prone regions is extremely valuable in determining the location of possible voids. Two geophysical techniques were employed: multichannel analysis of surface waves (MASW) and electric resistivity tomography (ERT).The MASW data was acquired at each test location using different array lengths and different array orientations (to increase the probability of getting interpretable data in karst terrain). The ERT data were acquired using a dipole-dipole array consisting of 168 electrodes. The MASW data was interpreted (re: estimated depth to physical top of rock) and used to constrain and verify the interpretation of the ERT data. The ERT data indicates poorer quality MASW data were acquired in areas where there was significant local variation in the depth to top of rock.Keywords: dipole-dipole, ERT, Karst terrains, MASW
Procedia PDF Downloads 315795 Expanding the Evaluation Criteria for a Wind Turbine Performance
Authors: Ivan Balachin, Geanette Polanco, Jiang Xingliang, Hu Qin
Abstract:
The problem of global warming raised up interest towards renewable energy sources. To reduce cost of wind energy is a challenge. Before building of wind park conditions such as: average wind speed, direction, time for each wind, probability of icing, must be considered in the design phase. Operation values used on the setting of control systems also will depend on mentioned variables. Here it is proposed a procedure to be include in the evaluation of the performance of a wind turbine, based on the amplitude of wind changes, the number of changes and their duration. A generic study case based on actual data is presented. Data analysing techniques were applied to model the power required for yaw system based on amplitude and data amount of wind changes. A theoretical model between time, amplitude of wind changes and angular speed of nacelle rotation was identified.Keywords: field data processing, regression determination, wind turbine performance, wind turbine placing, yaw system losses
Procedia PDF Downloads 390