Search results for: random flows
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2805

Search results for: random flows

2535 Investigating the Efficiency of Stratified Double Median Ranked Set Sample for Estimating the Population Mean

Authors: Mahmoud I. Syam

Abstract:

Stratified double median ranked set sampling (SDMRSS) method is suggested for estimating the population mean. The SDMRSS is compared with the simple random sampling (SRS), stratified simple random sampling (SSRS), and stratified ranked set sampling (SRSS). It is shown that SDMRSS estimator is an unbiased of the population mean and more efficient than SRS, SSRS, and SRSS. Also, by SDMRSS, we can increase the efficiency of mean estimator for specific value of the sample size. SDMRSS is applied on real life examples, and the results of the example agreed the theoretical results.

Keywords: efficiency, double ranked set sampling, median ranked set sampling, ranked set sampling, stratified

Procedia PDF Downloads 223
2534 An Analysis on Gravel of Sand-Gravel Bar at Gneiss or Granite Area of the Upper Hongcheon River in South Korea

Authors: Man Kyu Kim, Hansu Shin

Abstract:

This study is an analysis on gravel of sand-gravel bar that stretches variously in the Duchon and Naechon stream basins, which are situated on Hong-Cheon River (a well-developed sand-gravel bar in upstream river) basins in Korea. Naechon stream mostly flows through granite zone but Duchon stream mostly flows through gneiss zone. The characteristics of gravel in the sand-gravel bar of these two branches in the upper Hongcheon River were analyzed in this study in order to understand the geomorphic development of streams depending on the differences of bedrock. Through the analysis on the roundness and flatness of gravel, we figured out an irregular trend following the increase in supply of granite gravel and gneiss gravel as we traveled downstream. The result shows that the two basins have uppermost small basin condition reflecting the mountain valley environment although it may be difficult to do an equivalent comparison to other roundness researches in Korea or in Europe. This study conducted an analysis on gravels found in small scale streams unlike the previous studies trend which mostly studies large rivers. The research provides an opportunity to offer basic data for continuous comparison research on various small basins.

Keywords: flatness, geology, roundness, sand-gravel bar

Procedia PDF Downloads 340
2533 Hydrodynamic Characteristics of Single and Twin Offshore Rubble Mound Breakwaters under Regular and Random Waves

Authors: M. Alkhalidi, S. Neelamani, Z. Al-Zaqah

Abstract:

This paper investigates the interaction of single and twin offshore rubble mound breakwaters with regular and random water waves through physical modeling to assess their reflection, transmission and energy dissipation characteristics. Various combinations of wave heights and wave periods were utilized in a series of experiments, along with three different water depths. The single and twin permeable breakwater models were both constructed with one layer of rubbles. Both models had the same total volume; however, the single breakwater was of trapezoidal type while the twin breakwaters were of triangular type. Physical modeling experiments were carried out in the wave flume of the coastal engineering laboratory of Kuwait Institute for Scientific Research (KISR). Measurements of the six wave probes which were fixed in the two-dimensional wave flume were collected and used to determine the generated incident wave heights, as well as the reflected and transmitted wave heights resulting from the wave-breakwater interaction. The possible factors affecting the wave attenuation efficiency of the breakwater models are the relative water depth (d/L), wave steepness (H/L), relative wave height ((h-d)/Hi), relative height of the breakwater (h/d), and relative clear spacing between the twin breakwaters (S/h). The results indicated that the single and double breakwaters show different responds to the change in their relative height as well as the relative wave height which demonstrates that the effect of the relative water depth on wave reflection, transmission, and energy dissipation is highly influenced by the change in the relative breakwater height, the relative wave height and the relative breakwater spacing. In general, within the range of the relative water depth tested in this study, and under both regular and random waves, it is found that the single breakwater allows for lower wave transmission and shows higher energy dissipation effect than both of the tested twin breakwaters, and hence has the best overall performance.

Keywords: random waves, regular waves, relative water depth, relative wave height, single breakwater, twin breakwater, wave steepness

Procedia PDF Downloads 284
2532 Facial Recognition on the Basis of Facial Fragments

Authors: Tetyana Baydyk, Ernst Kussul, Sandra Bonilla Meza

Abstract:

There are many articles that attempt to establish the role of different facial fragments in face recognition. Various approaches are used to estimate this role. Frequently, authors calculate the entropy corresponding to the fragment. This approach can only give approximate estimation. In this paper, we propose to use a more direct measure of the importance of different fragments for face recognition. We propose to select a recognition method and a face database and experimentally investigate the recognition rate using different fragments of faces. We present two such experiments in the paper. We selected the PCNC neural classifier as a method for face recognition and parts of the LFW (Labeled Faces in the Wild) face database as training and testing sets. The recognition rate of the best experiment is comparable with the recognition rate obtained using the whole face.

Keywords: face recognition, labeled faces in the wild (LFW) database, random local descriptor (RLD), random features

Procedia PDF Downloads 331
2531 Using Combination of Different Sets of Features of Molecules for Improved Prediction of Solubility

Authors: Muhammet Baldan, Emel Timuçin

Abstract:

Generally, absorption and bioavailability increase if solubility increases; therefore, it is crucial to predict them in drug discovery applications. Molecular descriptors and Molecular properties are traditionally used for the prediction of water solubility. There are various key descriptors that are used for this purpose, namely Drogan Descriptors, Morgan Descriptors, Maccs keys, etc., and each has different prediction capabilities with differentiating successes between different data sets. Another source for the prediction of solubility is structural features; they are commonly used for the prediction of solubility. However, there are little to no studies that combine three or more properties or descriptors for prediction to produce a more powerful prediction model. Unlike available models, we used a combination of those features in a random forest machine learning model for improved solubility prediction to better predict and, therefore, contribute to drug discovery systems.

Keywords: solubility, molecular descriptors, machine learning, random forest

Procedia PDF Downloads 25
2530 On the Use of Analytical Performance Models to Design a High-Performance Active Queue Management Scheme

Authors: Shahram Jamali, Samira Hamed

Abstract:

One of the open issues in Random Early Detection (RED) algorithm is how to set its parameters to reach high performance for the dynamic conditions of the network. Although original RED uses fixed values for its parameters, this paper follows a model-based approach to upgrade performance of the RED algorithm. It models the routers queue behavior by using the Markov model and uses this model to predict future conditions of the queue. This prediction helps the proposed algorithm to make some tunings over RED's parameters and provide efficiency and better performance. Widespread packet level simulations confirm that the proposed algorithm, called Markov-RED, outperforms RED and FARED in terms of queue stability, bottleneck utilization and dropped packets count.

Keywords: active queue management, RED, Markov model, random early detection algorithm

Procedia PDF Downloads 516
2529 Experimental Investigation of Partially Premixed Laminar Methane/Air Co-Flow Flames Using Mach-Zehnder Interferometry

Authors: Misagh Irandoost Shahrestani, Mehdi Ashjaee, Shahrokh Zandieh Vakili

Abstract:

In this paper, partially premixed laminar methane/air co-flow flame is studied experimentally. Methane-air flame was established on an axisymmetric coannular burner. The fuel-air jet flows from the central tube while the secondary air flows from the region between the inner and the outer tube. The aim is to investigate the flame features and to develop a nonintrusive method for temperature measurement of methane/air partially premixed flame using Mach-Zehnder interferometry method. Different equivalence ratios and Reynolds numbers are considered. Flame generic visible appearance was also investigated and its various structures were studied. Three distinguished flame regimes were seen based on its appearance. A double flame structure can be seen for the equivalence ratio in the range of 1<Φ<2.1. By adding air to the mixture up to Φ=4 the flame has the characteristics of both premixed and non-premixed flames. Finally for 4<Φ<∞ the flame mainly becomes non-premixed like and the luminous sooting region on its tip is the obvious feature of this type of flames. The Mach-Zehnder method is used to obtain temperature field of a transparent fluid by means of index of refraction. Temperature obtained from optical techniques was compared with that of obtained from thermocouples in order to validate the results. Good agreement was observed for these two methods.

Keywords: flame structure, Mach-Zehnder interferometry, partially premixed flame, temperature field

Procedia PDF Downloads 440
2528 Sensitivity Analysis of Principal Stresses in Concrete Slab of Rigid Pavement Made From Recycled Materials

Authors: Aleš Florian, Lenka Ševelová

Abstract:

Complex sensitivity analysis of stresses in a concrete slab of the real type of rigid pavement made from recycled materials is performed. The computational model of the pavement is designed as a spatial (3D) model, is based on a nonlinear variant of the finite element method that respects the structural nonlinearity, enables to model different arrangements of joints, and the entire model can be loaded by the thermal load. Interaction of adjacent slabs in joints and contact of the slab and the subsequent layer are modeled with the help of special contact elements. Four concrete slabs separated by transverse and longitudinal joints and the additional structural layers and soil to the depth of about 3m are modeled. The thickness of individual layers, physical and mechanical properties of materials, characteristics of joints, and the temperature of the upper and lower surface of slabs are supposed to be random variables. The modern simulation technique Updated Latin Hypercube Sampling with 20 simulations is used. For sensitivity analysis the sensitivity coefficient based on the Spearman rank correlation coefficient is utilized. As a result, the estimates of influence of random variability of individual input variables on the random variability of principal stresses s1 and s3 in 53 points on the upper and lower surface of the concrete slabs are obtained.

Keywords: concrete, FEM, pavement, sensitivity, simulation

Procedia PDF Downloads 298
2527 Low Cost Inertial Sensors Modeling Using Allan Variance

Authors: A. A. Hussen, I. N. Jleta

Abstract:

Micro-electromechanical system (MEMS) accelerometers and gyroscopes are suitable for the inertial navigation system (INS) of many applications due to the low price, small dimensions and light weight. The main disadvantage in a comparison with classic sensors is a worse long term stability. The estimation accuracy is mostly affected by the time-dependent growth of inertial sensor errors, especially the stochastic errors. In order to eliminate negative effect of these random errors, they must be accurately modeled. Where the key is the successful implementation that depends on how well the noise statistics of the inertial sensors is selected. In this paper, the Allan variance technique will be used in modeling the stochastic errors of the inertial sensors. By performing a simple operation on the entire length of data, a characteristic curve is obtained whose inspection provides a systematic characterization of various random errors contained in the inertial-sensor output data.

Keywords: Allan variance, accelerometer, gyroscope, stochastic errors

Procedia PDF Downloads 409
2526 Coupling Random Demand and Route Selection in the Transportation Network Design Problem

Authors: Shabnam Najafi, Metin Turkay

Abstract:

Network design problem (NDP) is used to determine the set of optimal values for certain pre-specified decision variables such as capacity expansion of nodes and links by optimizing various system performance measures including safety, congestion, and accessibility. The designed transportation network should improve objective functions defined for the system by considering the route choice behaviors of network users at the same time. The NDP studies mostly investigated the random demand and route selection constraints separately due to computational challenges. In this work, we consider both random demand and route selection constraints simultaneously. This work presents a nonlinear stochastic model for land use and road network design problem to address the development of different functional zones in urban areas by considering both cost function and air pollution. This model minimizes cost function and air pollution simultaneously with random demand and stochastic route selection constraint that aims to optimize network performance via road capacity expansion. The Bureau of Public Roads (BPR) link impedance function is used to determine the travel time function in each link. We consider a city with origin and destination nodes which can be residential or employment or both. There are set of existing paths between origin-destination (O-D) pairs. Case of increasing employed population is analyzed to determine amount of roads and origin zones simultaneously. Minimizing travel and expansion cost of routes and origin zones in one side and minimizing CO emission in the other side is considered in this analysis at the same time. In this work demand between O-D pairs is random and also the network flow pattern is subject to stochastic user equilibrium, specifically logit route choice model. Considering both demand and route choice, random is more applicable to design urban network programs. Epsilon-constraint is one of the methods to solve both linear and nonlinear multi-objective problems. In this work epsilon-constraint method is used to solve the problem. The problem was solved by keeping first objective (cost function) as the objective function of the problem and second objective as a constraint that should be less than an epsilon, where epsilon is an upper bound of the emission function. The value of epsilon should change from the worst to the best value of the emission function to generate the family of solutions representing Pareto set. A numerical example with 2 origin zones and 2 destination zones and 7 links is solved by GAMS and the set of Pareto points is obtained. There are 15 efficient solutions. According to these solutions as cost function value increases, emission function value decreases and vice versa.

Keywords: epsilon-constraint, multi-objective, network design, stochastic

Procedia PDF Downloads 617
2525 Classification of Contexts for Mentioning Love in Interviews with Victims of the Holocaust

Authors: Marina Yurievna Aleksandrova

Abstract:

Research of the Holocaust retains value not only for history but also for sociology and psychology. One of the most important fields of study is how people were coping during and after this traumatic event. The aim of this paper is to identify the main contexts of the topic of love and to determine which contexts are more characteristic for different groups of victims of the Holocaust (gender, nationality, age). In this research, transcripts of interviews with Holocaust victims that were collected during 1946 for the "Voices of the Holocaust" project were used as data. Main contexts were analyzed with methods of network analysis and latent semantic analysis and classified by gender, age, and nationality with random forest. The results show that love is articulated and described significantly differently for male and female informants, nationality is shown results with lower values of quality metrics, as well as the age.

Keywords: Holocaust, latent semantic analysis, network analysis, text-mining, random forest

Procedia PDF Downloads 159
2524 Catchment Yield Prediction in an Ungauged Basin Using PyTOPKAPI

Authors: B. S. Fatoyinbo, D. Stretch, O. T. Amoo, D. Allopi

Abstract:

This study extends the use of the Drainage Area Regionalization (DAR) method in generating synthetic data and calibrating PyTOPKAPI stream yield for an ungauged basin at a daily time scale. The generation of runoff in determining a river yield has been subjected to various topographic and spatial meteorological variables, which integers form the Catchment Characteristics Model (CCM). Many of the conventional CCM models adapted in Africa have been challenged with a paucity of adequate, relevance and accurate data to parameterize and validate the potential. The purpose of generating synthetic flow is to test a hydrological model, which will not suffer from the impact of very low flows or very high flows, thus allowing to check whether the model is structurally sound enough or not. The employed physically-based, watershed-scale hydrologic model (PyTOPKAPI) was parameterized with GIS-pre-processing parameters and remote sensing hydro-meteorological variables. The validation with mean annual runoff ratio proposes a decent graphical understanding between observed and the simulated discharge. The Nash-Sutcliffe efficiency and coefficient of determination (R²) values of 0.704 and 0.739 proves strong model efficiency. Given the current climate variability impact, water planner can now assert a tool for flow quantification and sustainable planning purposes.

Keywords: catchment characteristics model, GIS, synthetic data, ungauged basin

Procedia PDF Downloads 298
2523 Rounding Technique's Application in Schnorr Signature Algorithm: Known Partially Most Significant Bits of Nonce

Authors: Wenjie Qin, Kewei Lv

Abstract:

In 1996, Boneh and Venkatesan proposed the Hidden Number Problem (HNP) and proved the most significant bits (MSB) of computational Diffie-Hellman key exchange scheme and related schemes are unpredictable bits. They also gave a method which is a lattice rounding technique to solve HNP in non-uniform model. In this paper, we put forward a new concept that is Schnorr-MSB-HNP. We also reduce the problem of solving Schnorr signature private key with a few consecutive most significant bits of random nonce (used at each signature generation) to Schnorr-MSB-HNP, then we use the rounding technique to solve the Schnorr-MSB-HNP. We have come to the conclusion that if there is a ‘miraculous box’ which inputs the random nonce and outputs 2loglogq (q is a prime number) most significant bits of nonce, the signature private key will be obtained by choosing 2logq signature messages randomly. Thus we get an attack on the Schnorr signature private key.

Keywords: rounding technique, most significant bits, Schnorr signature algorithm, nonce, Schnorr-MSB-HNP

Procedia PDF Downloads 203
2522 Efficient Credit Card Fraud Detection Based on Multiple ML Algorithms

Authors: Neha Ahirwar

Abstract:

In the contemporary digital era, the rise of credit card fraud poses a significant threat to both financial institutions and consumers. As fraudulent activities become more sophisticated, there is an escalating demand for robust and effective fraud detection mechanisms. Advanced machine learning algorithms have become crucial tools in addressing this challenge. This paper conducts a thorough examination of the design and evaluation of a credit card fraud detection system, utilizing four prominent machine learning algorithms: random forest, logistic regression, decision tree, and XGBoost. The surge in digital transactions has opened avenues for fraudsters to exploit vulnerabilities within payment systems. Consequently, there is an urgent need for proactive and adaptable fraud detection systems. This study addresses this imperative by exploring the efficacy of machine learning algorithms in identifying fraudulent credit card transactions. The selection of random forest, logistic regression, decision tree, and XGBoost for scrutiny in this study is based on their documented effectiveness in diverse domains, particularly in credit card fraud detection. These algorithms are renowned for their capability to model intricate patterns and provide accurate predictions. Each algorithm is implemented and evaluated for its performance in a controlled environment, utilizing a diverse dataset comprising both genuine and fraudulent credit card transactions.

Keywords: efficient credit card fraud detection, random forest, logistic regression, XGBoost, decision tree

Procedia PDF Downloads 29
2521 Behavior of Droplets in Microfluidic System with T-Junction

Authors: A. Guellati, F-M Lounis, N. Guemras, K. Daoud

Abstract:

Micro droplet formation is considered as a growing emerging area of research due to its wide-range application in chemistry as well as biology. The mechanism of micro droplet formation using two immiscible liquids running through a T-junction has been widely studied. We believe that the flow of these two immiscible phases can be of greater important factor that could have an impact on out-flow hydrodynamic behavior, the droplets generated and the size of the droplets. In this study, the type of the capillary tubes used also represents another important factor that can have an impact on the generation of micro droplets. The tygon capillary tubing with hydrophilic inner surface doesn't allow regular out-flows due to the fact that the continuous phase doesn't adhere to the wall of the capillary inner surface. Teflon capillary tubing, presents better wettability than tygon tubing, and allows to obtain steady and regular regimes of out-flow, and the micro droplets are homogeneoussize. The size of the droplets is directly dependent on the flows of the continuous and dispersed phases. Thus, as increasing the flow of the continuous phase, to flow of the dispersed phase stationary, the size of the drops decreases. Inversely, while increasing the flow of the dispersed phase, to flow of the continuous phase stationary, the size of the droplet increases.

Keywords: microfluidic system, micro droplets generation, t-junction, fluids engineering

Procedia PDF Downloads 316
2520 The Impact of Size of the Regional Economic Blocs to the Country’s Flows of Trade: Evidence from COMESA, EAC and Tanzania

Authors: Mosses E. Lufuke, Lorna M. Kamau

Abstract:

This paper attempted to assess whether the size of the regional economic bloc has an impact to the flow of trade to a particular country. Two different sized blocs (COMESA and EAC) and one country (Tanzania) have been used as the point of references. Using the results from of the analyses, the paper also was anticipated to establish whether it was rational for Tanzania to withdraw its membership from COMESA (the larger bloc) to join EAC (the small one). Gravity model has been used to estimate the relationship between the variables, from which the bilateral trade flows between Tanzania and the eighteen member countries of the two blocs (COMESA and EAC) was employed for the time between 2000 and 2013. In the model, the dummy variable for regional bloc (bloc) at which the Tanzania trade partner countries belong are also added to the model to understand which trade bloc exhibit higher trade flow with Tanzania. From the findings, it was noted that over the period of study (2000-2013) Tanzania acknowledged more than 257% of trade volume in EAC than in COMESA. Conclusive, it was noted that the flow of trade is explained by many other variables apart from the size of regional bloc; and that the size by itself offer insufficient evidence in causality relationship. The paper therefore remain neutral on such staggered switching decision since more analyses are required to establish the country’s trade flow, especially when if it had been in multiple membership of COMESA and EAC.

Keywords: economic bloc, flow of trade, size of bloc, switching

Procedia PDF Downloads 226
2519 Seismo-Volcanic Hazards in Great Ararat Region, Eastern Turkey

Authors: Mehmet Salih Bayraktutan, Emre Tokmak

Abstract:

Great Ararat Volcano is the highest peak in South Caucasus Volcanic Plateau. Uplifted by Quaternary basaltic pyroclastic and lava flows. Numerous volcanic cones formed along with the tensional fractures under N-S compressional geodynamic framework. Basaltic flows have fresh surface morphology give ages of 650-680 K years. Hyperstene andesites constitute a major mass of Greater Ararat gives ages of 450-490 K years. During the early eruption period, predominately pyroclastics, cinder, lapilly-ash volcanic bombs were extruded. Third-period eruptions dominantly basaltic lava flows. Andesitic domes aligned along with the NW-SE striking fractures. Hyalo basalt and hornblende basaltic lavas are the latest lava eruptions. Hyalo-basaltic eruptions occurred via parasitic cones distributed far from the center. Parasitic cones are most common at the foot of Mount covered by recent NW flowing basaltic lava. Some of the cones are distributed on a circular pattern. One of the most hazardous disasters recorded in Eastern Turkey was July 1840 Cehennem Canyon Flood. Volcanic activities seismically triggered resulted in melting of glacier cap, mixed with ash and pyroclastics, flowed down along the Valley. Mud rich Slush urged catastrophically northwards, crossed Ars River and damned Surmeli Basin, forming reservoir behind. Ararat volcanoes are located on NW-SE striking Agri Fault Zone. Right lateral extensional faults, along which a series of andesitic domes formed. Great Ararat, in general strato-type volcano. This huge structure, developed in two main parts with different topographic and morphological features. The large lower base covers a widespread area composed of predominantly pyroclastics, ignimbrites, aglomerates, thick pumice, perlite deposits. Approximately 1/3 of the Crest by height formed of this basement. And 2/3 of the upper part with a conic- shape composed of basaltic lava flows. The active tectonic structure consists of three different patterns. The first network is radially distributed fractures formed during the last stage of lava eruptions. The second group of active faults striking in NW direction, and continue in N30W strike, formes Igdir Fault Zone. The third set of faults, dipping in the northwest with 75-80 degrees, strikes NE- SW across the whole Mount, slicing Great Ararat into four segments. In the upper stage of Cehennem Canyon, this set cutting volcanic layers caused numerous Waterfalls, Rock Avalanches, Mud Flows along the canyon, threatens the Village of Yanidogan, at the apex of flood deposits. Great Ararat Region has high seismo-tectonic risk and by occurrence frequency and magnitude, which caused in history caused heavy disasters, at villages surrounding the Ararat Basement.

Keywords: Eastern Turkey, geohazard, great ararat volcano, seismo-tectonic features

Procedia PDF Downloads 162
2518 On a Single Server Queue with Arrivals in Batches of Variable Size, Generalized Coxian-2 Service and Compulsory Server Vacations

Authors: Kailash C. Madan

Abstract:

We study the steady state behaviour of a batch arrival single server queue in which the first service with general service times is compulsory and the second service with general service times is optional. We term such a two phase service as generalized Coxian-2 service. Just after completion of a service the server must take a vacation of random length of time with general vacation times. We obtain steady state probability generating functions for the queue size as well as the steady state mean queue size at a random epoch of time in explicit and closed forms. Some particular cases of interest including some known results have been derived.

Keywords: batch arrivals, compound Poisson process, generalized Coxian-2 service, steady state

Procedia PDF Downloads 429
2517 Mean Square Responses of a Cantilever Beam with Various Damping Mechanisms

Authors: Yaping Zhao, Yimin Zhang

Abstract:

In the present paper, the stationary random vibration of a uniform cantilever beam is investigated. Two types of damping mechanism, i.e. the external and internal viscous dampings, are taken into account simultaneously. The excitation form is the support motion, and it is ideal white. Because two type of damping mechanism are considered concurrently, the product of the modal damping ratio and the natural frequency is not a constant anymore. As a result, the infinite definite integral encountered in the process of computing the mean square response is more complex than that in the existing literature. One signal progress of this work is to have calculated these definite integrals accurately. The precise solution of the mean square response is thus obtained in the infinite series form finally. Numerical examples are supplied and the numerical outcomes acquired confirm the validity of the theoretical analyses.

Keywords: random vibration, cantilever beam, mean square response, white noise

Procedia PDF Downloads 366
2516 Evaluation of Reliability Indices Using Monte Carlo Simulation Accounting Time to Switch

Authors: Sajjad Asefi, Hossein Afrakhte

Abstract:

This paper presents the evaluation of reliability indices of an electrical distribution system using Monte Carlo simulation technique accounting Time To Switch (TTS) for each section. In this paper, the distribution system has been assumed by accounting random repair time omission. For simplicity, we have assumed the reliability analysis to be based on exponential law. Each segment has a specified rate of failure (λ) and repair time (r) which will give us the mean up time and mean down time of each section in distribution system. After calculating the modified mean up time (MUT) in years, mean down time (MDT) in hours and unavailability (U) in h/year, TTS have been added to the time which the system is not available, i.e. MDT. In this paper, we have assumed the TTS to be a random variable with Log-Normal distribution.

Keywords: distribution system, Monte Carlo simulation, reliability, repair time, time to switch (TTS)

Procedia PDF Downloads 403
2515 Deformation of Particle-Laden Droplet in Viscous Liquid under DC Electric Fields

Authors: Khobaib Khobaib, Alexander Mikkelsen, Zbigniew Rozynek

Abstract:

Electric fields have proven useful for inducing droplet deformation and to structure particles adsorbed at droplet interfaces. In this experimental research, direct current electric fields were applied to deform particle-covered droplets made out of silicone oil and immersed in castor oil. The viscosity of the drop and surrounding fluid were changed by external heating. We designed an experimental system in such a way that electric field-induced electrohydrodynamic (EHD) flows were asymmetric and only present on one side of the drop, i.e., the droplet adjoined a washer and adhered to one of the electrodes constituting the sample cell. The study investigated the influence of viscosity on the steady-state deformation magnitude of particle-laden droplets, droplet compression, and relaxation, as well as particle arrangements at drop interfaces. Initially, before the application of an electric field, we changed the viscosity of the fluids by heating the sample cell at different temperatures. The viscosity of the fluids was varied by changing the temperature of the fluids from 25 to 50°C. Under the application of a uniform electric field of strength 290 Vmm⁻¹, electric stress was induced at the drop interface, yielding drop deformation. In our study, we found that by lowering the fluid viscosity, the velocity of the EHD flows was increased, which also increases the deformation of the drop.

Keywords: drop deformation and relaxation, electric field, electrohydrodynamic flow, particle assembly, viscosity

Procedia PDF Downloads 237
2514 Comparison of Data Reduction Algorithms for Image-Based Point Cloud Derived Digital Terrain Models

Authors: M. Uysal, M. Yilmaz, I. Tiryakioğlu

Abstract:

Digital Terrain Model (DTM) is a digital numerical representation of the Earth's surface. DTMs have been applied to a diverse field of tasks, such as urban planning, military, glacier mapping, disaster management. In the expression of the Earth' surface as a mathematical model, an infinite number of point measurements are needed. Because of the impossibility of this case, the points at regular intervals are measured to characterize the Earth's surface and DTM of the Earth is generated. Hitherto, the classical measurement techniques and photogrammetry method have widespread use in the construction of DTM. At present, RADAR, LiDAR, and stereo satellite images are also used for the construction of DTM. In recent years, especially because of its superiorities, Airborne Light Detection and Ranging (LiDAR) has an increased use in DTM applications. A 3D point cloud is created with LiDAR technology by obtaining numerous point data. However recently, by the development in image mapping methods, the use of unmanned aerial vehicles (UAV) for photogrammetric data acquisition has increased DTM generation from image-based point cloud. The accuracy of the DTM depends on various factors such as data collection method, the distribution of elevation points, the point density, properties of the surface and interpolation methods. In this study, the random data reduction method is compared for DTMs generated from image based point cloud data. The original image based point cloud data set (100%) is reduced to a series of subsets by using random algorithm, representing the 75, 50, 25 and 5% of the original image based point cloud data set. Over the ANS campus of Afyon Kocatepe University as the test area, DTM constructed from the original image based point cloud data set is compared with DTMs interpolated from reduced data sets by Kriging interpolation method. The results show that the random data reduction method can be used to reduce the image based point cloud datasets to 50% density level while still maintaining the quality of DTM.

Keywords: DTM, Unmanned Aerial Vehicle (UAV), uniform, random, kriging

Procedia PDF Downloads 125
2513 Analysis of Cross-Correlations in Emerging Markets Using Random Matrix Theory

Authors: Thomas Chinwe Urama, Patrick Oseloka Ezepue, Peters Chimezie Nnanwa

Abstract:

This paper investigates the universal financial dynamics in two dominant stock markets in Sub-Saharan Africa, through an in-depth analysis of the cross-correlation matrix of price returns in Nigerian Stock Market (NSM) and Johannesburg Stock Exchange (JSE), for the period 2009 to 2013. The strength of correlations between stocks is known to be higher in JSE than that of the NSM. Particularly important for modelling Nigerian derivatives in the future, the interactions of other stocks with the oil sector are weak, whereas the banking sector has strong positive interactions with the other sectors in the stock exchange. For the JSE, it is the oil sector and beverages that have greater sectorial correlations, instead of the banks which have the weaker correlation with other sectors in the stock exchange.

Keywords: random matrix theory, cross-correlations, emerging markets, option pricing, eigenvalues eigenvectors, inverse participation ratios and implied volatility

Procedia PDF Downloads 270
2512 Water Reclamation and Reuse in Asia’s Largest Sewage Treatment Plant

Authors: Naveen Porika, Snigdho Majumdar, Niraj Sethi

Abstract:

Water, food and energy securities are emerging as increasingly important and vital issues for India and the world. Hyderabad urban agglomeration (HUA), the capital city of Andhra Pradesh State in India, is the sixth largest city has a population of about 8.2 million. The Musi River, which is a tributary of Krishna river flows from west to east right through the heart of Hyderabad, about 80% of the water used by people is released back as sewage, which flows back into Musi every day with detrimental effects on the environment and people downstream of the city. The average daily sewage generated in Hyderabad city is 950 MLD, however, treatment capacity exists only for 541 Million Liters per Day (MLD) but only 407 MLD of sewage is treated. As a result, 543 MLD of sewage daily flows into Musi river. Hyderabad’s current estimated water demand stands at 320 Million Gallons per Day (MGD). However, its installed capacity is merely 270 MGD; by 2020 estimated demand will grow to 400 MGD. There is huge gap between current supply and demand, and this is likely to widen by 2021. Developing new fresh water sources is a challenge for Hyderabad, as the fresh water sources are few and far from the City (about 150-200 km) and requires excessive pumping. The constraints presented above make the conventional alternatives for supply augmentation unsustainable and unattractive .One such dependable and captive source of easily available water is the treated sewage. With proper treatment, water of desired quality can be recovered from the waste water (sewage) for recycle and reuse. Hyderabad Amberpet sewage treatment of capacity 339 MLD is Asia’s largest sewage treatment plant. Tertiary sewage treatment Standard basic engineering modules of 30 MLD,60 MLD, 120MLD & 180 MLD for sewage treatment plants has been developed which are utilized for developing Sewage Reclamation & Reuse model in Asia’s largest sewage treatment plant. This paper will focus on Hyderabad Water Supply & Demand, Sewage Generation & Treatment, Technical aspects of Tertiary Sewage Treatment and Utilization of developed standard modules for reclamation & reuse of treated sewage to overcome the deficit of 130 MGD as projected by 2021.

Keywords: water reclamation, reuse, Andhra Pradesh, hyderabad, musi river, sewage, demand and supply, recycle, Amberpet, 339 MLD, engineering modules, tertiary treatment

Procedia PDF Downloads 594
2511 Parallel Opportunity for Water Conservation and Habitat Formation on Regulated Streams through Formation of Thermal Stratification in River Pools

Authors: Todd H. Buxton, Yong G. Lai

Abstract:

Temperature management in regulated rivers can involve significant expenditures of water to meet the cold-water requirements of species in summer. For this purpose, flows released from Lewiston Dam on the Trinity River in Northern California are 12.7 cms with temperatures around 11oC in July through September to provide adult spring Chinook cold water to hold in deep pools and mature until spawning in fall. The releases are more than double the flow and 10oC colder temperatures than the natural conditions before the dam was built. The high, cold releases provide springers the habitat they require but may suppress the stream food base and limit future populations of salmon by reducing the juvenile fish size and survival to adults via the positive relationship between the two. Field and modeling research was undertaken to explore whether lowering summer releases from Lewiston Dam may promote thermal stratification in river pools so that both the cold-water needs of adult salmon and warmer water requirements of other organisms in the stream biome may be met. For this investigation, a three-dimensional (3D) computational fluid dynamics (CFD) model was developed and validated with field measurements in two deep pools on the Trinity River. Modeling and field observations were then used to identify the flows and temperatures that may form and maintain thermal stratification under different meteorologic conditions. Under low flows, a pool was found to be well mixed and thermally homogenous until temperatures began to stratify shortly after sunrise. Stratification then strengthened through the day until shading from trees and mountains cooled the inlet flow and decayed the thermal gradient, which collapsed shortly before sunset and returned the pool to a well-mixed state. This diurnal process of stratification formation and destruction was closely predicted by the 3D CFD model. Both the model and field observations indicate that thermal stratification maintained the coldest temperatures of the day at ≥2m depth in a pool and provided water that was around 8oC warmer in the upper 2m of the pool. Results further indicate that the stratified pool under low flows provided almost the same daily average temperatures as when flows were an order of magnitude higher and stratification was prevented, indicating significant water savings may be realized in regulated streams while also providing a diversity in water temperatures the ecosystem requires. With confidence in the 3D CFD model, the model is now being applied to a dozen pools in the Trinity River to understand how pool bathymetry influences thermal stratification under variable flows and diurnal temperature variations. This knowledge will be used to expand the results to 52 pools in a 64 km reach below Lewiston Dam that meet the depth criteria (≥2 m) for spring Chinook holding. From this, rating curves will be developed to relate discharge to the volume of pool habitat that provides springers the temperature (<15.6oC daily average), velocity (0.15 to 0.4 m/s) and depths that accommodate the escapement target for spring Chinook (6,000 adults) under maximum fish densities measured in other streams (3.1 m3/fish) during the holding time of year (May through August). Flow releases that meet these goals will be evaluated for water savings relative to the current flow regime and their influence on indicator species, including the Foothill Yellow-Legged Frog, and aspects of the stream biome that support salmon populations, including macroinvertebrate production and juvenile Chinook growth rates.

Keywords: 3D CFD modeling, flow regulation, thermal stratification, chinook salmon, foothill yellow-legged frogs, water managment

Procedia PDF Downloads 38
2510 Racial Bias by Prosecutors: Evidence from Random Assignment

Authors: CarlyWill Sloan

Abstract:

Racial disparities in criminal justice outcomes are well-documented. However, there is little evidence on the extent to which racial bias by prosecutors is responsible for these disparities. This paper tests for racial bias in conviction by prosecutors. To identify effects, this paper leverages as good as random variation in prosecutor race using detailed administrative data on the case assignment process and case outcomes in New York County, New York. This paper shows that the assignment of an opposite-race prosecutor leads to a 5 percentage point (~ 8 percent) increase in the likelihood of conviction for property crimes. There is no evidence of effects for other types of crimes. Additional results indicate decreased dismissals by opposite-race prosecutors likely drive my property crime estimates.

Keywords: criminal justice, discrimination, prosecutors, racial disparities

Procedia PDF Downloads 171
2509 Implementation of a Lattice Boltzmann Method for Pulsatile Flow with Moment Based Boundary Condition

Authors: Zainab A. Bu Sinnah, David I. Graham

Abstract:

The Lattice Boltzmann Method has been developed and used to simulate both steady and unsteady fluid flow problems such as turbulent flows, multiphase flow and flows in the vascular system. As an example, the study of blood flow and its properties can give a greater understanding of atherosclerosis and the flow parameters which influence this phenomenon. The blood flow in the vascular system is driven by a pulsating pressure gradient which is produced by the heart. As a very simple model of this, we simulate plane channel flow under periodic forcing. This pulsatile flow is essentially the standard Poiseuille flow except that the flow is driven by the periodic forcing term. Moment boundary conditions, where various moments of the particle distribution function are specified, are applied at solid walls. We used a second-order single relaxation time model and investigated grid convergence using two distinct approaches. In the first approach, we fixed both Reynolds and Womersley numbers and varied relaxation time with grid size. In the second approach, we fixed the Womersley number and relaxation time. The expected second-order convergence was obtained for the second approach. For the first approach, however, the numerical method converged, but not necessarily to the appropriate analytical result. An explanation is given for these observations.

Keywords: Lattice Boltzmann method, single relaxation time, pulsatile flow, moment based boundary condition

Procedia PDF Downloads 211
2508 A Statistical Model for the Dynamics of Single Cathode Spot in Vacuum Cylindrical Cathode

Authors: Po-Wen Chen, Jin-Yu Wu, Md. Manirul Ali, Yang Peng, Chen-Te Chang, Der-Jun Jan

Abstract:

Dynamics of cathode spot has become a major part of vacuum arc discharge with its high academic interest and wide application potential. In this article, using a three-dimensional statistical model, we simulate the distribution of the ignition probability of a new cathode spot occurring in different magnetic pressure on old cathode spot surface and at different arcing time. This model for the ignition probability of a new cathode spot was proposed in two typical situations, one by the pure isotropic random walk in the absence of an external magnetic field, other by the retrograde motion in external magnetic field, in parallel with the cathode surface. We mainly focus on developed relationship between the ignition probability density distribution of a new cathode spot and the external magnetic field.

Keywords: cathode spot, vacuum arc discharge, transverse magnetic field, random walk

Procedia PDF Downloads 410
2507 Gaussian Particle Flow Bernoulli Filter for Single Target Tracking

Authors: Hyeongbok Kim, Lingling Zhao, Xiaohong Su, Junjie Wang

Abstract:

The Bernoulli filter is a precise Bayesian filter for single target tracking based on the random finite set theory. The standard Bernoulli filter often underestimates the number of targets. This study proposes a Gaussian particle flow (GPF) Bernoulli filter employing particle flow to migrate particles from prior to posterior positions to improve the performance of the standard Bernoulli filter. By employing the particle flow filter, the computational speed of the Bernoulli filters is significantly improved. In addition, the GPF Bernoulli filter provides a more accurate estimation compared with that of the standard Bernoulli filter. Simulation results confirm the improved tracking performance and computational speed in two- and three-dimensional scenarios compared with other algorithms.

Keywords: Bernoulli filter, particle filter, particle flow filter, random finite sets, target tracking

Procedia PDF Downloads 60
2506 Diversity of Voices: Audio Visual Continuous Speech Recognition with Traditional Approach

Authors: Partha Protim Majumder, Sajeeb Das, Sharun Akter Khushbu

Abstract:

Bengali is widely spoken in the world, but Bengali speech recognition has not received much attention. Here, we are conducting the toughest task because it must be performed in a noisy place in our study. Another challenge we overcome is dealing with speeches and collecting data on third genders, and our approach is to recognize the gender in speeches. All of the Bangla speech samples used in this study were short and were taken from real-life situations. We employed the male, female, and third-gender categories of speech. In this study, we derive the feature from the spoken word. We used MFCC(1-20), ZCR,rolloff,spec_cen, RMSE, and chroma_stft. Here, we used the algorithms Gboost, Random Forest, K-Nearest Neighbors (KNN), Decision Tree, Naive Bayes, and Logistic Regression (LR) to assess the performance of recognition metrics, and we got the highest performance from random forest in recognizing the gender of the speeches.

Keywords: MFCC, ZCR, Bengali, LR, RMSE, roll-off, Gboost

Procedia PDF Downloads 38