Search results for: Block and Basu bivariate distributions
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1835

Search results for: Block and Basu bivariate distributions

1565 A Case Study on the Estimation of Design Discharge for Flood Management in Lower Damodar Region, India

Authors: Susmita Ghosh

Abstract:

Catchment area of Damodar River, India experiences seasonal rains due to the south-west monsoon every year and depending upon the intensity of the storms, floods occur. During the monsoon season, the rainfall in the area is mainly due to active monsoon conditions. The upstream reach of Damodar river system has five dams store the water for utilization for various purposes viz, irrigation, hydro-power generation, municipal supplies and last but not the least flood moderation. But, in the downstream reach of Damodar River, known as Lower Damodar region, is severely and frequently suffering from flood due to heavy monsoon rainfall and also release from upstream reservoirs. Therefore, an effective flood management study is required to know in depth the nature and extent of flood, water logging, and erosion related problems, affected area, and damages in the Lower Damodar region, by conducting mathematical model study. The design flood or discharge is needed to decide to assign the respective model for getting several scenarios from the simulation runs. The ultimate aim is to achieve a sustainable flood management scheme from the several alternatives. there are various methods for estimating flood discharges to be carried through the rivers and their tributaries for quick drainage from inundated areas due to drainage congestion and excess rainfall. In the present study, the flood frequency analysis is performed to decide the design flood discharge of the study area. This, on the other hand, has limitations in respect of availability of long peak flood data record for determining long type of probability density function correctly. If sufficient past records are available, the maximum flood on a river with a given frequency can safely be determined. The floods of different frequency for the Damodar has been calculated by five candidate distributions i.e., generalized extreme value, extreme value-I, Pearson type III, Log Pearson and normal. Annual peak discharge series are available at Durgapur barrage for the period of 1979 to 2013 (35 years). The available series are subjected to frequency analysis. The primary objective of the flood frequency analysis is to relate the magnitude of extreme events to their frequencies of occurrence through the use of probability distributions. The design flood for return periods of 10, 15 and 25 years return period at Durgapur barrage are estimated by flood frequency method. It is necessary to develop flood hydrographs for the above floods to facilitate the mathematical model studies to find the depth and extent of inundation etc. Null hypothesis that the distributions fit the data at 95% confidence is checked with goodness of fit test, i.e., Chi Square Test. It is revealed from the goodness of fit test that the all five distributions do show a good fit on the sample population and is therefore accepted. However, it is seen that there is considerable variation in the estimation of frequency flood. It is therefore considered prudent to average out the results of these five distributions for required frequencies. The inundated area from past data is well matched using this flood.

Keywords: design discharge, flood frequency, goodness of fit, sustainable flood management

Procedia PDF Downloads 178
1564 Very Large Scale Integration Architecture of Finite Impulse Response Filter Implementation Using Retiming Technique

Authors: S. Jalaja, A. M. Vijaya Prakash

Abstract:

Recursive combination of an algorithm based on Karatsuba multiplication is exploited to design a generalized transpose and parallel Finite Impulse Response (FIR) Filter. Mid-range Karatsuba multiplication and Carry Save adder based on Karatsuba multiplication reduce time complexity for higher order multiplication implemented up to n-bit. As a result, we design modified N-tap Transpose and Parallel Symmetric FIR Filter Structure using Karatsuba algorithm. The mathematical formulation of the FFA Filter is derived. The proposed architecture involves significantly less area delay product (APD) then the existing block implementation. By adopting retiming technique, hardware cost is reduced further. The filter architecture is designed by using 90 nm technology library and is implemented by using cadence EDA Tool. The synthesized result shows better performance for different word length and block size. The design achieves switching activity reduction and low power consumption by applying with and without retiming for different combination of the circuit. The proposed structure achieves more than a half of the power reduction by adopting with and without retiming techniques compared to the earlier design structure. As a proof of the concept for block size 16 and filter length 64 for CKA method, it achieves a 51% as well as 70% less power by applying retiming technique, and for CSA method it achieves a 57% as well as 77% less power by applying retiming technique compared to the previously proposed design.

Keywords: carry save adder Karatsuba multiplication, mid range Karatsuba multiplication, modified FFA and transposed filter, retiming

Procedia PDF Downloads 210
1563 Fingerprint Image Encryption Using a 2D Chaotic Map and Elliptic Curve Cryptography

Authors: D. M. S. Bandara, Yunqi Lei, Ye Luo

Abstract:

Fingerprints are suitable as long-term markers of human identity since they provide detailed and unique individual features which are difficult to alter and durable over life time. In this paper, we propose an algorithm to encrypt and decrypt fingerprint images by using a specially designed Elliptic Curve Cryptography (ECC) procedure based on block ciphers. In addition, to increase the confusing effect of fingerprint encryption, we also utilize a chaotic-behaved method called Arnold Cat Map (ACM) for a 2D scrambling of pixel locations in our method. Experimental results are carried out with various types of efficiency and security analyses. As a result, we demonstrate that the proposed fingerprint encryption/decryption algorithm is advantageous in several different aspects including efficiency, security and flexibility. In particular, using this algorithm, we achieve a margin of about 0.1% in the test of Number of Pixel Changing Rate (NPCR) values comparing to the-state-of-the-art performances.

Keywords: arnold cat map, biometric encryption, block cipher, elliptic curve cryptography, fingerprint encryption, Koblitz’s encoding

Procedia PDF Downloads 178
1562 Water Resources Green Efficiency in China: Evaluation, Spatial Association Network Structure Analysis, and Influencing Factors

Authors: Tingyu Zhang

Abstract:

This paper utilizes the Super-SBM model to assess water resources green efficiency (WRGE) among provinces in China and investigate its spatial and temporal features, based on the characteristic framework of “economy-environment-society.” The social network analysis is employed to examine the network pattern and spatial interaction of WRGE. Further, the quadratic assignment procedure method is utilized for examining the influencing factors of the spatial association of WRGE regarding “relationship.” The study reveals that: (1) the spatial distribution of WRGE demonstrates a distribution pattern of Eastern>Western>Central; (2) a remarkable spatial association exists among provinces; however, no strict hierarchical structure is observed. The internal structure of the WRGE network is characterized by the feature of "Eastern strong and Western weak". The block model analysis discovers that the members of the “net spillover” and “two-way spillover” blocks are mostly in the eastern and central provinces; “broker” block, which plays an intermediary role, is mostly in the central provinces; and members of the “net beneficiary” block are mostly in the western region. (3) Differences in economic development, degree of urbanization, water use environment, and water management have significant impacts on the spatial connection of WRGE. This study is dedicated to the realization of regional linkages and synergistic enhancement of WRGE, which provides a meaningful basis for building a harmonious society of human and water coexistence.

Keywords: water resources green efficiency, super-SBM model, social network analysis, quadratic assignment procedure

Procedia PDF Downloads 35
1561 Statistical Modelling of Maximum Temperature in Rwanda Using Extreme Value Analysis

Authors: Emmanuel Iyamuremye, Edouard Singirankabo, Alexis Habineza, Yunvirusaba Nelson

Abstract:

Temperature is one of the most important climatic factors for crop production. However, severe temperatures cause drought, feverish and cold spells that have various consequences for human life, agriculture, and the environment in general. It is necessary to provide reliable information related to the incidents and the probability of such extreme events occurring. In the 21st century, the world faces a huge number of threats, especially from climate change, due to global warming and environmental degradation. The rise in temperature has a direct effect on the decrease in rainfall. This has an impact on crop growth and development, which in turn decreases crop yield and quality. Countries that are heavily dependent on agriculture use to suffer a lot and need to take preventive steps to overcome these challenges. The main objective of this study is to model the statistical behaviour of extreme maximum temperature values in Rwanda. To achieve such an objective, the daily temperature data spanned the period from January 2000 to December 2017 recorded at nine weather stations collected from the Rwanda Meteorological Agency were used. The two methods, namely the block maxima (BM) method and the Peaks Over Threshold (POT), were applied to model and analyse extreme temperature. Model parameters were estimated, while the extreme temperature return periods and confidence intervals were predicted. The model fit suggests Gumbel and Beta distributions to be the most appropriate models for the annual maximum of daily temperature. The results show that the temperature will continue to increase, as shown by estimated return levels.

Keywords: climate change, global warming, extreme value theory, rwanda, temperature, generalised extreme value distribution, generalised pareto distribution

Procedia PDF Downloads 152
1560 Lateritic Soils from Ceara, Brazil: Sustainable Use in Constructive Blocks for Social Housing

Authors: Ivelise M. Strozberg, Juliana Sales Frota, Lucas de Oliveira Vale

Abstract:

The state of Ceara, located in the northeast region of Brazil, is abundant in lateritic soil which has been usually discarded due to its lack of agricultural potential while materials of similar nature have been used as constituents of housing constructive elements in many parts of the world, such as India and Portugal, for decades. Since many of the semi-arid housing conditions in the state of Ceara fail to meet the minimum criteria regarding comfort and safety requirements, this research proposed to study the Ceara lateritic soil and the possibility of its use as a sustainable building block constituent for social housings, collaborating to the improvement of the region living conditions. In order to achieve this objective, soil samples were collected from five different locations within the specific region, three of which presented lateritic nature, being characterized according to the Unified Soil Classification System and the MCT methodology, which is a Brazilian methodology developed during the 80’s that aimed to better describe and approach tropical soils, its characterization and behavior. Two of these samples were used to build two different miniature block prototypes, which were manually molded, heated at low temperatures -( < 300 ºC) in order to save energy and lessen the CO₂ high emission rate common in traditional burning methods- and then submitted to load tests. Among the soils tested, the one with the highest degree of laterization and greater presence of fines constituted the block with the best performance in terms of flexural strength tensions, presenting resistance gains when heated at increasing temperatures, which can indicate that this type of soil has potential towards being used as constructing material.

Keywords: constructive blocks, lateritic soil, MCT methodology, sustainability

Procedia PDF Downloads 107
1559 User-Driven Product Line Engineering for Assembling Large Families of Software

Authors: Zhaopeng Xuan, Yuan Bian, C. Cailleaux, Jing Qin, S. Traore

Abstract:

Traditional software engineering allows engineers to propose to their clients multiple specialized software distributions assembled from a shared set of software assets. The management of these assets however requires a trade-off between client satisfaction and software engineering process. Clients have more and more difficult to find a distribution or components based on their needs from all of distributed repositories. This paper proposes a software engineering for a user-driven software product line in which engineers define a feature model but users drive the actual software distribution on demand. This approach makes the user become final actor as a release manager in software engineering process, increasing user product satisfaction and simplifying user operations to find required components. In addition, it provides a way for engineers to manage and assembly large software families. As a proof of concept, a user-driven software product line is implemented for eclipse, an integrated development environment. An eclipse feature model is defined, which is exposed to users on a cloud-based built platform from which clients can download individualized Eclipse distributions.

Keywords: software product line, model-driven development, reverse engineering and refactoring, agile method

Procedia PDF Downloads 413
1558 Development of Alpha Spectroscopy Method with Solid State Nuclear Track Detector Using Aluminium Thin Films

Authors: Nidal Dwaikat

Abstract:

This work presents the development of alpha spectroscopy method with Solid-state nuclear track detectors using aluminum thin films. The resolution of this method is high, and it is able to discriminate between alpha particles at different incident energy. It can measure the exact number of alpha particles at specific energy without needing a calibration of alpha track diameter versus alpha energy. This method was tested by using Cf-252 alpha standard source at energies 5.11 Mev, 3.86 MeV and 2.7 MeV, which produced by the variation of detector -standard source distance. On front side, two detectors were covered with two Aluminum thin films and the third detector was kept uncovered. The thickness of Aluminum thin films was selected carefully (using SRIM 2013) such that one of the films will block the lower two alpha particles (3.86 MeV and 2.7 MeV) and the alpha particles at higher energy (5.11 Mev) can penetrate the film and reach the detector’s surface. The second thin film will block alpha particles at lower energy of 2.7 MeV and allow alpha particles at higher two energies (5.11 Mev and 3.86 MeV) to penetrate and produce tracks. For uncovered detector, alpha particles at three different energies can produce tracks on it. For quality assurance and accuracy, the detectors were mounted on thick enough copper substrates to block exposure from the backside. The tracks on the first detector are due to alpha particles at energy of 5.11 MeV. The difference between the tracks number on the first detector and the tracks number on the second detector is due to alpha particles at energy of 3.8 MeV. Finally, by subtracting the tracks number on the second detector from the tracks number on the third detector (uncovered), we can find the tracks number due to alpha particles at energy 2.7 MeV. After knowing the efficiency calibration factor, we can exactly calculate the activity of standard source.

Keywords: aluminium thin film, alpha particles, copper substrate, CR-39 detector

Procedia PDF Downloads 343
1557 Electrospun Nanofibers from Amphiphlic Block Copolymers and Their Graphene Nanocomposites

Authors: Hussein M. Etmimi, Peter E. Mallon

Abstract:

Electrospinning uses an electrical charge to draw very fine fibers (typically on the micro or nano scale) from a liquid or molten precursor. Over the years, this method has become a widely used and a successful technique to process polymer materials and their composites into nanofibers. The main focus of this work is to study the electrospinning of multi-phase amphiphilic copolymers and their nanocomposites, which contain graphene as the nanofiller material. In such amphiphilic materials, the constituents segments are incompatible and thus the solid state morphology will be determined by the composition of the various constituents as well as the method of preparation. In this study, amphiphilic block copolymers of poly(dimethyl siloxane) and poly(methyl methacrylate) (PDMS-b-PMMA) with well-defined structures were synthesized and the solution electrospinning of these materials and their properties were investigated. Atom transfer radical polymerization (ATRP) was used to obtain the controlled block copolymers with relatively high molar masses and narrow dispersity. First, PDMS macroinitiators with different chain length of 1000, 5000 and 10000 g/mol were synthesized by the reaction of monocarbinol terminated PDMS with α-bromoisobutyryl bromide initiator. The obtained macroinitiators were used for the polymerization of methyl methacrylate monomer to obtain the desired block copolymers using the ATRP process. Graphene oxide (GO) of different loading was then added to the copolymer solution and the resultant nanocomposites were successfully electrospun into nanofibers. The electrospinning was achieved using dimethylformamide/chloroform mixture (60:40 vl%) as electrospinning solution medium. Scanning electron microscopy (SEM) showed the successful formation of the electrospun fibers with dimensions in the nanometer range. X-ray diffraction indicated that the GO nanosheets were of an exfoliated structure, irrespective of the filler loading. Thermogravimetric analysis also showed that the thermal stability of the nanofibers was improved in the presence of GO, which was not a function of the filler loading. Differential scanning calorimetry also showed that the mechanical properties (measured as glass transition temperature) of the nanofibers was improved significantly in the presence of GO, which was a function of the filler loading.

Keywords: elctrospinning, graphene oxide, nanofibers, polymethyl methacrylate (PMMA)

Procedia PDF Downloads 289
1556 Modeling Thermal Changes of Urban Blocks in Relation to the Landscape Structure and Configuration in Guilan Province

Authors: Roshanak Afrakhteh, Abdolrasoul Salman Mahini, Mahdi Motagh, Hamidreza Kamyab

Abstract:

Urban Heat Islands (UHIs) are distinctive urban areas characterized by densely populated central cores surrounded by less densely populated peripheral lands. These areas experience elevated temperatures, primarily due to impermeable surfaces and specific land use patterns. The consequences of these temperature variations are far-reaching, impacting the environment and society negatively, leading to increased energy consumption, air pollution, and public health concerns. This paper emphasizes the need for simplified approaches to comprehend UHI temperature dynamics and explains how urban development patterns contribute to land surface temperature variation. To illustrate this relationship, the study focuses on the Guilan Plain, utilizing techniques like principal component analysis and generalized additive models. The research centered on mapping land use and land surface temperature in the low-lying area of Guilan province. Satellite data from Landsat sensors for three different time periods (2002, 2012, and 2021) were employed. Using eCognition software, a spatial unit known as a "city block" was utilized through object-based analysis. The study also applied the normalized difference vegetation index (NDVI) method to estimate land surface radiance. Predictive variables for urban land surface temperature within residential city blocks were identified categorized as intrinsic (related to the block's structure) and neighboring (related to adjacent blocks) variables. Principal Component Analysis (PCA) was used to select significant variables, and a Generalized Additive Model (GAM) approach, implemented using R's mgcv package, modeled the relationship between urban land surface temperature and predictor variables.Notable findings included variations in urban temperature across different years attributed to environmental and climatic factors. Block size, shared boundary, mother polygon area, and perimeter-to-area ratio were identified as main variables for the generalized additive regression model. This model showed non-linear relationships, with block size, shared boundary, and mother polygon area positively correlated with temperature, while the perimeter-to-area ratio displayed a negative trend. The discussion highlights the challenges of predicting urban surface temperature and the significance of block size in determining urban temperature patterns. It also underscores the importance of spatial configuration and unit structure in shaping urban temperature patterns. In conclusion, this study contributes to the growing body of research on the connection between land use patterns and urban surface temperature. Block size, along with block dispersion and aggregation, emerged as key factors influencing urban surface temperature in residential areas. The proposed methodology enhances our understanding of parameter significance in shaping urban temperature patterns across various regions, particularly in Iran.

Keywords: urban heat island, land surface temperature, LST modeling, GAM, Gilan province

Procedia PDF Downloads 54
1555 Distribution Patterns of Trace Metals in Soils of Gbongan-Odeyinka-Orileowu Area, Southwestern Nigeria

Authors: T. A. Adesiyan, J. A. Adekoya A. Akinlua, N. Torto

Abstract:

One hundred and eighty six in situ soil samples of the B–horizon were collected around Gbongan–Odeyinka-Orileowu area, southwestern Nigeria, delineated by longitude 4°15l and 4°30l and latitude 7°14l and 7°31 for a reconnaissance geochemical soil survey. The objective was to determine the distribution pattern of some trace metals in the area with a view to discovering any indication of metallic mineralization. The samples were air–dried and sieved to obtain the minus 230 µ fractions which were used for pH determinations and subjected to hot aqua regia acid digestion. The solutions obtained were analyzed for Ag, As, Au, Cd, Co, Cr, Cu, Fe, Mn, Ni, Pb, Sn, and Zn using atomic absorption spectrometric methods. The resulting data were subjected to simple statistical treatment and used in preparing distribution maps of the elements. With these, the spatial distributions of the elements in the area were discussed. The pH of the soils range from 4.70 to 7.59 and this reflects the geochemical distribution patterns of trace metals in the area. The spatial distribution maps of the elements showed similarity in the distributions of Co, Cr, Fe, Ni, Mn and Pb. This suggests close associations between these elements none of which showed any significant anomaly in the study. The associations might be due to the scavenging actions of Fe–Mn oxides on the elements. Only Ag, Au and Sn on one hand and Zn on the other hand showed significant anomalies, which are thought to be due to mineralization and anthropogenic activities respectively.

Keywords: distribution, metals, Gbongan, Nigeria, mineralization anthropogenic

Procedia PDF Downloads 301
1554 Parkinson’s Disease Hand-Eye Coordination and Dexterity Evaluation System

Authors: Wann-Yun Shieh, Chin-Man Wang, Ya-Cheng Shieh

Abstract:

This study aims to develop an objective scoring system to evaluate hand-eye coordination and hand dexterity for Parkinson’s disease. This system contains three boards, and each of them is implemented with the sensors to sense a user’s finger operations. The operations include the peg test, the block test, and the blind block test. A user has to use the vision, hearing, and tactile abilities to finish these operations, and the board will record the results automatically. These results can help the physicians to evaluate a user’s reaction, coordination, dexterity function. The results will be collected to a cloud database for further analysis and statistics. A researcher can use this system to obtain systematic, graphic reports for an individual or a group of users. Particularly, a deep learning model is developed to learn the features of the data from different users. This model will help the physicians to assess the Parkinson’s disease symptoms by a more intellective algorithm.

Keywords: deep learning, hand-eye coordination, reaction, hand dexterity

Procedia PDF Downloads 45
1553 Distributional and Developmental Analysis of PM2.5 in Beijing, China

Authors: Alexander K. Guo

Abstract:

PM2.5 poses a large threat to people’s health and the environment and is an issue of large concern in Beijing, brought to the attention of the government by the media. In addition, both the United States Embassy in Beijing and the government of China have increased monitoring of PM2.5 in recent years, and have made real-time data available to the public. This report utilizes hourly historical data (2008-2016) from the U.S. Embassy in Beijing for the first time. The first objective was to attempt to fit probability distributions to the data to better predict a number of days exceeding the standard, and the second was to uncover any yearly, seasonal, monthly, daily, and hourly patterns and trends that may arise to better understand of air control policy. In these data, 66,650 hours and 2687 days provided valid data. Lognormal, gamma, and Weibull distributions were fit to the data through an estimation of parameters. The Chi-squared test was employed to compare the actual data with the fitted distributions. The data were used to uncover trends, patterns, and improvements in PM2.5 concentration over the period of time with valid data in addition to specific periods of time that received large amounts of media attention, analyzed to gain a better understanding of causes of air pollution. The data show a clear indication that Beijing’s air quality is unhealthy, with an average of 94.07µg/m3 across all 66,650 hours with valid data. It was found that no distribution fit the entire dataset of all 2687 days well, but each of the three above distribution types was optimal in at least one of the yearly data sets, with the lognormal distribution found to fit recent years better. An improvement in air quality beginning in 2014 was discovered, with the first five months of 2016 reporting an average PM2.5 concentration that is 23.8% lower than the average of the same period in all years, perhaps the result of various new pollution-control policies. It was also found that the winter and fall months contained more days in both good and extremely polluted categories, leading to a higher average but a comparable median in these months. Additionally, the evening hours, especially in the winter, reported much higher PM2.5 concentrations than the afternoon hours, possibly due to the prohibition of trucks in the city in the daytime and the increased use of coal for heating in the colder months when residents are home in the evening. Lastly, through analysis of special intervals that attracted media attention for either unnaturally good or bad air quality, the government’s temporary pollution control measures, such as more intensive road-space rationing and factory closures, are shown to be effective. In summary, air quality in Beijing is improving steadily and do follow standard probability distributions to an extent, but still needs improvement. Analysis will be updated when new data become available.

Keywords: Beijing, distribution, patterns, pm2.5, trends

Procedia PDF Downloads 224
1552 Topping Failure Analysis of Anti-Dip Bedding Rock Slopes Subjected to Crest Loads

Authors: Chaoyi Sun, Congxin Chen, Yun Zheng, Kaizong Xia, Wei Zhang

Abstract:

Crest loads are often encountered in hydropower, highway, open-pit and other engineering rock slopes. Toppling failure is one of the most common deformation failure types of anti-dip bedding rock slopes. Analysis on such failure of anti-dip bedding rock slopes subjected to crest loads has an important influence on engineering practice. Based on the step-by-step analysis approach proposed by Goodman and Bray, a geo-mechanical model was developed, and the related analysis approach was proposed for the toppling failure of anti-dip bedding rock slopes subjected to crest loads. Using the transfer coefficient method, a formulation was derived for calculating the residual thrust of slope toe and the support force required to meet the requirements of the slope stability under crest loads, which provided a scientific reference to design and support for such slopes. Through slope examples, the influence of crest loads on the residual thrust and sliding ratio coefficient was investigated for cases of different block widths and slope cut angles. The results show that there exists a critical block width for such slope. The influence of crest loads on the residual thrust is non-negligible when the block thickness is smaller than the critical value. Moreover, the influence of crest loads on the slope stability increases with the slope cut angle and the sliding ratio coefficient of anti-dip bedding rock slopes increases with the crest loads. Finally, the theoretical solutions and numerical simulations using Universal Distinct Element Code (UDEC) were compared, in which the consistent results show the applicability of both approaches.

Keywords: anti-dip bedding rock slope, crest loads, stability analysis, toppling failure

Procedia PDF Downloads 162
1551 Regional Hydrological Extremes Frequency Analysis Based on Statistical and Hydrological Models

Authors: Hadush Kidane Meresa

Abstract:

The hydrological extremes frequency analysis is the foundation for the hydraulic engineering design, flood protection, drought management and water resources management and planning to utilize the available water resource to meet the desired objectives of different organizations and sectors in a country. This spatial variation of the statistical characteristics of the extreme flood and drought events are key practice for regional flood and drought analysis and mitigation management. For different hydro-climate of the regions, where the data set is short, scarcity, poor quality and insufficient, the regionalization methods are applied to transfer at-site data to a region. This study aims in regional high and low flow frequency analysis for Poland River Basins. Due to high frequent occurring of hydrological extremes in the region and rapid water resources development in this basin have caused serious concerns over the flood and drought magnitude and frequencies of the river in Poland. The magnitude and frequency result of high and low flows in the basin is needed for flood and drought planning, management and protection at present and future. Hydrological homogeneous high and low flow regions are formed by the cluster analysis of site characteristics, using the hierarchical and C- mean clustering and PCA method. Statistical tests for regional homogeneity are utilized, by Discordancy and Heterogeneity measure tests. In compliance with results of the tests, the region river basin has been divided into ten homogeneous regions. In this study, frequency analysis of high and low flows using AM for high flow and 7-day minimum low flow series is conducted using six statistical distributions. The use of L-moment and LL-moment method showed a homogeneous region over entire province with Generalized logistic (GLOG), Generalized extreme value (GEV), Pearson type III (P-III), Generalized Pareto (GPAR), Weibull (WEI) and Power (PR) distributions as the regional drought and flood frequency distributions. The 95% percentile and Flow duration curves of 1, 7, 10, 30 days have been plotted for 10 stations. However, the cluster analysis performed two regions in west and east of the province where L-moment and LL-moment method demonstrated the homogeneity of the regions and GLOG and Pearson Type III (PIII) distributions as regional frequency distributions for each region, respectively. The spatial variation and regional frequency distribution of flood and drought characteristics for 10 best catchment from the whole region was selected and beside the main variable (streamflow: high and low) we used variables which are more related to physiographic and drainage characteristics for identify and delineate homogeneous pools and to derive best regression models for ungauged sites. Those are mean annual rainfall, seasonal flow, average slope, NDVI, aspect, flow length, flow direction, maximum soil moisture, elevation, and drainage order. The regional high-flow or low-flow relationship among one streamflow characteristics with (AM or 7-day mean annual low flows) some basin characteristics is developed using Generalized Linear Mixed Model (GLMM) and Generalized Least Square (GLS) regression model, providing a simple and effective method for estimation of flood and drought of desired return periods for ungauged catchments.

Keywords: flood , drought, frequency, magnitude, regionalization, stochastic, ungauged, Poland

Procedia PDF Downloads 580
1550 NFResNet: Multi-Scale and U-Shaped Networks for Deblurring

Authors: Tanish Mittal, Preyansh Agrawal, Esha Pahwa, Aarya Makwana

Abstract:

Multi-Scale and U-shaped Networks are widely used in various image restoration problems, including deblurring. Keeping in mind the wide range of applications, we present a comparison of these architectures and their effects on image deblurring. We also introduce a new block called as NFResblock. It consists of a Fast Fourier Transformation layer and a series of modified Non-Linear Activation Free Blocks. Based on these architectures and additions, we introduce NFResnet and NFResnet+, which are modified multi-scale and U-Net architectures, respectively. We also use three differ-ent loss functions to train these architectures: Charbonnier Loss, Edge Loss, and Frequency Reconstruction Loss. Extensive experiments on the Deep Video Deblurring dataset, along with ablation studies for each component, have been presented in this paper. The proposed architectures achieve a considerable increase in Peak Signal to Noise (PSNR) ratio and Structural Similarity Index (SSIM) value.

Keywords: multi-scale, Unet, deblurring, FFT, resblock, NAF-block, nfresnet, charbonnier, edge, frequency reconstruction

Procedia PDF Downloads 101
1549 Transformations of Spatial Distributions of Bio-Polymers and Nanoparticles in Water Suspensions Induced by Resonance-Like Low Frequency Electrical Fields

Authors: A. A. Vasin, N. V. Klassen, A. M. Likhter

Abstract:

Water suspensions of in-organic (metals and oxides) and organic nano-objects (chitozan and collagen) were subjected to the treatment of direct and alternative electrical fields. In addition to quasi-periodical spatial patterning resonance-like performance of spatial distributions of these suspensions has been found at low frequencies of alternating electrical field. These resonances are explained as the result of creation of equilibrium states of groups of charged nano-objects with opposite signs of charges at the interparticle distances where the forces of Coulomb attraction are compensated by the repulsion forces induced by relatively negative polarization of hydrated regions surrounding the nanoparticles with respect to pure water. The low frequencies of these resonances are explained by comparatively big distances between the particles and their big masses with t\respect to masses of atoms constituting molecules with high resonance frequencies. These new resonances open a new approach to detailed modeling and understanding of mechanisms of the influence of electrical fields on the functioning of internal organs of living organisms at the level of cells and neurons.

Keywords: bio-polymers, chitosan, collagen, nanoparticles, coulomb attraction, polarization repulsion, periodical patterning, electrical low frequency resonances, transformations

Procedia PDF Downloads 528
1548 Comparison of Blockchain Ecosystem for Identity Management

Authors: K. S. Suganya, R. Nedunchezhian

Abstract:

In recent years, blockchain technology has been found to be the most significant discovery in this digital era, after the discovery of the Internet and Cloud Computing. Blockchain is a simple, distributed public ledger that contains all the user’s transaction details in a block. The global copy of the block is then shared among all its peer-peer network users after validation by the Blockchain miners. Once a block is validated and accepted, it cannot be altered by any users making it a trust-free transaction. It also resolves the problem of double-spending by using traditional cryptographic methods. Since the advent of bitcoin, blockchain has been the backbone for all its transactions. But in recent years, it has found its roots and uses in many fields like Smart Contracts, Smart City management, healthcare, etc. Identity management against digital identity theft has become a major concern among financial and other organizations. To solve this digital identity theft, blockchain technology can be employed with existing identity management systems, which maintain a distributed public ledger containing details of an individual’s identity containing information such as Digital birth certificates, Citizenship number, Bank details, voter details, driving license in the form of blocks verified on the blockchain becomes time-stamped, unforgeable and publicly visible for any legitimate users. The main challenge in using blockchain technology to prevent digital identity theft is ensuring the pseudo-anonymity and privacy of the users. This survey paper will exert to study the blockchain concepts, consensus protocols, and various blockchain-based Digital Identity Management systems with their research scope. This paper also discusses the role of Blockchain in COVID-19 pandemic management by self-sovereign identity and supply chain management.

Keywords: blockchain, consensus protocols, bitcoin, identity theft, digital identity management, pandemic, COVID-19, self-sovereign identity

Procedia PDF Downloads 104
1547 Treatment and Reuse of Nonmetallic PCBs Waste

Authors: Johan Sohaili, Siti Suhaila Mohamad, Shantha Kumari Muniyandi

Abstract:

The strength development, durability and leachability aspects of mortar added with nonmetallic printed circuit board (NMPCBs) were investigated. This study aims to propose methods for treatment and reuse of NMPCBs waste. The leachability of raw NMPCBs was tested for toxicity by performing the Crushed Block Leachability (CBL) test. The effectiveness of the treatment was evaluated by performing compressive, flexural strength, durability and whole block leachability (WBL) tests on the mortar. The results indicated that the concentration of metals leach from the raw NMPCBs are within the standard limits and higher than the concentration of metals from WBL test. The compressive and flexural strength of the NMPCBs mortar was generally lower than the standard mortar. From durability tests, weight and compressive strength both of mortars was decrease after soaking in acid solution. As a conclusion, the treated NMPCBs can be reused in profitable and environmentally friendly ways and has broad application prospects.

Keywords: nonmetallic, printed circuit board, treatment, reuse

Procedia PDF Downloads 447
1546 High Temperature Behaviour of Various Limestone Used in Heritage Buildings at Material and Block Scales

Authors: Ayoub Daoudi, Javad Eslami, Anne-Lise Beaucour, Martin Vigroux, Albert Noumowé

Abstract:

As a fact, many cultural heritage masonry buildings have undergone violent fires during their history. In order to investigate the high temperature behaviour of stone masonry, six French limestones were heated to 600 °C at a rate of 9 °C/min. The main focus is the comparison between the high temperature behaviour of stones at the material and at the structural scale. In order to evaluate the risk of spalling, the tests have been carried out on the stone blocks (12x30x30 cm) instrumented with thermocouples and subjected to an unidirectional heating on one face. Thereafter, visual assessments and non-destructive measurements (dynamic elastic modulus) performed on blocks demonstrate a different behaviour from what was observed at the material scale. Finally, a series of thermo-mechanical computations, using finite element method, allowed us to highlight the difference between the behaviour of stones at material and block scales.

Keywords: limestones, hight temperature behaviour, damage, thermo-mechanical modeling, material and blocks scales, color change

Procedia PDF Downloads 86
1545 Efficient Motion Estimation by Fast Three Step Search Algorithm

Authors: S. M. Kulkarni, D. S. Bormane, S. L. Nalbalwar

Abstract:

The rapid development in the technology have dramatic impact on the medical health care field. Medical data base obtained with latest machines like CT Machine, MRI scanner requires large amount of memory storage and also it requires large bandwidth for transmission of data in telemedicine applications. Thus, there is need for video compression. As the database of medical images contain number of frames (slices), hence while coding of these images there is need of motion estimation. Motion estimation finds out movement of objects in an image sequence and gets motion vectors which represents estimated motion of object in the frame. In order to reduce temporal redundancy between successive frames of video sequence, motion compensation is preformed. In this paper three step search (TSS) block matching algorithm is implemented on different types of video sequences. It is shown that three step search algorithm produces better quality performance and less computational time compared with exhaustive full search algorithm.

Keywords: block matching, exhaustive search motion estimation, three step search, video compression

Procedia PDF Downloads 462
1544 Investigating the Effects of Data Transformations on a Bi-Dimensional Chi-Square Test

Authors: Alexandru George Vaduva, Adriana Vlad, Bogdan Badea

Abstract:

In this research, we conduct a Monte Carlo analysis on a two-dimensional χ2 test, which is used to determine the minimum distance required for independent sampling in the context of chaotic signals. We investigate the impact of transforming initial data sets from any probability distribution to new signals with a uniform distribution using the Spearman rank correlation on the χ2 test. This transformation removes the randomness of the data pairs, and as a result, the observed distribution of χ2 test values differs from the expected distribution. We propose a solution to this problem and evaluate it using another chaotic signal.

Keywords: chaotic signals, logistic map, Pearson’s test, Chi Square test, bivariate distribution, statistical independence

Procedia PDF Downloads 72
1543 The Incidence of Obesity among Adult Women in Pekanbaru City, Indonesia, Related to High Fat Consumption, Stress Level, and Physical Activity

Authors: Yudia Mailani Putri, Martalena Purba, B. J. Istiti Kandarina

Abstract:

Background: Obesity has been recognized as a global health problem. Individuals classified as overweight and obese are increasing at an alarming rate. This condition is associated with psychological and physiological problems. as a person reaches adulthood, somatic growth ceases. At this stage, the human body has developed fully, to a stable state. As the capital of Riau Province in Indonesia, Pekanbaru is dominated by Malay ethnic population habitually consuming cholesterol-rich fatty foods as a daily menu, a trigger to the onset of obesity resulting in high prevalence of degenerative diseases. Research objectives: The aim of this study is elaborating the relationship between high-fat consumption pattern, stress level, physical activity and the incidence of obesity in adult women in Pekanbaru city. Research Methods: Among the combined research methods applied in this study, the first stage is quantitative observational, analytical cross-sectional research design with adult women aged 20-40 living in Pekanbaru city. The sample consists of 200 women with BMI≥25. Sample data is processed with univariate, bivariate (correlation and simple linear regression) and multivariate (multiple linear regression) analysis. The second phase is qualitative descriptive study purposive sampling by in-depth interviews. six participants withdrew from the study. Results: According to the results of the bivariate analysis, there are relationships between the incidence of obesity and the pattern of high fat foods consumption (energy intake (p≤0.000; r = 0.536), protein intake (p≤0.000; r=0.307), fat intake (p≤0.000; r=0.416), carbohydrate intake (p≤0.000; r=0.430), frequency of fatty food consumption (p≤0.000; r=0.506) and frequency of viscera foods consumption (p≤0.000; r=0.535). There is a relationship between physical activity and incidence of obesity (p≤0.000; r=-0.631). However, there is no relationship between the level of stress (p=0.741; r=0.019-) and the incidence of obesity. Physical activity is a predominant factor in the incidence of obesity in adult women in Pekanbaru city. Conclusion: There are relationships between high-fat food consumption pattern, physical activity and the incidence of obesity in Pekanbaru city whereas physical activity is a predominant factor in the occurrence of obesity, supported by the unchangeable pattern of high-fat foods consumption.

Keywords: obesity, adult, high in fat, stress, physical activity, consumption pattern

Procedia PDF Downloads 216
1542 Conflation Methodology Applied to Flood Recovery

Authors: Eva L. Suarez, Daniel E. Meeroff, Yan Yong

Abstract:

Current flooding risk modeling focuses on resilience, defined as the probability of recovery from a severe flooding event. However, the long-term damage to property and well-being by nuisance flooding and its long-term effects on communities are not typically included in risk assessments. An approach was developed to address the probability of recovering from a severe flooding event combined with the probability of community performance during a nuisance event. A consolidated model, namely the conflation flooding recovery (&FR) model, evaluates risk-coping mitigation strategies for communities based on the recovery time from catastrophic events, such as hurricanes or extreme surges, and from everyday nuisance flooding events. The &FR model assesses the variation contribution of each independent input and generates a weighted output that favors the distribution with minimum variation. This approach is especially useful if the input distributions have dissimilar variances. The &FR is defined as a single distribution resulting from the product of the individual probability density functions. The resulting conflated distribution resides between the parent distributions, and it infers the recovery time required by a community to return to basic functions, such as power, utilities, transportation, and civil order, after a flooding event. The &FR model is more accurate than averaging individual observations before calculating the mean and variance or averaging the probabilities evaluated at the input values, which assigns the same weighted variation to each input distribution. The main disadvantage of these traditional methods is that the resulting measure of central tendency is exactly equal to the average of the input distribution’s means without the additional information provided by each individual distribution variance. When dealing with exponential distributions, such as resilience from severe flooding events and from nuisance flooding events, conflation results are equivalent to the weighted least squares method or best linear unbiased estimation. The combination of severe flooding risk with nuisance flooding improves flood risk management for highly populated coastal communities, such as in South Florida, USA, and provides a method to estimate community flood recovery time more accurately from two different sources, severe flooding events and nuisance flooding events.

Keywords: community resilience, conflation, flood risk, nuisance flooding

Procedia PDF Downloads 78
1541 Investigation of Dynamic Heat Transfer in Masonry Walls

Authors: Joelle Al Fakhoury, Emilio Sassine, Yassine Cherif, Joseph Dgheim, Emmanuel Antczak

Abstract:

Hollow block masonry is the most used building technology in the Lebanese context. These blocks are manufactured in an artisanal way and have unknown thermal properties; their overall thermos-physical performance is thus unknown and also poorly investigated scientifically in both single wall and also double wall configurations. In this work, experimental measurements and numerical simulations are performed for a better understanding of the heat transfer in masonry walls. This study was realized using an experimental setup consisting of a masonry hollow block wall (0.1m x 1m x 1m) and two heat boxes, such that each covers one side of the wall. The first is a reference box having a constant interior temperature, and the other is a control box having an adjustable interior temperature. At first, the numerical model is validated using an experimental setup; then 3D numerical analyzes are held in order to investigate the effect of the air gap, the mortar joints, and the plastering on the thermal performance of masonry walls for a better understanding of the heat transfer process and the recommendation of suitable thermal improvements.

Keywords: masonry wall, hollow blocks, heat transfer, wall instrumentation, thermal improvement

Procedia PDF Downloads 206
1540 Delineating Concern Ground in Block Caving – Underground Mine Using Ground Penetrating Radar

Authors: Eric Sitorus, Septian Prahastudhi, Turgod Nainggolan, Erwin Riyanto

Abstract:

Mining by block or panel caving is a mining method that takes advantage of fractures within an ore body, coupled with gravity, to extract material from a predetermined column of ore. The caving column is weakened from beneath through the use of undercutting, after which the ore breaks up and is extracted from below in a continuous cycle. The nature of this method induces cyclical stresses on the pillars of excavations as stress is built up and released over time, which has a detrimental effect on both the installed ground support and the rock mass itself. Ground support capacity, especially on the production where excavation void ratio is highest, is subjected to heavy loading. Strain above threshold of the elongation of support capacity can yield resulting in damage to excavations. Geotechnical engineers must evaluate not only the remnant capacity of ground support systems but also investigate depth of rock mass yield within pillars, backs and floors. Ground Penetrating Radar (GPR) is a geophysical method that has the ability to evaluate rock mass damage using electromagnetic waves. This paper illustrates a case study from the Grasberg mining complex where non-invasive information on the depth of damage and condition of the remaining rock mass was required. GPR with 100 MHz antenna resolution was used to obtain images of the subsurface to determine rehabilitation requirements prior to recommencing production activities. The GPR surveys were used to calibrate the reflection coefficient response of varying rock mass conditions to known Rock Quality Designation (RQD) parameters observed at the mine. The calibrated GPR survey allowed site engineers to map subsurface conditions and plan rehabilitation accordingly.

Keywords: block caving, ground penetrating radar, reflectivity, RQD

Procedia PDF Downloads 115
1539 Bit Error Rate Performance of MIMO Systems for Wireless Communications

Authors: E. Ghayoula, M. Haj Taieb, A. Bouallegue, J. Y. Chouinard, R. Ghayoula

Abstract:

This paper evaluates the bit error rate (BER) performance of MIMO systems for wireless communication. MIMO uses multiple transmitting antennas, multiple receiving antennas and the space-time block codes to provide diversity. MIMO transmits signal encoded by space-time block (STBC) encoder through different transmitting antennas. These signals arrive at the receiver at slightly different times. Spatially separated multiple receiving antennas are employed to provide diversity reception to combat the effect of fading in the channel. This paper presents a detailed study of diversity coding for MIMO systems. STBC techniques are implemented and simulation results in terms of the BER performance with varying number of MIMO transmitting and receiving antennas are presented. Our results show how increasing the number of both transmit and receive antenna improves system performance and reduces the bit error rate.

Keywords: MIMO systems, diversity, BER, MRRC, SIMO, MISO, STBC, alamouti, SNR

Procedia PDF Downloads 478
1538 Factors Affecting Cesarean Section among Women in Qatar Using Multiple Indicator Cluster Survey Database

Authors: Sahar Elsaleh, Ghada Farhat, Shaikha Al-Derham, Fasih Alam

Abstract:

Background: Cesarean section (CS) delivery is one of the major concerns both in developing and developed countries. The rate of CS deliveries are on the rise globally, and especially in Qatar. Many socio-economic, demographic, clinical and institutional factors play an important role for cesarean sections. This study aims to investigate factors affecting the prevalence of CS among women in Qatar using the UNICEF’s Multiple Indicator Cluster Survey (MICS) 2012 database. Methods: The study has focused on the women’s questionnaire of the MICS, which was successfully distributed to 5699 participants. Following study inclusion and exclusion criteria, a final sample of 761 women aged 19- 49 years who had at least one delivery of giving birth in their lifetime before the survey were included. A number of socio-economic, demographic, clinical and institutional factors, identified through literature review and available in the data, were considered for the analyses. Bivariate and multivariate logistic regression models, along with a multi-level modeling to investigate clustering effect, were undertaken to identify the factors that affect CS prevalence in Qatar. Results: From the bivariate analyses the study has shown that, a number of categorical factors are statistically significantly associated with the dependent variable (CS). When identifying the factors from a multivariate logistic regression, the study found that only three categorical factors -‘age of women’, ‘place at delivery’ and ‘baby weight’ appeared to be significantly affecting the CS among women in Qatar. Although the MICS dataset is based on a cluster survey, an exploratory multi-level analysis did not show any clustering effect, i.e. no significant variation in results at higher level (households), suggesting that all analyses at lower level (individual respondent) are valid without any significant bias in results. Conclusion: The study found a statistically significant association between the dependent variable (CS delivery) and age of women, frequency of TV watching, assistance at birth and place of birth. These results need to be interpreted cautiously; however, it can be used as evidence-base for further research on cesarean section delivery in Qatar.

Keywords: cesarean section, factors, multiple indicator cluster survey, MICS database, Qatar

Procedia PDF Downloads 97
1537 Parameter Estimation for the Mixture of Generalized Gamma Model

Authors: Wikanda Phaphan

Abstract:

Mixture generalized gamma distribution is a combination of two distributions: generalized gamma distribution and length biased generalized gamma distribution. These two distributions were presented by Suksaengrakcharoen and Bodhisuwan in 2014. The findings showed that probability density function (pdf) had fairly complexities, so it made problems in estimating parameters. The problem occurred in parameter estimation was that we were unable to calculate estimators in the form of critical expression. Thus, we will use numerical estimation to find the estimators. In this study, we presented a new method of the parameter estimation by using the expectation – maximization algorithm (EM), the conjugate gradient method, and the quasi-Newton method. The data was generated by acceptance-rejection method which is used for estimating α, β, λ and p. λ is the scale parameter, p is the weight parameter, α and β are the shape parameters. We will use Monte Carlo technique to find the estimator's performance. Determining the size of sample equals 10, 30, 100; the simulations were repeated 20 times in each case. We evaluated the effectiveness of the estimators which was introduced by considering values of the mean squared errors and the bias. The findings revealed that the EM-algorithm had proximity to the actual values determined. Also, the maximum likelihood estimators via the conjugate gradient and the quasi-Newton method are less precision than the maximum likelihood estimators via the EM-algorithm.

Keywords: conjugate gradient method, quasi-Newton method, EM-algorithm, generalized gamma distribution, length biased generalized gamma distribution, maximum likelihood method

Procedia PDF Downloads 201
1536 A Generalisation of Pearson's Curve System and Explicit Representation of the Associated Density Function

Authors: S. B. Provost, Hossein Zareamoghaddam

Abstract:

A univariate density approximation technique whereby the derivative of the logarithm of a density function is assumed to be expressible as a rational function is introduced. This approach which extends Pearson’s curve system is solely based on the moments of a distribution up to a determinable order. Upon solving a system of linear equations, the coefficients of the polynomial ratio can readily be identified. An explicit solution to the integral representation of the resulting density approximant is then obtained. It will be explained that when utilised in conjunction with sample moments, this methodology lends itself to the modelling of ‘big data’. Applications to sets of univariate and bivariate observations will be presented.

Keywords: density estimation, log-density, moments, Pearson's curve system

Procedia PDF Downloads 258