Search results for: water cycle algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13457

Search results for: water cycle algorithm

9797 The Diverse and Flexible Coping Strategies Simulation for Maanshan Nuclear Power Plant

Authors: Chin-Hsien Yeh, Shao-Wen Chen, Wen-Shu Huang, Chun-Fu Huang, Jong-Rong Wang, Jung-Hua Yang, Yuh-Ming Ferng, Chunkuan Shih

Abstract:

In this research, a Fukushima-like conditions is simulated with TRACE and RELAP5. Fukushima Daiichi Nuclear Power Plant (NPP) occurred the disaster which caused by the earthquake and tsunami. This disaster caused extended loss of all AC power (ELAP). Hence, loss of ultimate heat sink (LUHS) happened finally. In order to handle Fukushima-like conditions, Taiwan Atomic Energy Council (AEC) commanded that Taiwan Power Company should propose strategies to ensure the nuclear power plant safety. One of the diverse and flexible coping strategies (FLEX) is a different water injection strategy. It can execute core injection at 20 Kg/cm2 without depressurization. In this study, TRACE and RELAP5 were used to simulate Maanshan nuclear power plant, which is a three loops PWR in Taiwan, under Fukushima-like conditions and make sure the success criteria of FLEX. Reducing core cooling ability is due to failure of emergency core cooling system (ECCS) in extended loss of all AC power situation. The core water level continues to decline because of the seal leakage, and then FLEX is used to save the core water level and make fuel rods covered by water. The result shows that this mitigation strategy can cool the reactor pressure vessel (RPV) as soon as possible under Fukushima-like conditions, and keep the core water level higher than Top of Active Fuel (TAF). The FLEX can ensure the peak cladding temperature (PCT) below than the criteria 1088.7 K. Finally, the FLEX can provide protection for nuclear power plant and make plant safety.

Keywords: TRACE, RELAP5/MOD3.3, ELAP, FLEX

Procedia PDF Downloads 234
9796 Pattern Synthesis of Nonuniform Linear Arrays Including Mutual Coupling Effects Based on Gaussian Process Regression and Genetic Algorithm

Authors: Ming Su, Ziqiang Mu

Abstract:

This paper proposes a synthesis method for nonuniform linear antenna arrays that combine Gaussian process regression (GPR) and genetic algorithm (GA). In this method, the GPR model can be used to calculate the array radiation pattern in the presence of mutual coupling effects, and then the GA is used to optimize the excitations and locations of the elements so as to generate the desired radiation pattern. In this paper, taking a 9-element nonuniform linear array as an example and the desired radiation pattern corresponding to a Chebyshev distribution as the optimization objective, optimize the excitations and locations of the elements. Finally, the optimization results are verified by electromagnetic simulation software CST, which shows that the method is effective.

Keywords: nonuniform linear antenna arrays, GPR, GA, mutual coupling effects, active element pattern

Procedia PDF Downloads 94
9795 Habitat Studies of Etheria elliptica in Some Water Bodies (River Ogbese and Owena Reservoir) in Ondo State, Nigeria

Authors: O. O. Olawusi-Peters, M. O. Adediran, O. A. Ajibare

Abstract:

Etheria elliptica population is declining due to various human activities on the freshwater habitat. This necessitate the habitat study of the mussel in river Ogbese and Owena reservoir in Ondo state, Nigeria in order to know the status of the organism within the ecosystem. Thirty (30) specimens each from River Ogbese and Owena reservoir were sampled between May and August 2012. The meristic variables such as length, breadth, shell thickness and weight of the mussel were measured. Also, some physico-chemical parameters, flow rate and soil profile of the two rivers were studied. In River Ogbese, the weight, length, breadth and thickness variables obtained were; 49.73g, 8.42cm, 3.78cm and 0.53cm respectively. In Owena reservoir, the values were; 111.17g, 8.80cm, 6.64cm, 0.22cm respectively. The condition factor showed that the samples from Owena reservoir (K = 16.33) were healthier than River Ogbese (K = 8.34). Also, the length-weight relationship indicated isometric growth in both water bodies (Ogbese r2 = 0.68; Owena r2 = 0.66). In River Ogbese, the physico-chemical parameters obtained were; temperature (24.3oC), pH (7.12), TDS (72ppm), DO (3.2mg/l), conductivity (145µ), BOD (0.7mg/l). The mean temperature (24.1oC), pH (7.69), TDS (102ppm), DO (3.1mg/l), conductivity (183µ), BOD (0.8mg/l) were obtained from Owena reservoir. The soil samples values obtained from both water bodies are; River Ogbese –phosphorus; 78.78, calcium; 3.60, magnesium; 1.90 and organic matter; 0.17. Owena reservoir - Phosphorus; 3.34, calcium; 4.40, magnesium; 1.20 and organic matter; 0.66. The river flow rate was 0.22m/s for Owena reservoir and 0.26m/s for river Ogbese. The study revealed that Etheria elliptica in Owena reservoir and Ogbese were in good and healthy conditions despite the various human activities on the water bodies. The water quality parameters obtained were within the preferred requirements of the mussels.

Keywords: Etheria elliptica, mussels, Owena reservoir, River Ogbese

Procedia PDF Downloads 488
9794 Exergy Model for a Solar Water Heater with Flat Plate Collector

Authors: P. Sathyakala, G. Sai Sundara Krishnan

Abstract:

The objective of this paper is to derive an exergy model for a solar water heater with honey comb structure in order to identify the element which has larger irreversibility in the system. This will help us in finding the means to reduce the wasted work potential so that the overall efficiency of the system can be improved by finding the ways to reduce those wastages.

Keywords: exergy, energy balance, entropy balance, work potential, degradation, honey comb, flat plate collector

Procedia PDF Downloads 463
9793 Applying Hybrid Graph Drawing and Clustering Methods on Stock Investment Analysis

Authors: Mouataz Zreika, Maria Estela Varua

Abstract:

Stock investment decisions are often made based on current events of the global economy and the analysis of historical data. Conversely, visual representation could assist investors’ gain deeper understanding and better insight on stock market trends more efficiently. The trend analysis is based on long-term data collection. The study adopts a hybrid method that combines the Clustering algorithm and Force-directed algorithm to overcome the scalability problem when visualizing large data. This method exemplifies the potential relationships between each stock, as well as determining the degree of strength and connectivity, which will provide investors another understanding of the stock relationship for reference. Information derived from visualization will also help them make an informed decision. The results of the experiments show that the proposed method is able to produced visualized data aesthetically by providing clearer views for connectivity and edge weights.

Keywords: clustering, force-directed, graph drawing, stock investment analysis

Procedia PDF Downloads 288
9792 Water Heating System with Solar Energy from Solar Panel as Absorber to Reduce the Reduction of Efficiency Solar Panel Use

Authors: Mas Aji Rizki Widjayanto, Rizka Yunita

Abstract:

The building which has an efficient and low-energy today followed by the developers. It’s not because trends on the building nowaday, but rather because of its positive effects in the long term, where the cost of energy per month to be much cheaper, along with the high price of electricity. The use of solar power (Photovoltaic System) becomes one source of electrical energy for the apartment so that will efficiently use energy, water, and other resources in the operations of the apartment. However, more than 80% of the solar radiation is not converted into electrical energy, but reflected and converted into heat energy. This causes an increase on the working temperature of solar panels and consequently decrease the efficiency of conversion to electrical energy. The high temperature solar panels work caused by solar radiation can be used as medium heat exchanger or heating water for the apartments, so that the working temperature of the solar panel can be lowered to reduce the reduction on the efficiency of conversion to electrical energy.

Keywords: photovoltaic system, efficient, heat energy, heat exchanger, efficiency of conversion

Procedia PDF Downloads 338
9791 Cobalt Ions Adsorption by Quartz and Illite and Calcite from Waste Water

Authors: Saad A. Aljlil

Abstract:

Adsorption of cobalt ions on quartz and illite and calcite from waste water was investigated. The effect of pH on the adsorption of cobalt ions was studied. The maximum capacities of cobalt ions of the three adsorbents increase with increasing cobalt solution temperature. The maximum capacities were (4.66) mg/g for quartz, (3.94) mg/g for illite, and (3.44) mg/g for calcite. The enthalpy, Gibbs free energy, and entropy for adsorption of cobalt ions on the three adsorbents were calculated. It was found that the adsorption process of the cobalt ions of the adsorbent was an endothermic process. consequently increasing the temperature causes the increase of the cobalt ions adsorption of the adsorbents. Therefore, the adsorption process is preferred at high temperature levels. The equilibrium adsorption data were correlated using Langmuir model, Freundlich model. The experimental data of cobalt ions of the adsorbents correlated well with Freundlich model.

Keywords: adsorption, Langmuir, Freundlich, quartz, illite, calcite, waste water

Procedia PDF Downloads 355
9790 Seismic Response Control of Multi-Span Bridge Using Magnetorheological Dampers

Authors: B. Neethu, Diptesh Das

Abstract:

The present study investigates the performance of a semi-active controller using magneto-rheological dampers (MR) for seismic response reduction of a multi-span bridge. The application of structural control to the structures during earthquake excitation involves numerous challenges such as proper formulation and selection of the control strategy, mathematical modeling of the system, uncertainty in system parameters and noisy measurements. These problems, however, need to be tackled in order to design and develop controllers which will efficiently perform in such complex systems. A control algorithm, which can accommodate un-certainty and imprecision compared to all the other algorithms mentioned so far, due to its inherent robustness and ability to cope with the parameter uncertainties and imprecisions, is the sliding mode algorithm. A sliding mode control algorithm is adopted in the present study due to its inherent stability and distinguished robustness to system parameter variation and external disturbances. In general a semi-active control scheme using an MR damper requires two nested controllers: (i) an overall system controller, which derives the control force required to be applied to the structure and (ii) an MR damper voltage controller which determines the voltage required to be supplied to the damper in order to generate the desired control force. In the present study a sliding mode algorithm is used to determine the desired optimal force. The function of the voltage controller is to command the damper to produce the desired force. The clipped optimal algorithm is used to find the command voltage supplied to the MR damper which is regulated by a semi active control law based on sliding mode algorithm. The main objective of the study is to propose a robust semi active control which can effectively control the responses of the bridge under real earthquake ground motions. Lumped mass model of the bridge is developed and time history analysis is carried out by solving the governing equations of motion in the state space form. The effectiveness of MR dampers is studied by analytical simulations by subjecting the bridge to real earthquake records. In this regard, it may also be noted that the performance of controllers depends, to a great extent, on the characteristics of the input ground motions. Therefore, in order to study the robustness of the controller in the present study, the performance of the controllers have been investigated for fourteen different earthquake ground motion records. The earthquakes are chosen in such a way that all possible characteristic variations can be accommodated. Out of these fourteen earthquakes, seven are near-field and seven are far-field. Also, these earthquakes are divided into different frequency contents, viz, low-frequency, medium-frequency, and high-frequency earthquakes. The responses of the controlled bridge are compared with the responses of the corresponding uncontrolled bridge (i.e., the bridge without any control devices). The results of the numerical study show that the sliding mode based semi-active control strategy can substantially reduce the seismic responses of the bridge showing a stable and robust performance for all the earthquakes.

Keywords: bridge, semi active control, sliding mode control, MR damper

Procedia PDF Downloads 115
9789 A Local Tensor Clustering Algorithm to Annotate Uncharacterized Genes with Many Biological Networks

Authors: Paul Shize Li, Frank Alber

Abstract:

A fundamental task of clinical genomics is to unravel the functions of genes and their associations with disorders. Although experimental biology has made efforts to discover and elucidate the molecular mechanisms of individual genes in the past decades, still about 40% of human genes have unknown functions, not to mention the diseases they may be related to. For those biologists who are interested in a particular gene with unknown functions, a powerful computational method tailored for inferring the functions and disease relevance of uncharacterized genes is strongly needed. Studies have shown that genes strongly linked to each other in multiple biological networks are more likely to have similar functions. This indicates that the densely connected subgraphs in multiple biological networks are useful in the functional and phenotypic annotation of uncharacterized genes. Therefore, in this work, we have developed an integrative network approach to identify the frequent local clusters, which are defined as those densely connected subgraphs that frequently occur in multiple biological networks and consist of the query gene that has few or no disease or function annotations. This is a local clustering algorithm that models multiple biological networks sharing the same gene set as a three-dimensional matrix, the so-called tensor, and employs the tensor-based optimization method to efficiently find the frequent local clusters. Specifically, massive public gene expression data sets that comprehensively cover dynamic, physiological, and environmental conditions are used to generate hundreds of gene co-expression networks. By integrating these gene co-expression networks, for a given uncharacterized gene that is of biologist’s interest, the proposed method can be applied to identify the frequent local clusters that consist of this uncharacterized gene. Finally, those frequent local clusters are used for function and disease annotation of this uncharacterized gene. This local tensor clustering algorithm outperformed the competing tensor-based algorithm in both module discovery and running time. We also demonstrated the use of the proposed method on real data of hundreds of gene co-expression data and showed that it can comprehensively characterize the query gene. Therefore, this study provides a new tool for annotating the uncharacterized genes and has great potential to assist clinical genomic diagnostics.

Keywords: local tensor clustering, query gene, gene co-expression network, gene annotation

Procedia PDF Downloads 122
9788 Elephant Herding Optimization for Service Selection in QoS-Aware Web Service Composition

Authors: Samia Sadouki Chibani, Abdelkamel Tari

Abstract:

Web service composition combines available services to provide new functionality. Given the number of available services with similar functionalities and different non functional aspects (QoS), the problem of finding a QoS-optimal web service composition is considered as an optimization problem belonging to NP-hard class. Thus, an optimal solution cannot be found by exact algorithms within a reasonable time. In this paper, a meta-heuristic bio-inspired is presented to address the QoS aware web service composition; it is based on Elephant Herding Optimization (EHO) algorithm, which is inspired by the herding behavior of elephant group. EHO is characterized by a process of dividing and combining the population to sub populations (clan); this process allows the exchange of information between local searches to move toward a global optimum. However, with Applying others evolutionary algorithms the problem of early stagnancy in a local optimum cannot be avoided. Compared with PSO, the results of experimental evaluation show that our proposition significantly outperforms the existing algorithm with better performance of the fitness value and a fast convergence.

Keywords: bio-inspired algorithms, elephant herding optimization, QoS optimization, web service composition

Procedia PDF Downloads 309
9787 Alphabet Recognition Using Pixel Probability Distribution

Authors: Vaidehi Murarka, Sneha Mehta, Dishant Upadhyay

Abstract:

Our project topic is “Alphabet Recognition using pixel probability distribution”. The project uses techniques of Image Processing and Machine Learning in Computer Vision. Alphabet recognition is the mechanical or electronic translation of scanned images of handwritten, typewritten or printed text into machine-encoded text. It is widely used to convert books and documents into electronic files etc. Alphabet Recognition based OCR application is sometimes used in signature recognition which is used in bank and other high security buildings. One of the popular mobile applications includes reading a visiting card and directly storing it to the contacts. OCR's are known to be used in radar systems for reading speeders license plates and lots of other things. The implementation of our project has been done using Visual Studio and Open CV (Open Source Computer Vision). Our algorithm is based on Neural Networks (machine learning). The project was implemented in three modules: (1) Training: This module aims “Database Generation”. Database was generated using two methods: (a) Run-time generation included database generation at compilation time using inbuilt fonts of OpenCV library. Human intervention is not necessary for generating this database. (b) Contour–detection: ‘jpeg’ template containing different fonts of an alphabet is converted to the weighted matrix using specialized functions (contour detection and blob detection) of OpenCV. The main advantage of this type of database generation is that the algorithm becomes self-learning and the final database requires little memory to be stored (119kb precisely). (2) Preprocessing: Input image is pre-processed using image processing concepts such as adaptive thresholding, binarizing, dilating etc. and is made ready for segmentation. “Segmentation” includes extraction of lines, words, and letters from the processed text image. (3) Testing and prediction: The extracted letters are classified and predicted using the neural networks algorithm. The algorithm recognizes an alphabet based on certain mathematical parameters calculated using the database and weight matrix of the segmented image.

Keywords: contour-detection, neural networks, pre-processing, recognition coefficient, runtime-template generation, segmentation, weight matrix

Procedia PDF Downloads 368
9786 Material Use & Life cycle GHG Emissions of Different Electrification Options for Long-Haul Trucks

Authors: Nafisa Mahbub, Hajo Ribberink

Abstract:

Electrification of long-haul trucks has been in discussion as a potential strategy to decarbonization. These trucks will require large batteries because of their weight and long daily driving distances. Around 245 million battery electric vehicles are predicted to be on the road by the year 2035. This huge increase in the number of electric vehicles (EVs) will require intensive mining operations for metals and other materials to manufacture millions of batteries for the EVs. These operations will add significant environmental burdens and there is a significant risk that the mining sector will not be able to meet the demand for battery materials, leading to higher prices. Since the battery is the most expensive component in the EVs, technologies that can enable electrification with smaller batteries sizes have substantial potential to reduce the material usage and associated environmental and cost burdens. One of these technologies is an ‘electrified road’ (eroad), where vehicles receive power while they are driving, for instance through an overhead catenary (OC) wire (like trolleybuses and electric trains), through wireless (inductive) chargers embedded in the road, or by connecting to an electrified rail in or on the road surface. This study assessed the total material use and associated life cycle GHG emissions of two types of eroads (overhead catenary and in-road wireless charging) for long-haul trucks in Canada and compared them to electrification using stationary plug-in fast charging. As different electrification technologies require different amounts of materials for charging infrastructure and for the truck batteries, the study included the contributions of both for the total material use. The study developed a bottom-up approach model comparing the three different charging scenarios – plug in fast chargers, overhead catenary and in-road wireless charging. The investigated materials for charging technology and batteries were copper (Cu), steel (Fe), aluminium (Al), and lithium (Li). For the plug-in fast charging technology, different charging scenarios ranging from overnight charging (350 kW) to megawatt (MW) charging (2 MW) were investigated. A 500 km of highway (1 lane of in-road charging per direction) was considered to estimate the material use for the overhead catenary and inductive charging technologies. The study considered trucks needing an 800 kWh battery under the plug-in charger scenario but only a 200 kWh battery for the OC and inductive charging scenarios. Results showed that overall the inductive charging scenario has the lowest material use followed by OC and plug-in charger scenarios respectively. The materials use for the OC and plug-in charger scenarios were 50-70% higher than for the inductive charging scenarios for the overall system including the charging infrastructure and battery. The life cycle GHG emissions from the construction and installation of the charging technology material were also investigated.

Keywords: charging technology, eroad, GHG emissions, material use, overhead catenary, plug in charger

Procedia PDF Downloads 34
9785 Evaluation of Risk and the Beneficial Effects of Synthesized Nano Silver-Based Disinfectant on Poultry Mortality and Health

Authors: Indrajeet Kumar, Jayanta Bhattacharya

Abstract:

This study was evaluated for the potential use of nanosilver (nAg) as a disinfectant and antimicrobial growth promoter supplement for the poultry. The experiments were conducted in the Kangsabati river basin region, in West Medinipur district, West Bengal, India for six months. Two poultry farms were adopted for the experiment. The rural economy of this region from Jhargram to Barkola is heavily dependent on contract poultry farming. The water samples were collected from the water source of poultry farm which has been used for poultry drinking purpose. The bacteriological analysis of water sample revealed that the total bacterial count (total coliform and E. coli) were higher than the acceptable standards. The bacterial loads badly affected the growth performance and health of the poultry. For disinfection, a number of chemical compounds (like formaldehyde, calcium hypochloride, sodium hypochloride, and sodium bicarbonate) have been used in typical commercial formulations. However, the effects of all these chemical compounds have not been significant over time. As a part of our research-to-market initiative, we used nanosilver (nAg) formulation as a disinfectant. The nAg formulation was synthesized by hydrothermal technique and characterized by UV-visible, TEM, SEM, and EDX. The obtained results revealed that the mortality rate of poultry was reduced due to nAg formulation compared to the mortality rate of the negative control. Moreover, the income of the farmer family was increased by 10-20% due to less mortality and better health of the poultry.

Keywords: farm water, nanosilver, field application, and poultry performance

Procedia PDF Downloads 140
9784 Simulation of Flood Inundation in Kedukan River Using HEC-RAS and GIS

Authors: Reini S. Ilmiaty, Muhammad B. Al Amin, Sarino, Muzamil Jariski

Abstract:

Kedukan River is an artificial river which serves as a Watershed Boang drainage channel in Palembang. The river has upstream and downstream connected to Musi River, that often overflowing and flooding caused by the huge runoff discharge and high tide water level of Musi River. This study aimed to analyze the flood water surface profile on Kedukan River continued with flood inundation simulation to determine flooding prone areas in research area. The analysis starts from the peak runoff discharge calculations using rational method followed by water surface profile analysis using HEC-RAS program controlled by manual calculations using standard stages. The analysis followed by running flood inundation simulation using ArcGIS program that has been integrated with HEC-GeoRAS. Flood inundation simulation on Kedukan River creates inundation characteristic maps with depth, area, and circumference of inundation as the parameters. The inundation maps are very useful in providing an overview of flood prone areas in Kedukan River.

Keywords: flood modelling, HEC-GeoRAS, HEC-RAS, inundation map

Procedia PDF Downloads 493
9783 Detecting Tomato Flowers in Greenhouses Using Computer Vision

Authors: Dor Oppenheim, Yael Edan, Guy Shani

Abstract:

This paper presents an image analysis algorithm to detect and count yellow tomato flowers in a greenhouse with uneven illumination conditions, complex growth conditions and different flower sizes. The algorithm is designed to be employed on a drone that flies in greenhouses to accomplish several tasks such as pollination and yield estimation. Detecting the flowers can provide useful information for the farmer, such as the number of flowers in a row, and the number of flowers that were pollinated since the last visit to the row. The developed algorithm is designed to handle the real world difficulties in a greenhouse which include varying lighting conditions, shadowing, and occlusion, while considering the computational limitations of the simple processor in the drone. The algorithm identifies flowers using an adaptive global threshold, segmentation over the HSV color space, and morphological cues. The adaptive threshold divides the images into darker and lighter images. Then, segmentation on the hue, saturation and volume is performed accordingly, and classification is done according to size and location of the flowers. 1069 images of greenhouse tomato flowers were acquired in a commercial greenhouse in Israel, using two different RGB Cameras – an LG G4 smartphone and a Canon PowerShot A590. The images were acquired from multiple angles and distances and were sampled manually at various periods along the day to obtain varying lighting conditions. Ground truth was created by manually tagging approximately 25,000 individual flowers in the images. Sensitivity analyses on the acquisition angle of the images, periods throughout the day, different cameras and thresholding types were performed. Precision, recall and their derived F1 score were calculated. Results indicate better performance for the view angle facing the flowers than any other angle. Acquiring images in the afternoon resulted with the best precision and recall results. Applying a global adaptive threshold improved the median F1 score by 3%. Results showed no difference between the two cameras used. Using hue values of 0.12-0.18 in the segmentation process provided the best results in precision and recall, and the best F1 score. The precision and recall average for all the images when using these values was 74% and 75% respectively with an F1 score of 0.73. Further analysis showed a 5% increase in precision and recall when analyzing images acquired in the afternoon and from the front viewpoint.

Keywords: agricultural engineering, image processing, computer vision, flower detection

Procedia PDF Downloads 305
9782 Exploring the Role of Building Information Modeling for Delivering Successful Construction Projects

Authors: Muhammad Abu Bakar Tariq

Abstract:

Construction industry plays a crucial role in the progress of societies and economies. Furthermore, construction projects have social as well as economic implications, thus, their success/failure have wider impacts. However, the industry is lagging behind in terms of efficiency and productivity. Building Information Modeling (BIM) is recognized as a revolutionary development in Architecture, Engineering and Construction (AEC) industry. There are numerous interest groups around the world providing definitions of BIM, proponents describing its advantages and opponents identifying challenges/barriers regarding adoption of BIM. This research is aimed at to determine what actually BIM is, along with its potential role in delivering successful construction projects. The methodology is critical analysis of secondary data sources i.e. information present in public domain, which include peer reviewed journal articles, industry and government reports, conference papers, books, case studies etc. It is discovered that clash detection and visualization are two major advantages of BIM. Clash detection option identifies clashes among structural, architectural and MEP designs before construction actually commences, which subsequently saves time as well as cost and ensures quality during execution phase of a project. Visualization is a powerful tool that facilitates in rapid decision-making in addition to communication and coordination among stakeholders throughout project’s life cycle. By eliminating inconsistencies that consume time besides cost during actual construction, improving collaboration among stakeholders throughout project’s life cycle, BIM can play a positive role to achieve efficiency and productivity that consequently deliver successful construction projects.

Keywords: building information modeling, clash detection, construction project success, visualization

Procedia PDF Downloads 239
9781 Environmental Impact Assessment of Municipal Solid Waste Disposal Site in Shahrood City

Authors: Mehri Bagherkazemi, Naser Hafezi Moghaddas

Abstract:

This study investigates the environmental impact of the disposal site located in Shahrood city, focusing on the geological characteristics of the region. Shahrood's disposal site primarily consists of limestone bedrock, overlaid by substantial alluvial deposits. The area's highly permeable soil is anticipated to have a significant influence on groundwater pollution. Spanning 52 hectares, the Shahrood disposal site is situated in the eastern sector of the city. Historically, waste disposal took place on the surface, but recent practices involve on-site trenching. This research involved the collection of soil and water samples near the disposal site, with subsequent analysis of 11 soil samples and 3 water samples. The soil's particle size distribution was determined, and comprehensive analyses were conducted for 35 elements in the soil and 42 elements in water. The study combines the results of these tests with field investigations to evaluate the landfill's impact on the surrounding soil and groundwater contamination.

Keywords: environmental geology, environmental impact assessment, disposal site, heavy metals contamination

Procedia PDF Downloads 55
9780 Investigation of Water Absorption and Compressive Strength of Resin Coated Mortar

Authors: Yasir Ali, Zain Ul Abdin, Muhammad Wisal Khattak

Abstract:

Nowadays various advanced techniques are used to enhance the performance of materials in the field of construction engineering. Structures exposed to an aggressive, humid and hostile environment are experiencing severe negative impacts which lead to premature failure. Polyester resin is one of the advanced material used for improving performance of structural materials especially for repair/ refurbish purpose of structures and protection from contaminated environmental effect/ hazards. This study investigated the aptness of the polyester resin as coating agent on the mortar and assessed its performance in an ambient environment of Pakistan. Cubical specimens of mortar were fabricated. These specimens were tested for water absorption and compressive strength after one day and sixty days. These tests were performed under different exposure conditions (ambient environment and submerged in water). The specimens were coated with one, two and three layers and results were compared to control (no/ zero resin layer) specimens. Test results indicated that there is a significant decrease in water absorption of mortar coated with resin when compared to controlled specimens. The compressive strength test results revealed that resin coated specimen had higher strength when compared to controlled specimens. The results suggested that resin is a promising material and can be used effectively in structures which are exposed to high temperatures. The study would be helpful in improving performance of the structural material in a hazardous environment.

Keywords: ambient environment, coating, mortar, polyester resin

Procedia PDF Downloads 343
9779 Fabrication of Durable and Renegerable Superhydrophobic Coatings on Metallic Surfaces for Potential Industrial Applications

Authors: Priya Varshney, Soumya S. Mohapatra

Abstract:

Fabrication of anti-corrosion and self-cleaning superhydrophobic coatings for metallic surfaces which are regenerable and durable in the aggressive conditions has shown tremendous interest in materials science. In this work, the superhydrophobic coatings on metallic surfaces (aluminum, steel, copper) were prepared by two-step and one-step chemical etching process. In two-step process, roughness on surface was created by chemical etching and then passivation of roughened surface with low surface energy materials whereas, in one-step process, roughness on surface by chemical etching and passivation of surface with low surface energy materials were done in a single step. Beside this, the effect of etchant concentration and etching time on wettability and morphology was also studied. Thermal, mechanical, ultra-violet stability of these coatings were also tested. Along with this, regeneration of coatings and self-cleaning, corrosion resistance and water repelling characteristics were also studied. The surface morphology shows the presence of a rough microstuctures on the treated surfaces and the contact angle measurements confirms the superhydrophobic nature. It is experimentally observed that the surface roughness and contact angle increases with increase in etching time as well as with concentration of etchant. Superhydrophobic surfaces show the excellent self-cleaning behaviour. Coatings are found to be stable and maintain their superhydrophobicity in acidic and alkaline solutions. Water jet impact, floatation on water surface, and low temperature condensation tests prove the water-repellent nature of the coatings. These coatings are found to be thermal, mechanical and ultra-violet stable. These durable superhydrophobic metallic surfaces have potential industrial applications.

Keywords: superhydrophobic, water-repellent, anti-corrosion, self-cleaning

Procedia PDF Downloads 261
9778 Analysis of Financial Time Series by Using Ornstein-Uhlenbeck Type Models

Authors: Md Al Masum Bhuiyan, Maria C. Mariani, Osei K. Tweneboah

Abstract:

In the present work, we develop a technique for estimating the volatility of financial time series by using stochastic differential equation. Taking the daily closing prices from developed and emergent stock markets as the basis, we argue that the incorporation of stochastic volatility into the time-varying parameter estimation significantly improves the forecasting performance via Maximum Likelihood Estimation. While using the technique, we see the long-memory behavior of data sets and one-step-ahead-predicted log-volatility with ±2 standard errors despite the variation of the observed noise from a Normal mixture distribution, because the financial data studied is not fully Gaussian. Also, the Ornstein-Uhlenbeck process followed in this work simulates well the financial time series, which aligns our estimation algorithm with large data sets due to the fact that this algorithm has good convergence properties.

Keywords: financial time series, maximum likelihood estimation, Ornstein-Uhlenbeck type models, stochastic volatility model

Procedia PDF Downloads 222
9777 Efficacy of Preimplantation Genetic Screening in Women with a Spontaneous Abortion History with Eukaryotic or Aneuploidy Abortus

Authors: Jayeon Kim, Eunjung Yu, Taeki Yoon

Abstract:

Most spontaneous miscarriage is believed to be a consequence of embryo aneuploidies. Transferring eukaryotic embryos selected by PGS is expected to decrease the miscarriage rate. Current PGS indications include advanced maternal age, recurrent pregnancy loss, repeated implantation failure. Recently, use of PGS for healthy women without above indications for the purpose of improving in vitro fertilization (IVF) outcomes is on the rise. However, it is still controversy about the beneficial effect of PGS in this population, especially, in women with a history of no more than 2 miscarriages or miscarriage of eukaryotic abortus. This study aimed to investigate if karyotyping result of abortus is a good indicator of preimplantation genetic screening (PGS) in subsequent IVF cycle in women with a history of spontaneous abortion. A single-center retrospective cohort study was performed. Women who had spontaneous abortion(s) (less than 3) and dilatation and evacuation, and subsequent IVF from January 2016 to November 2016 were included. Their medical information was extracted from the charts. Clinical pregnancy was defined as presence of a gestational sac with fetal heart beat detected on ultrasound in week 7. Statistical analysis was performed using SPSS software. Total 234 women were included. 121 out of 234 (51.7%) underwent karyotyping of the abortus, and 113 did not have the abortus karyotyped. Embryo biopsy was performed on 3 or 5 days after oocyte retrieval, followed by embryo transfer (ET) on a fresh or frozen cycle. The biopsied materials were subjected to microarray comparative genomic hybridization. Clinical pregnancy rate per ET was compared between PGS and non-PGS group in each study group. Patients were grouped by two criteria: karyotype of the abortus from previous miscarriage (unknown fetal karyotype (n=89, Group 1), eukaryotic abortus (n=36, Group 2) or aneuploidy abortus (n=67, Group 3)), and pursuing PGS in subsequent IVF cycle (pursuing PGS (PGS group, n=105) or not pursuing PGS (non-PGS group, n=87)). The PGS group was significantly older and had higher number of retrieved oocytes and prior miscarriages compared to non-PGS group. There were no differences in BMI and AMH level between those two groups. In PGS group, the mean number of transferable embryos (eukaryotic embryo) was 1.3 ± 0.7, 1.5 ± 0.5 and 1.4 ± 0.5, respectively (p = 0.049). In 42 cases, ET was cancelled because all embryos biopsied turned out to be abnormal. In all three groups (group 1, 2, and 3), clinical pregnancy rates were not statistically different between PGS and non-PGS group (Group 1: 48.8% vs. 52.2% (p=0.858), Group 2: 70% vs. 73.1% (p=0.730), Group 3: 42.3% vs. 46.7% (p=0.640), in PGS and non-PGS group, respectively). In both groups who had miscarriage with eukaryotic and aneuploidy abortus, the clinical pregnancy rate between IVF cycles with and without PGS was not different. When we compare miscarriage and ongoing pregnancy rate, there were no significant differences between PGS and non-PGS group in all three groups. Our results show that the routine application of PGS in women who had less than 3 miscarriages would not be beneficial, even in cases that previous miscarriage had been caused by fetal aneuploidy.

Keywords: preimplantation genetic diagnosis, miscarriage, kpryotyping, in vitro fertilization

Procedia PDF Downloads 161
9776 A Next-Generation Pin-On-Plate Tribometer for Use in Arthroplasty Material Performance Research

Authors: Lewis J. Woollin, Robert I. Davidson, Paul Watson, Philip J. Hyde

Abstract:

Introduction: In-vitro testing of arthroplasty materials is of paramount importance when ensuring that they can withstand the performance requirements encountered in-vivo. One common machine used for in-vitro testing is a pin-on-plate tribometer, an early stage screening device that generates data on the wear characteristics of arthroplasty bearing materials. These devices test vertically loaded rotating cylindrical pins acting against reciprocating plates, representing the bearing surfaces. In this study, a pin-on-plate machine has been developed that provides several improvements over current technology, thereby progressing arthroplasty bearing research. Historically, pin-on-plate tribometers have been used to investigate the performance of arthroplasty bearing materials under conditions commonly encountered during a standard gait cycle; nominal operating pressures of 2-6 MPa and an operating frequency of 1 Hz are typical. There has been increased interest in using pin-on-plate machines to test more representative in-vivo conditions, due to the drive to test 'beyond compliance', as well as their testing speed and economic advantages over hip simulators. Current pin-on-plate machines do not accommodate the increased performance requirements associated with more extreme kinematic conditions, therefore a next-generation pin-on-plate tribometer has been developed to bridge the gap between current technology and future research requirements. Methodology: The design was driven by several physiologically relevant requirements. Firstly, an increased loading capacity was essential to replicate the peak pressures that occur in the natural hip joint during running and chair-rising, as well as increasing the understanding of wear rates in obese patients. Secondly, the introduction of mid-cycle load variation was of paramount importance, as this allows for an approximation of the loads present in a gait cycle to be applied and to test the fatigue properties of materials. Finally, the rig must be validated against previous-generation pin-on-plate and arthroplasty wear data. Results: The resulting machine is a twelve station device that is split into three sets of four stations, providing an increased testing capacity compared to most current pin-on-plate tribometers. The loading of the pins is generated using a pneumatic system, which can produce contact pressures of up to 201 MPa on a 3.2 mm² round pin face. This greatly exceeds currently achievable contact pressures in literature and opens new research avenues such as testing rim wear of mal-positioned hip implants. Additionally, the contact pressure of each set can be changed independently of the others, allowing multiple loading conditions to be tested simultaneously. Using pneumatics also allows the applied pressure to be switched ON/OFF mid-cycle, another feature not currently reported elsewhere, which allows for investigation into intermittent loading and material fatigue. The device is currently undergoing a series of validation tests using Ultra-High-Molecular-Weight-Polyethylene pins and 316L Stainless Steel Plates (polished to a Ra < 0.05 µm). The operating pressures will be between 2-6 MPa, operating at 1 Hz, allowing for validation of the machine against results reported previously in the literature. The successful production of this next-generation pin-on-plate tribometer will, following its validation, unlock multiple previously unavailable research avenues.

Keywords: arthroplasty, mechanical design, pin-on-plate, total joint replacement, wear testing

Procedia PDF Downloads 80
9775 Optimal Reactive Power Dispatch under Various Contingency Conditions Using Whale Optimization Algorithm

Authors: Khaled Ben Oualid Medani, Samir Sayah

Abstract:

The Optimal Reactive Power Dispatch (ORPD) problem has been solved and analysed usually in the normal conditions. However, network collapses appear in contingency conditions. In this paper, ORPD under several contingencies is presented using the proposed method WOA. To ensure viability of the power system in contingency conditions, several critical cases are simulated in order to prevent and prepare the power system to face such situations. The results obtained are carried out in IEEE 30 bus test system for the solution of ORPD problem in which control of bus voltages, tap position of transformers and reactive power sources are involved. Moreover, another method, namely, Particle Swarm Optimization with Time Varying Acceleration Coefficient (PSO-TVAC) has been compared with the proposed technique. Simulation results indicate that the proposed WOA gives remarkable solution in terms of effectiveness in case of outages.

Keywords: optimal reactive power dispatch, power system analysis, real power loss minimization, contingency condition, metaheuristic technique, whale optimization algorithm

Procedia PDF Downloads 103
9774 An Assessment of Floodplain Vegetation Response to Groundwater Changes Using the Soil & Water Assessment Tool Hydrological Model, Geographic Information System, and Machine Learning in the Southeast Australian River Basin

Authors: Newton Muhury, Armando A. Apan, Tek N. Marasani, Gebiaw T. Ayele

Abstract:

The changing climate has degraded freshwater availability in Australia that influencing vegetation growth to a great extent. This study assessed the vegetation responses to groundwater using Terra’s moderate resolution imaging spectroradiometer (MODIS), Normalised Difference Vegetation Index (NDVI), and soil water content (SWC). A hydrological model, SWAT, has been set up in a southeast Australian river catchment for groundwater analysis. The model was calibrated and validated against monthly streamflow from 2001 to 2006 and 2007 to 2010, respectively. The SWAT simulated soil water content for 43 sub-basins and monthly MODIS NDVI data for three different types of vegetation (forest, shrub, and grass) were applied in the machine learning tool, Waikato Environment for Knowledge Analysis (WEKA), using two supervised machine learning algorithms, i.e., support vector machine (SVM) and random forest (RF). The assessment shows that different types of vegetation response and soil water content vary in the dry and wet seasons. The WEKA model generated high positive relationships (r = 0.76, 0.73, and 0.81) between NDVI values of all vegetation in the sub-basins against soil water content (SWC), the groundwater flow (GW), and the combination of these two variables, respectively, during the dry season. However, these responses were reduced by 36.8% (r = 0.48) and 13.6% (r = 0.63) against GW and SWC, respectively, in the wet season. Although the rainfall pattern is highly variable in the study area, the summer rainfall is very effective for the growth of the grass vegetation type. This study has enriched our knowledge of vegetation responses to groundwater in each season, which will facilitate better floodplain vegetation management.

Keywords: ArcSWAT, machine learning, floodplain vegetation, MODIS NDVI, groundwater

Procedia PDF Downloads 81
9773 Concentration of Waste Waters by Enzyme-Assisted Low-Temperature Evaporation

Authors: Ahokas Mikko, Taskila Sanna, Varrio Kalle, Tanskanen Juha

Abstract:

The present research aimed at the development of an energy efficient process for the concentration of starchy waste waters. The selected principle is mechanical vapor recompression evaporation (MVR) which leads to concentrated solid material and evaporated water phase. Evaporation removes water until a certain viscosity limit is reached. Materials with high viscosity cannot be concentrated using standard evaporators due to limitations of pumps and other constraints, such as wetting. Control of viscosity is thus essential for efficient evaporation. This applies especially to fluids in which due starch or other compounds the viscosity tends to increase via removal of water. In the present research, the effect of enzymes on evaporation of highly viscous starch industry waste waters was investigated. Wastewater samples were received from starch industry at pH of 4.8. Response surface methodology (RSM) was applied for the investigation of factor effects on the behaviour of concentrate during evaporation. The RSM was prepared using quadratic face-centered central composite design (CCF). The evaporation performance was evaluated by monitoring the viscosity of fluid during processing. Based on viscosity curves, the addition of glucoamylase reduced the viscosity during evaporation. This assumption was confirmed by CCF, suggesting that the use of starch decomposing glucoamylase allowed evaporation of the starchy wastewater to a relatively high total solid concentration without a detrimental increase in the viscosity. The results suggest that use of enzymes for reduction of viscosity during the evaporation allows more effective concentration of the wastewater and thereby recovery of potable water.

Keywords: viscous, wastewater, treatment, evaporation, concentration

Procedia PDF Downloads 228
9772 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 176
9771 Preparation of Corn Flour Based Extruded Product and Evaluate Its Physical Characteristics

Authors: C. S. Saini

Abstract:

The composite flour blend consisting of corn, pearl millet, black gram and wheat bran in the ratio of 80:5:10:5 was taken to prepare the extruded product and their effect on physical properties of extrudate was studied. The extrusion process was conducted in laboratory by using twin screw extruder. The physical characteristics evaluated include lateral expansion, bulk density, water absorption index, water solubility index, rehydration ratio and moisture retention. The Central Composite Rotatable Design (CCRD) was used to decide the level of processing variables i.e. feed moisture content (%), screw speed (rpm), and barrel temperature (oC) for the experiment. The data obtained after extrusion process were analyzed by using response surface methodology. A second order polynomial model for the dependent variables was established to fit the experimental data. The numerical optimization studies resulted in 127°C of barrel temperature, 246 rpm of screw speed, and 14.5% of feed moisture as optimum variables to produce acceptable extruded product. The responses predicted by the software for the optimum process condition resulted in lateral expansion 126 %, bulk density 0.28 g/cm3, water absorption index 4.10 g/g, water solubility index 39.90 %, rehydration ratio 544 % and moisture retention 11.90 % with 75 % desirability.

Keywords: black gram, corn flour, extrusion, physical characteristics

Procedia PDF Downloads 461
9770 Image Compression on Region of Interest Based on SPIHT Algorithm

Authors: Sudeepti Dayal, Neelesh Gupta

Abstract:

Image abbreviation is utilized for reducing the size of a file without demeaning the quality of the image to an objectionable level. The depletion in file size permits more images to be deposited in a given number of spaces. It also minimizes the time necessary for images to be transferred. Storage of medical images is a most researched area in the current scenario. To store a medical image, there are two parameters on which the image is divided, regions of interest and non-regions of interest. The best way to store an image is to compress it in such a way that no important information is lost. Compression can be done in two ways, namely lossy, and lossless compression. Under that, several compression algorithms are applied. In the paper, two algorithms are used which are, discrete cosine transform, applied to non-region of interest (lossy), and discrete wavelet transform, applied to regions of interest (lossless). The paper introduces SPIHT (set partitioning hierarchical tree) algorithm which is applied onto the wavelet transform to obtain good compression ratio from which an image can be stored efficiently.

Keywords: Compression ratio, DWT, SPIHT, DCT

Procedia PDF Downloads 331
9769 Spectrum Assignment Algorithms in Optical Networks with Protection

Authors: Qusay Alghazali, Tibor Cinkler, Abdulhalim Fayad

Abstract:

In modern optical networks, the flex grid spectrum usage is most widespread, where higher bit rate streams get larger spectrum slices while lower bit rate traffic streams get smaller spectrum slices. To our practice, under the ITU-T recommendation, G.694.1, spectrum slices of 50, 75, and 100 GHz are being used with central frequency at 193.1 THz. However, when these spectrum slices are not sufficient, multiple spectrum slices can use either one next to another or anywhere in the optical wavelength. In this paper, we propose the analysis of the wavelength assignment problem. We compare different algorithms for this spectrum assignment with and without protection. As a reference for comparisons, we concluded that the Integer Linear Programming (ILP) provides the global optimum for all cases. The most scalable algorithm is the greedy one, which yields results in subsequent ranges even for more significant network instances. The algorithms’ benchmark implemented using the LEMON C++ optimization library and simulation runs based on a minimum number of spectrum slices assigned to lightpaths and their execution time.

Keywords: spectrum assignment, integer linear programming, greedy algorithm, international telecommunication union, library for efficient modeling and optimization in networks

Procedia PDF Downloads 156
9768 Modeling Floodplain Vegetation Response to Groundwater Variability Using ArcSWAT Hydrological Model, Moderate Resolution Imaging Spectroradiometer - Normalised Difference Vegetation Index Data, and Machine Learning

Authors: Newton Muhury, Armando A. Apan, Tek Maraseni

Abstract:

This study modelled the relationships between vegetation response and available water below the soil surface using the Terra’s Moderate Resolution Imaging Spectroradiometer (MODIS) generated Normalised Difference Vegetation Index (NDVI) and soil water content (SWC) data. The Soil & Water Assessment Tool (SWAT) interface known as ArcSWAT was used in ArcGIS for the groundwater analysis. The SWAT model was calibrated and validated in SWAT-CUP software using 10 years (2001-2010) of monthly streamflow data. The average Nash-Sutcliffe Efficiency during the calibration and validation was 0.54 and 0.51, respectively, indicating that the model performances were good. Twenty years (2001-2020) of monthly MODIS NDVI data for three different types of vegetation (forest, shrub, and grass) and soil water content for 43 sub-basins were analysed using the WEKA, machine learning tool with a selection of two supervised machine learning algorithms, i.e., support vector machine (SVM) and random forest (RF). The modelling results show that different types of vegetation response and soil water content vary in the dry and wet season. For example, the model generated high positive relationships (r=0.76, 0.73, and 0.81) between the measured and predicted NDVI values of all vegetation in the study area against the groundwater flow (GW), soil water content (SWC), and the combination of these two variables, respectively, during the dry season. However, these relationships were reduced by 36.8% (r=0.48) and 13.6% (r=0.63) against GW and SWC, respectively, in the wet season. On the other hand, the model predicted a moderate positive relationship (r=0.63) between shrub vegetation type and soil water content during the dry season, which was reduced by 31.7% (r=0.43) during the wet season. Our models also predicted that vegetation in the top location (upper part) of the sub-basin is highly responsive to GW and SWC (r=0.78, and 0.70) during the dry season. The results of this study indicate the study region is suitable for seasonal crop production in dry season. Moreover, the results predicted that the growth of vegetation in the top-point location is highly dependent on groundwater flow in both dry and wet seasons, and any instability or long-term drought can negatively affect these floodplain vegetation communities. This study has enriched our knowledge of vegetation responses to groundwater in each season, which will facilitate better floodplain vegetation management.

Keywords: ArcSWAT, machine learning, floodplain vegetation, MODIS NDVI, groundwater

Procedia PDF Downloads 99