Search results for: dimension reducing distribution load flow algorithm
16197 Hysteresis in Sustainable Two-layer Circular Tube under a Lateral Compression Load
Authors: Ami Nomura, Ken Imanishi, Etsuko Ueda, Tadahiro Wada, Shinichi Enoki
Abstract:
Recently, there have been a lot of earthquakes in Japan. It is necessary to promote seismic isolation devices for buildings. The devices have been hardly diffused in attached houses, because the devices are very expensive. We should develop a low-cost seismic isolation device for detached houses. We suggested a new seismic isolation device which uses a two-layer circular tube as a unit. If hysteresis is produced in the two-layer circular tube under lateral compression load, we think that the two-layer circular tube can have energy absorbing capacity. It is necessary to contact the outer layer and the inner layer to produce hysteresis. We have previously reported how the inner layer comes in contact with the outer layer from a perspective of analysis used mechanics of materials. We have clarified that the inner layer comes in contact with the outer layer under a lateral compression load. In this paper, we explored contact area between the outer layer and the inner layer under a lateral compression load by using FEA. We think that changing the inner layer’s thickness is effective in increase the contact area. In order to change the inner layer’s thickness, we changed the shape of the inner layer. As a result, the contact area changes depending on the inner layer’s thickness. Additionally, we experimented to check whether hysteresis occurs in fact. As a consequence, we can reveal hysteresis in the two-layer circular tube under the condition.Keywords: contact area, energy absorbing capacity, hysteresis, seismic isolation device
Procedia PDF Downloads 29516196 LES Investigation of the Natural Vortex Length in a Small-Scale Gas Cyclone
Authors: Dzmitry Misiulia, Sergiy Antonyuk
Abstract:
Small-scale cyclone separators are widely used in aerosol sampling. The flow field in a cyclone sampler is very complex, especially the vortex behavior. Most of the existing models for calculating cyclone efficiency use the same stable vortex structure while the vortex demonstrates dynamic variations rather than the steady-state picture. It can spontaneously ‘end’ at some point within the body of the separator. Natural vortex length is one of the most critical issues when designing and operating gas cyclones and is crucial to proper cyclone performance. The particle transport along the wall to the grid pot is not effective beyond this point. The flow field and vortex behavior inside the aerosol sampler have been investigated for a wide range of Reynolds numbers using Large Eddy Simulations. Two characteristics types of vortex behavior have been found with simulations. At low flow rates the vortex created in the cyclone dissipates in free space (without attaching to a surface) while at higher flow rates it attaches to the cyclone wall. The effects of the Reynolds number on the natural vortex length and the rotation frequency of the end of the vortex have been revealed.Keywords: cyclone, flow field, natural vortex length, pressure drop
Procedia PDF Downloads 15816195 Assessing Functional Structure in European Marine Ecosystems Using a Vector-Autoregressive Spatio-Temporal Model
Authors: Katyana A. Vert-Pre, James T. Thorson, Thomas Trancart, Eric Feunteun
Abstract:
In marine ecosystems, spatial and temporal species structure is an important component of ecosystems’ response to anthropological and environmental factors. Although spatial distribution patterns and fish temporal series of abundance have been studied in the past, little research has been allocated to the joint dynamic spatio-temporal functional patterns in marine ecosystems and their use in multispecies management and conservation. Each species represents a function to the ecosystem, and the distribution of these species might not be random. A heterogeneous functional distribution will lead to a more resilient ecosystem to external factors. Applying a Vector-Autoregressive Spatio-Temporal (VAST) model for count data, we estimate the spatio-temporal distribution, shift in time, and abundance of 140 species of the Eastern English Chanel, Bay of Biscay and Mediterranean Sea. From the model outputs, we determined spatio-temporal clusters, calculating p-values for hierarchical clustering via multiscale bootstrap resampling. Then, we designed a functional map given the defined cluster. We found that the species distribution within the ecosystem was not random. Indeed, species evolved in space and time in clusters. Moreover, these clusters remained similar over time deriving from the fact that species of a same cluster often shifted in sync, keeping the overall structure of the ecosystem similar overtime. Knowing the co-existing species within these clusters could help with predicting data-poor species distribution and abundance. Further analysis is being performed to assess the ecological functions represented in each cluster.Keywords: cluster distribution shift, European marine ecosystems, functional distribution, spatio-temporal model
Procedia PDF Downloads 19416194 Nanofluids and Hybrid Nanofluids: Comparative Study of Mixed Convection in a Round Bottom Flask
Authors: Hicham Salhi
Abstract:
This research project focuses on the numerical investigation of the mixed convection of Hybrid nanofluids in a round bottom flask commonly used in organic chemistry synthesis. The aim of this study is to improve the thermal properties of the reaction medium and enhance the rate of chemical reactions by using hybrid nanofluids. The flat bottom wall of the flask is maintained at a constant high temperature, while the top, left, and right walls are kept at a low temperature. The nanofluids used in this study contain suspended Cu and Al2O3 nanoparticles in pure water. The governing equations are solved numerically using the finite-volume approach and the Boussinesq approximation. The effects of the volume fraction of nanoparticles (φ) ranging from 0% to 5%, the Rayleigh number from 103 to 106, and the type of nanofluid (Cu and Al2O3) on the flow streamlines, isotherm distribution, and Nusselt number are examined in the simulation. The results indicate that the addition of Cu and Al2O3 nanoparticles increases the mean Nusselt number, which improves heat transfer and significantly alters the flow pattern. Moreover, the mean Nusselt number increases with increasing Rayleigh number and volume fraction, with Cu- Al2O3 hybrid nanofluid producing the best results. This research project focuses on the numerical investigation of the mixed convection of Hybrid nanofluids in a round bottom flask commonly used in organic chemistry synthesis. The aim of this study is to improve the thermal properties of the reaction medium and enhance the rate of chemical reactions by using hybrid nanofluids. The flat bottom wall of the flask is maintained at a constant high temperature, while the top, left, and right walls are kept at a low temperature. The nanofluids used in this study contain suspended Cu and Al2O3 nanoparticles in pure water. The governing equations are solved numerically using the finite-volume approach and the Boussinesq approximation. The effects of the volume fraction of nanoparticles (φ) ranging from 0% to 5%, the Rayleigh number from 103 to 106, and the type of nanofluid (Cu and Al2O3) on the flow streamlines, isotherm distribution, and Nusselt number are examined in the simulation. The results indicate that the addition of Cu and Al2O3 nanoparticles increases the mean Nusselt number, which improves heat transfer and significantly alters the flow pattern. Moreover, the mean Nusselt number increases with increasing Rayleigh number and volume fraction, with Cu- Al2O3 hybrid nanofluid producing the best results.Keywords: bottom flask, mixed convection, hybrid nanofluids, numerical simulation
Procedia PDF Downloads 8816193 Mobile and Hot Spot Measurement with Optical Particle Counting Based Dust Monitor EDM264
Authors: V. Ziegler, F. Schneider, M. Pesch
Abstract:
With the EDM264, GRIMM offers a solution for mobile short- and long-term measurements in outdoor areas and at production sites. For research as well as permanent areal observations on a near reference quality base. The model EDM264 features a powerful and robust measuring cell based on optical particle counting (OPC) principle with all the advantages that users of GRIMM's portable aerosol spectrometers are used to. The system is embedded in a compact weather-protection housing with all-weather sampling, heated inlet system, data logger, and meteorological sensor. With TSP, PM10, PM4, PM2.5, PM1, and PMcoarse, the EDM264 provides all fine dust fractions real-time, valid for outdoor applications and calculated with the proven GRIMM enviro-algorithm, as well as six additional dust mass fractions pm10, pm2.5, pm1, inhalable, thoracic and respirable for IAQ and workplace measurements. This highly versatile instrument performs real-time monitoring of particle number, particle size and provides information on particle surface distribution as well as dust mass distribution. GRIMM's EDM264 has 31 equidistant size channels, which are PSL traceable. A high-end data logger enables data acquisition and wireless communication via LTE, WLAN, or wired via Ethernet. Backup copies of the measurement data are stored in the device directly. The rinsing air function, which protects the laser and detector in the optical cell, further increases the reliability and long term stability of the EDM264 under different environmental and climatic conditions. The entire sample volume flow of 1.2 L/min is analyzed by 100% in the optical cell, which assures excellent counting efficiency at low and high concentrations and complies with the ISO 21501-1standard for OPCs. With all these features, the EDM264 is a world-leading dust monitor for precise monitoring of particulate matter and particle number concentration. This highly reliable instrument is an indispensable tool for many users who need to measure aerosol levels and air quality outdoors, on construction sites, or at production facilities.Keywords: aerosol research, aerial observation, fence line monitoring, wild fire detection
Procedia PDF Downloads 15116192 On the Stability Exact Analysis of Tall Buildings with Outrigger System
Authors: Mahrooz Abed, Amir R. Masoodi
Abstract:
Many structural lateral systems are used in tall buildings such as rigid frames, braced frames, shear walls, tubular structures and core structures. Some efficient structures for drift control and base moment reduction in tall buildings is outrigger and belt truss systems. When adopting outrigger beams in building design, their location should be in an optimum position for an economical design. A range of different strategies has been employed to identify the optimum locations of these outrigger beams under wind load. However, there is an absence of scientific research or case studies dealing with optimum outrigger location using buckling analysis. In this paper, one outrigger system is considered at the middle of height of structure. The optimum location of outrigger will be found based on the buckling load limitation. The core of structure is modeled by a clamped tapered beam. The exact stiffness matrix of tapered beam is formulated based on the Euler-Bernoulli theory. Finally, based on the buckling load of structure, the optimal location of outrigger will be found.Keywords: tall buildings, outrigger system, buckling load, second-order effects, Euler-Bernoulli beam theory
Procedia PDF Downloads 39716191 The Grand Unified Theory of Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow Model
Authors: Tory Erickson
Abstract:
The "Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow" (BST-SCWPDF) Model introduces a framework aimed at unifying general relativity (GR) and quantum mechanics (QM). By proposing a concept of bidirectional spacetime, this model suggests that time can flow in more than one direction, thus offering a perspective on temporal dynamics. Integrated with spatial covariance and wave-particle duality in spacetime flow, the BST-SCWPDF Model resolves long-standing discrepancies between GR and QM. This unified theory has profound implications for quantum gravity, potentially offering insights into quantum entanglement, the collapse of the wave function, and the fabric of spacetime itself. The Bidirectional Spacetime with Spatial Covariance and Wave-Particle Duality in Spacetime Flow" (BST-SCWPDF) Model offers researchers a framework for a better understanding of theoretical physics.Keywords: astrophysics, quantum mechanics, general relativity, unification theory, theoretical physics
Procedia PDF Downloads 8816190 Inverse Mapping of Weld Bead Geometry in Shielded Metal Arc-Welding: Genetic Algorithm Approach
Authors: D. S. Nagesh, G. L. Datta
Abstract:
In the field of welding, various studies had been made by some of the previous investigators to predict as well as optimize weld bead geometric descriptors. Modeling of weld bead shape is important for predicting the quality of welds. In most of the cases, design of experiments technique to postulate multiple linear regression equations have been used. Nowadays, Genetic Algorithm (GA) an intelligent information treatment system with the characteristics of treating complex relationships as seen in welding processes used as a tool for inverse mapping/optimization of the process is attempted.Keywords: smaw, genetic algorithm, bead geometry, optimization/inverse mapping
Procedia PDF Downloads 45316189 Genetic Algorithm Approach for Inverse Mapping of Weld Bead Geometry in Shielded Metal Arc-Welding
Authors: D. S. Nagesh, G. L. Datta
Abstract:
In the field of welding, various studies had been made by some of the previous investigators to predict as well as optimize weld bead geometric descriptors. Modeling of weld bead shape is important for predicting the quality of welds. In most of the cases design of experiments technique to postulate multiple linear regression equations have been used. Nowadays Genetic Algorithm (GA) an intelligent information treatment system with the characteristics of treating complex relationships as seen in welding processes used as a tool for inverse mapping/optimization of the process is attempted.Keywords: SMAW, genetic algorithm, bead geometry, optimization/inverse mapping
Procedia PDF Downloads 42116188 Maximum Efficiency of the Photovoltaic Cells Using a Genetic Algorithm
Authors: Latifa Sabri, Mohammed Benzirar, Mimoun Zazoui
Abstract:
The installation of photovoltaic systems is one of future sources to generate electricity without emitting pollutants. The photovoltaic cells used in these systems have demonstrated enormous efficiencies and advantages. Several researches have discussed the maximum efficiency of these technologies, but only a few experiences have succeeded to right weather conditions to get these results. In this paper, two types of cells were selected: crystalline and amorphous silicon. Using the method of genetic algorithm, the results show that for an ambient temperature of 25°C and direct irradiation of 625 W/m², the efficiency of crystalline silicon is 12% and 5% for amorphous silicon.Keywords: PV, maximum efficiency, solar cell, genetic algorithm
Procedia PDF Downloads 42416187 Development of Probability Distribution Models for Degree of Bending (DoB) in Chord Member of Tubular X-Joints under Bending Loads
Authors: Hamid Ahmadi, Amirreza Ghaffari
Abstract:
Fatigue life of tubular joints in offshore structures is not only dependent on the value of hot-spot stress, but is also significantly influenced by the through-the-thickness stress distribution characterized by the degree of bending (DoB). The DoB exhibits considerable scatter calling for greater emphasis in accurate determination of its governing probability distribution which is a key input for the fatigue reliability analysis of a tubular joint. Although the tubular X-joints are commonly found in offshore jacket structures, as far as the authors are aware, no comprehensive research has been carried out on the probability distribution of the DoB in tubular X-joints. What has been used so far as the probability distribution of the DoB in reliability analyses is mainly based on assumptions and limited observations, especially in terms of distribution parameters. In the present paper, results of parametric equations available for the calculation of the DoB have been used to develop probability distribution models for the DoB in the chord member of tubular X-joints subjected to four types of bending loads. Based on a parametric study, a set of samples was prepared and density histograms were generated for these samples using Freedman-Diaconis method. Twelve different probability density functions (PDFs) were fitted to these histograms. The maximum likelihood method was utilized to determine the parameters of fitted distributions. In each case, Kolmogorov-Smirnov test was used to evaluate the goodness of fit. Finally, after substituting the values of estimated parameters for each distribution, a set of fully defined PDFs have been proposed for the DoB in tubular X-joints subjected to bending loads.Keywords: tubular X-joint, degree of bending (DoB), probability density function (PDF), Kolmogorov-Smirnov goodness-of-fit test
Procedia PDF Downloads 71916186 Measurement of Steady Streaming from an Oscillating Bubble Using Particle Image Velocimetry
Authors: Yongseok Kwon, Woowon Jeong, Eunjin Cho, Sangkug Chung, Kyehan Rhee
Abstract:
Steady streaming flow fields induced by a 500 um bubble oscillating at 12 kHz were measured using microscopic particle image velocimetry (PIV). The accuracy of velocity measurement using a micro PIV system was checked by comparing the measured velocity fields with the theoretical velocity profiles in fully developed laminar flow. The steady streaming flow velocities were measured in the saggital plane of the bubble attached on the wall. Measured velocity fields showed upward jet flow with two symmetric counter-rotating vortices, and the maximum streaming velocity was about 12 mm/s, which was within the velocity ranges measured by other researchers. The measured streamlines were compared with the analytic solution, and they also showed a reasonable agreement.Keywords: oscillating bubble, particle image velocimetry, microstreaming, vortices,
Procedia PDF Downloads 41316185 A Comparative Analysis on QRS Peak Detection Using BIOPAC and MATLAB Software
Authors: Chandra Mukherjee
Abstract:
The present paper is a representation of the work done in the field of ECG signal analysis using MATLAB 7.1 Platform. An accurate and simple ECG feature extraction algorithm is presented in this paper and developed algorithm is validated using BIOPAC software. To detect the QRS peak, ECG signal is processed by following mentioned stages- First Derivative, Second Derivative and then squaring of that second derivative. Efficiency of developed algorithm is tested on ECG samples from different database and real time ECG signals acquired using BIOPAC system. Firstly we have lead wise specified threshold value the samples above that value is marked and in the original signal, where these marked samples face change of slope are spotted as R-peak. On the left and right side of the R-peak, faces change of slope identified as Q and S peak, respectively. Now the inbuilt Detection algorithm of BIOPAC software is performed on same output sample and both outputs are compared. ECG baseline modulation correction is done after detecting characteristics points. The efficiency of the algorithm is tested using some validation parameters like Sensitivity, Positive Predictivity and we got satisfied value of these parameters.Keywords: first derivative, variable threshold, slope reversal, baseline modulation correction
Procedia PDF Downloads 41116184 A Partially Accelerated Life Test Planning with Competing Risks and Linear Degradation Path under Tampered Failure Rate Model
Authors: Fariba Azizi, Firoozeh Haghighi, Viliam Makis
Abstract:
In this paper, we propose a method to model the relationship between failure time and degradation for a simple step stress test where underlying degradation path is linear and different causes of failure are possible. It is assumed that the intensity function depends only on the degradation value. No assumptions are made about the distribution of the failure times. A simple step-stress test is used to shorten failure time of products and a tampered failure rate (TFR) model is proposed to describe the effect of the changing stress on the intensities. We assume that some of the products that fail during the test have a cause of failure that is only known to belong to a certain subset of all possible failures. This case is known as masking. In the presence of masking, the maximum likelihood estimates (MLEs) of the model parameters are obtained through an expectation-maximization (EM) algorithm by treating the causes of failure as missing values. The effect of incomplete information on the estimation of parameters is studied through a Monte-Carlo simulation. Finally, a real example is analyzed to illustrate the application of the proposed methods.Keywords: cause of failure, linear degradation path, reliability function, expectation-maximization algorithm, intensity, masked data
Procedia PDF Downloads 33416183 Numerical Investigation of Divergence and Rib Orientation Effects on Thermal Performance in a Divergent Duct, as an Application of Inner Cooling of Turbine Blades
Authors: Heidar Jafarizadeh, Hossein Keshtkar, Ahmad Sohankar
Abstract:
Heat transfer and turbulent flow structure have been studied in a divergent ribbed duct with a varying duct geometry with Reynolds numbers of 7000 to 90000 using numerical methods. In this study, we confirmed our numerical results of a ribbed duct with an Initial slope of zero to 3 degree by comparing them to experimental data we had and investigated the impact of the ducts divergence on heat transfer and flow pattern in the 2-dimensional flow. Then we investigated the effect of tilting the ribs, on heat transfer and flow behavior. We achieved this by changing the ribs angles from a range of 40 to 75 degrees in a divergent duct and simulated the flow in 3-dimensions. Our results show that with an increase in duct divergence, heat transfer increases linearly and the coefficient of friction increases exponentially. As the results show, a duct with a divergence angle of 1.5 degree presents better thermal performance in comparison with all the angle range’s we studied. Besides, a ribbed duct with 40 degree rib orientation had the best thermal performance considering the simultaneous effects of pressure drop and heat transfer which were imposed on it.Keywords: divergent ribbed duct, heat transfer, thermal performance, turbulent flow structure
Procedia PDF Downloads 30216182 Reliability Analysis of Computer Centre at Yobe State University Using LRU Algorithm
Authors: V. V. Singh, Yusuf Ibrahim Gwanda, Rajesh Prasad
Abstract:
In this paper, we focus on the reliability and performance analysis of Computer Centre (CC) at Yobe State University, Damaturu, Nigeria. The CC consists of three servers: one database mail server, one redundant and one for sharing with the client computers in the CC (called as a local server). Observing the different possibilities of the functioning of the CC, the analysis has been done to evaluate the various popular measures of reliability such as availability, reliability, mean time to failure (MTTF), profit analysis due to the operation of the system. The system can ultimately fail due to the failure of router, redundant server before repairing the mail server and switch failure. The system can also partially fail when a local server fails. The failed devices have restored according to Least Recently Used (LRU) techniques. The system can also fail entirely due to a cooling failure of the server, electricity failure or some natural calamity like earthquake, fire tsunami, etc. All the failure rates are assumed to be constant and follow exponential time distribution, while the repair follows two types of distributions: i.e. general and Gumbel-Hougaard family copula distribution.Keywords: reliability, availability Gumbel-Hougaard family copula, MTTF, internet data centre
Procedia PDF Downloads 53116181 Graph-Based Semantical Extractive Text Analysis
Authors: Mina Samizadeh
Abstract:
In the past few decades, there has been an explosion in the amount of available data produced from various sources with different topics. The availability of this enormous data necessitates us to adopt effective computational tools to explore the data. This leads to an intense growing interest in the research community to develop computational methods focused on processing this text data. A line of study focused on condensing the text so that we are able to get a higher level of understanding in a shorter time. The two important tasks to do this are keyword extraction and text summarization. In keyword extraction, we are interested in finding the key important words from a text. This makes us familiar with the general topic of a text. In text summarization, we are interested in producing a short-length text which includes important information about the document. The TextRank algorithm, an unsupervised learning method that is an extension of the PageRank (algorithm which is the base algorithm of Google search engine for searching pages and ranking them), has shown its efficacy in large-scale text mining, especially for text summarization and keyword extraction. This algorithm can automatically extract the important parts of a text (keywords or sentences) and declare them as a result. However, this algorithm neglects the semantic similarity between the different parts. In this work, we improved the results of the TextRank algorithm by incorporating the semantic similarity between parts of the text. Aside from keyword extraction and text summarization, we develop a topic clustering algorithm based on our framework, which can be used individually or as a part of generating the summary to overcome coverage problems.Keywords: keyword extraction, n-gram extraction, text summarization, topic clustering, semantic analysis
Procedia PDF Downloads 7116180 Retrofitting of Bridge Piers against the Scour Damages: Case Study of the Marand-Soofian Route Bridge
Authors: Shatirah Akib, Hossein Basser, Hojat Karami, Afshin Jahangirzadeh
Abstract:
Bridge piers which are constructed in the track of high water rivers cause some variations in the flow patterns. This variation mostly is a result of the changes in river sections. Decreasing the river section, bridge piers significantly impress the flow patterns. Once the flow approaches the piers, the stream lines change their order, causing the appearance of different flow patterns around the bridge piers. New flow patterns are created following the geometry and the other technical characteristics of the piers. One of the most significant consequences of this event is the scour generated around the bridge piers which threatens the safety of the structure. In order to determine the properties of scour holes, to find maximum depth of the scour is an important factor. In this manuscript a numerical simulation of the scour around Marand-Soofian route bridge piers has been carried out via SSIIM 2.0 Software and the amount of maximum scour has been achieved subsequently. Eventually the methods for retrofitting of bridge piers against scours and also the methods for decreasing the amount of scour have been offered.Keywords: scour, bridge pier, numerical simulation, SSIIM 2.0
Procedia PDF Downloads 47316179 Knowledge-Driven Decision Support System Based on Knowledge Warehouse and Data Mining by Improving Apriori Algorithm with Fuzzy Logic
Authors: Pejman Hosseinioun, Hasan Shakeri, Ghasem Ghorbanirostam
Abstract:
In recent years, we have seen an increasing importance of research and study on knowledge source, decision support systems, data mining and procedure of knowledge discovery in data bases and it is considered that each of these aspects affects the others. In this article, we have merged information source and knowledge source to suggest a knowledge based system within limits of management based on storing and restoring of knowledge to manage information and improve decision making and resources. In this article, we have used method of data mining and Apriori algorithm in procedure of knowledge discovery one of the problems of Apriori algorithm is that, a user should specify the minimum threshold for supporting the regularity. Imagine that a user wants to apply Apriori algorithm for a database with millions of transactions. Definitely, the user does not have necessary knowledge of all existing transactions in that database, and therefore cannot specify a suitable threshold. Our purpose in this article is to improve Apriori algorithm. To achieve our goal, we tried using fuzzy logic to put data in different clusters before applying the Apriori algorithm for existing data in the database and we also try to suggest the most suitable threshold to the user automatically.Keywords: decision support system, data mining, knowledge discovery, data discovery, fuzzy logic
Procedia PDF Downloads 33516178 A Novel Heuristic for Analysis of Large Datasets by Selecting Wrapper-Based Features
Authors: Bushra Zafar, Usman Qamar
Abstract:
Large data sample size and dimensions render the effectiveness of conventional data mining methodologies. A data mining technique are important tools for collection of knowledgeable information from variety of databases and provides supervised learning in the form of classification to design models to describe vital data classes while structure of the classifier is based on class attribute. Classification efficiency and accuracy are often influenced to great extent by noisy and undesirable features in real application data sets. The inherent natures of data set greatly masks its quality analysis and leave us with quite few practical approaches to use. To our knowledge first time, we present a new approach for investigation of structure and quality of datasets by providing a targeted analysis of localization of noisy and irrelevant features of data sets. Machine learning is based primarily on feature selection as pre-processing step which offers us to select few features from number of features as a subset by reducing the space according to certain evaluation criterion. The primary objective of this study is to trim down the scope of the given data sample by searching a small set of important features which may results into good classification performance. For this purpose, a heuristic for wrapper-based feature selection using genetic algorithm and for discriminative feature selection an external classifier are used. Selection of feature based on its number of occurrence in the chosen chromosomes. Sample dataset has been used to demonstrate proposed idea effectively. A proposed method has improved average accuracy of different datasets is about 95%. Experimental results illustrate that proposed algorithm increases the accuracy of prediction of different diseases.Keywords: data mining, generic algorithm, KNN algorithms, wrapper based feature selection
Procedia PDF Downloads 31616177 Theoretical Framework and Empirical Simulation of Policy Design on Trans-Dimensional Resource Recycling
Authors: Yufeng Wu, Yifan Gu, Bin Li, Wei Wang
Abstract:
Resource recycling process contains a subsystem with interactions of three dimensions including coupling allocation of primary and secondary resources, responsibility coordination of stakeholders in forward and reverse supply chains, and trans-boundary transfer of hidden resource and environmental responsibilities between regions. Overlap or lack of responsibilities is easy to appear at the intersection of the three management dimensions. It is urgent to make an overall design of the policy system for recycling resources. From theoretical perspective, this paper analyzes the unique external differences of resource and environment in various dimensions and explores the reason why the effects of trans-dimensional policies are strongly correlated. Taking the example of the copper resources contained in the waste electrical and electronic equipment, this paper constructs reduction effect accounting model of resources recycling and set four trans-dimensional policy scenarios including resources tax and environmental tax reform of the raw and secondary resources, application of extended producer responsibility system, promotion of clean development mechanism, and strict entry barriers of imported wastes. In these ways, the paper simulates the impact effect of resources recycling process on resource deduction and emission reduction of waste water and gas, and constructs trans-dimensional policy mix scenario through integrating dominant strategy. The results show that combined application of various dimensional policies can achieve incentive compatibility and the trans-dimensional policy mix scenario can reach a better effect. Compared with baseline scenario, this scenario will increase 91.06% copper resources reduction effect and improve emission reduction of waste water and gas by eight times from 2010 to 2030. This paper further analyzes the development orientation of policies in various dimension. In resource dimension, the combined application of compulsory, market and authentication methods should be promoted to improve the use ratio of secondary resources. In supply chain dimension, resource value, residual functional value and potential information value contained in waste products should be fully excavated to construct a circular business system. In regional dimension, it should give full play to the comparative advantages of manufacturing power to improve China’s voice in resource recycling in the world.Keywords: resource recycling, trans-dimension, policy design, incentive compatibility, life cycle
Procedia PDF Downloads 12616176 Fermentable Bio-Ethanol Using Bakers and Palmwine Yeasts: Indices of Bioavailability of Carbohydrate and Sugar from Fungal Treated Rice Husk
Authors: Ezeonu, Chukwuma Stephen, Onwurah, Ikechukwu Noel Emmanuel
Abstract:
Pure strains of Aspergillus fumigatus (AF), aspergillus niger (AN), aspergillus oryzae (AO), trichophyton mentagrophyte (TM), trichophyton rubrum (TR) and Trichophyton soudanense (TS) were isolated from decomposing rice husk. Freshly processed rice husk in Mandle’s medium were heat pre-treated using an autoclave at 121oC for 20 minutes. The isolated fungi as monoculture and di-culture combinations were inoculated into each of the pre-treated rice husk with the exception of two controls. Seven days hydrolysis was followed by estimation of carbohydrate, reducing sugar and non-reducing sugar. Fungal treated rice husks were left to ferment for 7 days with introduction of both baker’s and palm wine yeast. The result obtained in the work gave the highest carbohydrate (20.53 ± 2.73 %) from rice husks treated with TS + TR di-culture. The highest soluble reducing sugar (2.66 ± 0.14 %) was obtained from rice husk treated with TM. The highest soluble nonreducing sugar (18.08 ± 2.61 %) was from AF. The introduction of yeasts from palm wine gave the highest bio-ethanol (12.82 ± 0.39 %) from AO. The highest bio-ethanol (6.60 ± 0.10 %) from baker's yeast fermentation was in AO + TS treated rice husk. There was increased availability of sugar and moderate yield of bio-ethanol, especially from palm wine yeast.Keywords: fungi, rice husk, carbohydrate, reducing sugar, non-reducing sugar, ethanol, fermentation
Procedia PDF Downloads 44016175 Blind Watermarking Using Discrete Wavelet Transform Algorithm with Patchwork
Authors: Toni Maristela C. Estabillo, Michaela V. Matienzo, Mikaela L. Sabangan, Rosette M. Tienzo, Justine L. Bahinting
Abstract:
This study is about blind watermarking on images with different categories and properties using two algorithms namely, Discrete Wavelet Transform and Patchwork Algorithm. A program is created to perform watermark embedding, extraction and evaluation. The evaluation is based on three watermarking criteria namely: image quality degradation, perceptual transparency and security. Image quality is measured by comparing the original properties with the processed one. Perceptual transparency is measured by a visual inspection on a survey. Security is measured by implementing geometrical and non-geometrical attacks through a pass or fail testing. Values used to measure the following criteria are mostly based on Mean Squared Error (MSE) and Peak Signal to Noise Ratio (PSNR). The results are based on statistical methods used to interpret and collect data such as averaging, z Test and survey. The study concluded that the combined DWT and Patchwork algorithms were less efficient and less capable of watermarking than DWT algorithm only.Keywords: blind watermarking, discrete wavelet transform algorithm, patchwork algorithm, digital watermark
Procedia PDF Downloads 26816174 Experimental Challenges and Solutions in Design and Operation of the Test Rig for Water Lubricated Journal Bearing
Authors: Ravindra Mallya, B. Satish Shenoy, B. Raghuvir Pai
Abstract:
The study deals with the challenges in developing a test rig to test the performance of water lubricated journal bearing. The test rig is designed to simulate the working conditions of the bearing in order to understand their performance before they are put in operation. The bearing that is studied is the commercially available water lubricated bearing which has a rubber liner bonded with a rigid metal shell. The lubricant enters the bearing axially through a pressurized inlet tank and exits to an outlet tank which is at sufficiently low pressure. The load on the bearing is applied through the dead weight system which acts both in upward and downward direction so that net load acts on the bearing. The issues in feeding the lubricant into the bearing from the inlet side and preventing the leakage of the lubricant is discussed. The application of the load on the test bearing while maintaining the bearing afloat is also discussed.Keywords: axial groove, hydrodynamic pressure, journal bearing, test rig, water lubrication
Procedia PDF Downloads 50216173 A Monolithic Arbitrary Lagrangian-Eulerian Finite Element Strategy for Partly Submerged Solid in Incompressible Fluid with Mortar Method for Modeling the Contact Surface
Authors: Suman Dutta, Manish Agrawal, C. S. Jog
Abstract:
Accurate computation of hydrodynamic forces on floating structures and their deformation finds application in the ocean and naval engineering and wave energy harvesting. This manuscript presents a monolithic, finite element strategy for fluid-structure interaction involving hyper-elastic solids partly submerged in an incompressible fluid. A velocity-based Arbitrary Lagrangian-Eulerian (ALE) formulation has been used for the fluid and a displacement-based Lagrangian approach has been used for the solid. The flexibility of the ALE technique permits us to treat the free surface of the fluid as a Lagrangian entity. At the interface, the continuity of displacement, velocity and traction are enforced using the mortar method. In the mortar method, the constraints are enforced in a weak sense using the Lagrange multiplier method. In the literature, the mortar method has been shown to be robust in solving various contact mechanics problems. The time-stepping strategy used in this work reduces to the generalized trapezoidal rule in the Eulerian setting. In the Lagrangian limit, in the absence of external load, the algorithm conserves the linear and angular momentum and the total energy of the system. The use of monolithic coupling with an energy-conserving time-stepping strategy gives an unconditionally stable algorithm and allows the user to take large time steps. All the governing equations and boundary conditions have been mapped to the reference configuration. The use of the exact tangent stiffness matrix ensures that the algorithm converges quadratically within each time step. The robustness and good performance of the proposed method are demonstrated by solving benchmark problems from the literature.Keywords: ALE, floating body, fluid-structure interaction, monolithic, mortar method
Procedia PDF Downloads 27416172 Exploring Data Stewardship in Fog Networking Using Blockchain Algorithm
Authors: Ruvaitha Banu, Amaladhithyan Krishnamoorthy
Abstract:
IoT networks today solve various consumer problems, from home automation systems to aiding in driving autonomous vehicles with the exploration of multiple devices. For example, in an autonomous vehicle environment, multiple sensors are available on roads to monitor weather and road conditions and interact with each other to aid the vehicle in reaching its destination safely and timely. IoT systems are predominantly dependent on the cloud environment for data storage, and computing needs that result in latency problems. With the advent of Fog networks, some of this storage and computing is pushed to the edge/fog nodes, saving the network bandwidth and reducing the latency proportionally. Managing the data stored in these fog nodes becomes crucial as it might also store sensitive information required for a certain application. Data management in fog nodes is strenuous because Fog networks are dynamic in terms of their availability and hardware capability. It becomes more challenging when the nodes in the network also live a short span, detaching and joining frequently. When an end-user or Fog Node wants to access, read, or write data stored in another Fog Node, then a new protocol becomes necessary to access/manage the data stored in the fog devices as a conventional static way of managing the data doesn’t work in Fog Networks. The proposed solution discusses a protocol that acts by defining sensitivity levels for the data being written and read. Additionally, a distinct data distribution and replication model among the Fog nodes is established to decentralize the access mechanism. In this paper, the proposed model implements stewardship towards the data stored in the Fog node using the application of Reinforcement Learning so that access to the data is determined dynamically based on the requests.Keywords: IoT, fog networks, data stewardship, dynamic access policy
Procedia PDF Downloads 5916171 Optimum Dimensions of Hydraulic Structures Foundation and Protections Using Coupled Genetic Algorithm with Artificial Neural Network Model
Authors: Dheyaa W. Abbood, Rafa H. AL-Suhaili, May S. Saleh
Abstract:
A model using the artificial neural networks and genetic algorithm technique is developed for obtaining optimum dimensions of the foundation length and protections of small hydraulic structures. The procedure involves optimizing an objective function comprising a weighted summation of the state variables. The decision variables considered in the optimization are the upstream and downstream cutoffs length sand their angles of inclination, the foundation length, and the length of the downstream soil protection. These were obtained for a given maximum difference in head, depth of impervious layer and degree of anisotropy.The optimization carried out subjected to constraints that ensure a safe structure against the uplift pressure force and sufficient protection length at the downstream side of the structure to overcome an excessive exit gradient. The Geo-studios oft ware, was used to analyze 1200 different cases. For each case the length of protection and volume of structure required to satisfy the safety factors mentioned previously were estimated. An ANN model was developed and verified using these cases input-output sets as its data base. A MatLAB code was written to perform a genetic algorithm optimization modeling coupled with this ANN model using a formulated optimization model. A sensitivity analysis was done for selecting the cross-over probability, the mutation probability and level ,the number of population, the position of the crossover and the weights distribution for all the terms of the objective function. Results indicate that the most factor that affects the optimum solution is the number of population required. The minimum value that gives stable global optimum solution of this parameters is (30000) while other variables have little effect on the optimum solution.Keywords: inclined cutoff, optimization, genetic algorithm, artificial neural networks, geo-studio, uplift pressure, exit gradient, factor of safety
Procedia PDF Downloads 32416170 A Simple Design Procedure for Calculating the Column Ultimate Load of Steel Frame Structures
Authors: Abdul Hakim Chikho
Abstract:
Calculating the ultimate load of a column in a sway framed structure involves, in the currently used design method, the calculation of the column effective length and utilizing the interaction formulas or tables. Therefore, no allowance is usually made for the effects of the presence of semi rigid connections or the presence of infill panels. In this paper, a new and simple design procedure is recommend to calculate the ultimate load of a framed Column allowing for the presence of rotational end restraints, semi rigid connections, the column end moments resulted from the applied vertical and horizontal loading and infill panels in real steel structure. In order to verify the accuracy of the recommended method to predict good and safe estimations of framed column ultimate loads, several examples have been solved utilizing the recommended procedure, and the results were compared to those obtained using a second order computer program, and good correlation had been obtained. Therefore, the accuracy of the proposed method to predict the Behaviour of practical steel columns in framed structures has been verified.Keywords: column ultimate load, semi rigid connections, steel column, infill panel, steel structure
Procedia PDF Downloads 17816169 Service Flow in Multilayer Networks: A Method for Evaluating the Layout of Urban Medical Resources
Authors: Guanglin Song
Abstract:
(Objective) Situated within the context of China's tiered medical treatment system, this study aims to analyze spatial causes of urban healthcare access difficulties from the perspective of the configuration of healthcare facilities. (Methods) A social network analysis approach is employed to construct a healthcare demand and supply flow network between major residential clusters and various tiers of hospitals in the city.(Conclusion) The findings reveal that:1.there exists overall maldistribution and over-concentration of healthcare resources in Study Area, characterized by structural imbalance; 2.the low rate of primary care utilization in Study Area is a key factor contributing to congestion at higher-tier hospitals, as excessive reliance on these institutions by neighboring communities exacerbates the problem; 3.gradual optimization of the healthcare facility layout in Study Area, encompassing holistic, local, and individual institutional levels, can enhance systemic efficiency and resource balance.(Prospects) This research proposes a method for evaluating urban healthcare resource distribution structures based on service flows within hierarchical networks. It offers spatially targeted optimization suggestions for promoting the implementation of the tiered healthcare system and alleviating challenges related to accessibility and congestion in seeking medical care. Provide some new ideas for researchers and healthcare managers in countries, cities, and healthcare management around the world with similar challenges.Keywords: flow of public services, urban networks, healthcare facilities, spatial planning, urban networks
Procedia PDF Downloads 6916168 A Preliminary Study of Economic Dimension of Underground Rock Caverns for Water Storage at Singapore
Authors: Junlong Shang, Zhengxian Chua, Hoongping Peh, Zhiye Zhao
Abstract:
Due to scarce land resources in Singapore, it is imperative to increase water storage capacities to meet the increasing demand of water to secure a sustainable development, which can be achieved in the underground by rock caverns. In this paper, a preliminary study on the effects of cavern span, height and radius on the cavern stability is presented to provide a guidance on the cavern construction in the context of Singapore. It is found that the radius of caverns should be around half of the span width (i.e., B/R=2) to reduce vertical displacement at the crown of cavern. The smaller the rock cover, the smaller displacement. The minimum rock thickness should be at least the same as the cavern span to eliminate excessive yielded element. Finally, rock support system is introduced to maintain the profile of caverns.Keywords: cavern dimension, numerical modelling, sustainable development, underground rock cavern
Procedia PDF Downloads 320