Search results for: e-content producing algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4855

Search results for: e-content producing algorithm

3925 Design of Chaos Algorithm Based Optimal PID Controller for SVC

Authors: Saeid Jalilzadeh

Abstract:

SVC is one of the most significant devices in FACTS technology which is used in parallel compensation, enhancing the transient stability, limiting the low frequency oscillations and etc. designing a proper controller is effective in operation of svc. In this paper the equations that describe the proposed system have been linearized and then the optimum PID controller has been designed for svc which its optimal coefficients have been earned by chaos algorithm. Quick damping of oscillations of generator is the aim of designing of optimum PID controller for svc whether the input power of generator has been changed suddenly. The system with proposed controller has been simulated for a special disturbance and the dynamic responses of generator have been presented. The simulation results showed that a system composed with proposed controller has suitable operation in fast damping of oscillations of generator.

Keywords: chaos, PID controller, SVC, frequency oscillation

Procedia PDF Downloads 429
3924 Geopolymer Concrete: A Review of Properties, Applications and Limitations

Authors: Abbas Ahmed Albu Shaqraa

Abstract:

The concept of a safe environment and low greenhouse gas emissions is a common concern especially in the construction industry. The produced carbon dioxide (CO2) emissions are nearly a ton in producing only one ton of Portland cement, which is the primary ingredient of concrete. Current studies had investigated the utilization of several waste materials in producing a cement free concrete. The geopolymer concrete is a green material that results from the reaction of aluminosilicate material with an alkaline liquid. A summary of several recent researches in geopolymer concrete will be presented in this manuscript. In addition, the offered presented review considers the use of several waste materials including fly ash, granulated blast furnace slag, cement kiln dust, kaolin, metakaolin, and limestone powder as binding materials in making geopolymer concrete. Moreover, the mechanical, chemical and thermal properties of geopolymer concrete will be reviewed. In addition, the geopolymer concrete applications and limitations will be discussed as well. The results showed a high early compressive strength gain in geopolymer concrete when dry- heating or steam curing was performed. Also, it was stated that the outstanding acidic resistance of the geopolymer concrete made it possible to be used where the ordinary Portland cement concrete was doubtable. Thus, the commercial geopolymer concrete pipes were favored for sewer system in case of high acidic conditions. Furthermore, it was reported that the geopolymer concrete could stand up to 1200 °C in fire without losing its strength integrity whereas the Portland cement concrete was losing its function upon heating to some 100s °C only. However, the geopolymer concrete still considered as an emerging field and occupied mainly by the precast industries.

Keywords: geopolymer concrete, Portland cement concrete, alkaline liquid, compressive strength

Procedia PDF Downloads 202
3923 Optimization of Structures Subjected to Earthquake

Authors: Alireza Lavaei, Alireza Lohrasbi, Mohammadali M. Shahlaei

Abstract:

To reduce the overall time of structural optimization for earthquake loads two strategies are adopted. In the first strategy, a neural system consisting self-organizing map and radial basis function neural networks, is utilized to predict the time history responses. In this case, the input space is classified by employing a self-organizing map neural network. Then a distinct RBF neural network is trained in each class. In the second strategy, an improved genetic algorithm is employed to find the optimum design. A 72-bar space truss is designed for optimal weight using exact and approximate analysis for the El Centro (S-E 1940) earthquake loading. The numerical results demonstrate the computational advantages and effectiveness of the proposed method.

Keywords: optimization, genetic algorithm, neural networks, self-organizing map

Procedia PDF Downloads 285
3922 Developing a Recommendation Library System based on Android Application

Authors: Kunyanuth Kularbphettong, Kunnika Tenprakhon, Pattarapan Roonrakwit

Abstract:

In this paper, we present a recommendation library application on Android system. The objective of this system is to support and advice user to use library resources based on mobile application. We describe the design approaches and functional components of this system. The system was developed based on under association rules, Apriori algorithm. In this project, it was divided the result by the research purposes into 2 parts: developing the Mobile application for online library service and testing and evaluating the system. Questionnaires were used to measure user satisfaction with system usability by specialists and users. The results were satisfactory both specialists and users.

Keywords: online library, Apriori algorithm, Android application, black box

Procedia PDF Downloads 467
3921 Non-Targeted Adversarial Image Classification Attack-Region Modification Methods

Authors: Bandar Alahmadi, Lethia Jackson

Abstract:

Machine Learning model is used today in many real-life applications. The safety and security of such model is important, so the results of the model are as accurate as possible. One challenge of machine learning model security is the adversarial examples attack. Adversarial examples are designed by the attacker to cause the machine learning model to misclassify the input. We propose a method to generate adversarial examples to attack image classifiers. We are modifying the successfully classified images, so a classifier misclassifies them after the modification. In our method, we do not update the whole image, but instead we detect the important region, modify it, place it back to the original image, and then run it through a classifier. The algorithm modifies the detected region using two methods. First, it will add abstract image matrix on back of the detected image matrix. Then, it will perform a rotation attack to rotate the detected region around its axes, and embed the trace of image in image background. Finally, the attacked region is placed in its original position, from where it was removed, and a smoothing filter is applied to smooth the background with foreground. We test our method in cascade classifier, and the algorithm is efficient, the classifier confident has dropped to almost zero. We also try it in CNN (Convolutional neural network) with higher setting and the algorithm was successfully worked.

Keywords: adversarial examples, attack, computer vision, image processing

Procedia PDF Downloads 321
3920 Image Enhancement Algorithm of Photoacoustic Tomography Using Active Contour Filtering

Authors: Prasannakumar Palaniappan, Dong Ho Shin, Chul Gyu Song

Abstract:

The photoacoustic images are obtained from a custom developed linear array photoacoustic tomography system. The biological specimens are imitated by conducting phantom tests in order to retrieve a fully functional photoacoustic image. The acquired image undergoes the active region based contour filtering to remove the noise and accurately segment the object area for further processing. The universal back projection method is used as the image reconstruction algorithm. The active contour filtering is analyzed by evaluating the signal to noise ratio and comparing it with the other filtering methods.

Keywords: contour filtering, linear array, photoacoustic tomography, universal back projection

Procedia PDF Downloads 388
3919 Recognition of Early Enterococcus Faecalis through Image Treatment by Using Octave

Authors: Laura Victoria Vigoya Morales, David Rolando Suarez Mora

Abstract:

The problem of detecting enterococcus faecalis is receiving considerable attention with the new cases of beachgoers infected with the bacteria, which can be found in fecal matter. The process detection of this kind of bacteria would be taking a long time, which waste time and money as a result of closing recreation place, like beach or pools. Hence, new methods for automating the process of detecting and recognition of this bacteria has become in a challenge. This article describes a novel approach to detect the enterococcus faecalis bacteria in water by using an octave algorithm, which embody a network neural. This document shows result of performance, quality and integrity of the algorithm.

Keywords: Enterococcus faecalis, image treatment, octave and network neuronal

Procedia PDF Downloads 209
3918 Application of the Discrete-Event Simulation When Optimizing of Business Processes in Trading Companies

Authors: Maxat Bokambayev, Bella Tussupova, Aisha Mamyrova, Erlan Izbasarov

Abstract:

Optimization of business processes in trading companies is reviewed in the report. There is the presentation of the “Wholesale Customer Order Handling Process” business process model applicable for small and medium businesses. It is proposed to apply the algorithm for automation of the customer order processing which will significantly reduce labor costs and time expenditures and increase the profitability of companies. An optimized business process is an element of the information system of accounting of spare parts trading network activity. The considered algorithm may find application in the trading industry as well.

Keywords: business processes, discrete-event simulation, management, trading industry

Procedia PDF Downloads 329
3917 Transmission Line Congestion Management Using Hybrid Fish-Bee Algorithm with Unified Power Flow Controller

Authors: P. Valsalal, S. Thangalakshmi

Abstract:

There is a widespread changeover in the electrical power industry universally from old-style monopolistic outline towards a horizontally distributed competitive structure to come across the demand of rising consumption. When the transmission lines of derestricted system are incapable to oblige the entire service needs, the lines are overloaded or congested. The governor between customer and power producer is nominated as Independent System Operator (ISO) to lessen the congestion without obstructing transmission line restrictions. Among the existing approaches for congestion management, the frequently used approaches are reorganizing the generation and load curbing. There is a boundary for reorganizing the generators, and further loads may not be supplemented with the prevailing resources unless more private power producers are added in the system by considerably raising the cost. Hence, congestion is relaxed by appropriate Flexible AC Transmission Systems (FACTS) devices which boost the existing transfer capacity of transmission lines. The FACTs device, namely, Unified Power Flow Controller (UPFC) is preferred, and the correct placement of UPFC is more vital and should be positioned in the highly congested line. Hence, the weak line is identified by using power flow performance index with the new objective function with proposed hybrid Fish – Bee algorithm. Further, the location of UPFC at appropriate line reduces the branch loading and minimizes the voltage deviation. The power transfer capacity of lines is determined with and without UPFC in the identified congested line of IEEE 30 bus structure and the simulated results are compared with prevailing algorithms. It is observed that the transfer capacity of existing line is increased with the presented algorithm and thus alleviating the congestion.

Keywords: available line transfer capability, congestion management, FACTS device, Hybrid Fish-Bee Algorithm, ISO, UPFC

Procedia PDF Downloads 365
3916 Life Cycle Assessment Comparison between Methanol and Ethanol Feedstock for the Biodiesel from Soybean Oil

Authors: Pawit Tangviroon, Apichit Svang-Ariyaskul

Abstract:

As the limited availability of petroleum-based fuel has been a major concern, biodiesel is one of the most attractive alternative fuels because it is renewable and it also has advantages over the conventional petroleum-base diesel. At Present, productions of biodiesel generally perform by transesterification of vegetable oils with low molecular weight alcohol, mainly methanol, using chemical catalysts. Methanol is petrochemical product that makes biodiesel producing from methanol to be not pure renewable energy source. Therefore, ethanol as a product produced by fermentation processes. It appears as a potential feed stock that makes biodiesel to be pure renewable alternative fuel. The research is conducted based on two biodiesel production processes by reacting soybean oils with methanol and ethanol. Life cycle assessment was carried out in order to evaluate the environmental impacts and to identify the process alternative. Nine mid-point impact categories are investigated. The results indicate that better performance on Abiotic Depletion Potential (ADP) and Acidification Potential (AP) are observed in biodiesel production from methanol when compared with biodiesel production from ethanol due to less energy consumption during the production processes. Except for ADP and AP, using methanol as feed stock does not show any advantages over biodiesel from ethanol. The single score method is also included in this study in order to identify the best option between two processes of biodiesel production. The global normalization and weighting factor based on eco-taxes are used and it shows that producing biodiesel form ethanol has less environmental load compare to biodiesel from methanol.

Keywords: biodiesel, ethanol, life cycle assessment, methanol, soybean oil

Procedia PDF Downloads 199
3915 Design and Performance Evaluation of Plasma Spouted Bed Reactor for Converting Waste Plastic into Green Hydrogen

Authors: Palash Kumar Mollick, Leire Olazar, Laura Santamaria, Pablo Comendador, Gartzen Lopez, Martin Olazar

Abstract:

Average calorific value of a mixure of waste plastic is approximately 38 MJ/kg. Present work aims to extract maximum possible energy from a mixure of waste plastic using a DC thermal plasma in a spouted bed reactor. Plasma pyrolysis and steam reforming process has shown a potential to generate hydrogen from plastic with much below of legal limit of producing dioxins and furans as the carcinogenic gases. A spouted bed pyrolysis rector can continuously process plastic beads to produce organic volatiles, which later react with steam in presence of catalyst to results in syngas. lasma being the fourth state of matter, can carry high impact electrons to favour the activation energy of any chemical reactions. Computational Fluid Dynamic (CFD) simulation using COMSOL Multiphysics software has been performed to evaluate performance of a plasma spouted bed reactor in producing contamination free hydrogen as a green energy from waste plastic beads. The simulation results will showcase a design of a plasma spouted bed reactor for converting plastic waste into green hydrogen in a single step process. The high temperature hydrodynamics of spouted bed with plastic beads and the corresponding temperature distribution inside the reaction chamber will be critically examined for it’s near future installation of demonstration plant.

Keywords: green hydrogen, plastic waste, synthetic gas, pyrolysis, steam reforming, spouted bed, reactor design, plasma, dc palsma, cfd simulation

Procedia PDF Downloads 84
3914 Aerodynamic Design an UAV and Stability Analysis with Method of Genetic Algorithm Optimization

Authors: Saul A. Torres Z., Eduardo Liceaga C., Alfredo Arias M.

Abstract:

We seek to develop a UAV for agricultural spraying at a maximum altitude of 5000 meters above sea level, with a payload of 100 liters of fumigant. For the developing the aerodynamic design of the aircraft is using computational tools such as the "Vortex Lattice Athena" software, "MATLAB", "ANSYS FLUENT", "XFoil" package among others. Also methods are being used structured programming, exhaustive analysis of optimization methods and search. The results have a very low margin of error, and the multi-objective problems can be helpful for future developments. Also we developed method for Stability Analysis (Lateral-Directional and Longitudinal).

Keywords: aerodynamics design, optimization, algorithm genetic, multi-objective problem, longitudinal stability, lateral-directional stability

Procedia PDF Downloads 574
3913 Dynamic Synthesis of a Flexible Multibody System

Authors: Mohamed Amine Ben Abdallah, Imed Khemili, Nizar Aifaoui

Abstract:

This work denotes an insight into dynamic synthesis of multibody systems. A set of mechanism parameters design variable are synthetized based on a desired mechanism response, such as, velocity, acceleration and bodies deformations. Moreover, knowing the work space, for a robot, and mechanism response allow defining optimal parameters mechanism handling with the desired target response. To this end, evolutionary genetic algorithm has been deployed. A demonstrative example for imperfect mechanism has been treated, mainly, a slider crank mechanism with a flexible connecting rod. The transversal deflection of the connecting rod has been chosen as response to identify the mechanism design parameters.

Keywords: dynamic response, evolutionary genetic algorithm, flexible bodies, optimization

Procedia PDF Downloads 298
3912 Signs, Signals and Syndromes: Algorithmic Surveillance and Global Health Security in the 21st Century

Authors: Stephen L. Roberts

Abstract:

This article offers a critical analysis of the rise of syndromic surveillance systems for the advanced detection of pandemic threats within contemporary global health security frameworks. The article traces the iterative evolution and ascendancy of three such novel syndromic surveillance systems for the strengthening of health security initiatives over the past two decades: 1) The Program for Monitoring Emerging Diseases (ProMED-mail); 2) The Global Public Health Intelligence Network (GPHIN); and 3) HealthMap. This article demonstrates how each newly introduced syndromic surveillance system has become increasingly oriented towards the integration of digital algorithms into core surveillance capacities to continually harness and forecast upon infinitely generating sets of digital, open-source data, potentially indicative of forthcoming pandemic threats. This article argues that the increased centrality of the algorithm within these next-generation syndromic surveillance systems produces a new and distinct form of infectious disease surveillance for the governing of emergent pathogenic contingencies. Conceptually, the article also shows how the rise of this algorithmic mode of infectious disease surveillance produces divergences in the governmental rationalities of global health security, leading to the rise of an algorithmic governmentality within contemporary contexts of Big Data and these surveillance systems. Empirically, this article demonstrates how this new form of algorithmic infectious disease surveillance has been rapidly integrated into diplomatic, legal, and political frameworks to strengthen the practice of global health security – producing subtle, yet distinct shifts in the outbreak notification and reporting transparency of states, increasingly scrutinized by the algorithmic gaze of syndromic surveillance.

Keywords: algorithms, global health, pandemic, surveillance

Procedia PDF Downloads 166
3911 Frequent Item Set Mining for Big Data Using MapReduce Framework

Authors: Tamanna Jethava, Rahul Joshi

Abstract:

Frequent Item sets play an essential role in many data Mining tasks that try to find interesting patterns from the database. Typically it refers to a set of items that frequently appear together in transaction dataset. There are several mining algorithm being used for frequent item set mining, yet most do not scale to the type of data we presented with today, so called “BIG DATA”. Big Data is a collection of large data sets. Our approach is to work on the frequent item set mining over the large dataset with scalable and speedy way. Big Data basically works with Map Reduce along with HDFS is used to find out frequent item sets from Big Data on large cluster. This paper focuses on using pre-processing & mining algorithm as hybrid approach for big data over Hadoop platform.

Keywords: frequent item set mining, big data, Hadoop, MapReduce

Procedia PDF Downloads 402
3910 Unlocking Green Hydrogen Potential: A Machine Learning-Based Assessment

Authors: Said Alshukri, Mazhar Hussain Malik

Abstract:

Green hydrogen is hydrogen produced using renewable energy sources. In the last few years, Oman aimed to reduce its dependency on fossil fuels. Recently, the hydrogen economy has become a global trend, and many countries have started to investigate the feasibility of implementing this sector. Oman created an alliance to establish the policy and rules for this sector. With motivation coming from both global and local interest in green hydrogen, this paper investigates the potential of producing hydrogen from wind and solar energies in three different locations in Oman, namely Duqm, Salalah, and Sohar. By using machine learning-based software “WEKA” and local metrological data, the project was designed to figure out which location has the highest wind and solar energy potential. First, various supervised models were tested to obtain their prediction accuracy, and it was found that the Random Forest (RF) model has the best prediction performance. The RF model was applied to 2021 metrological data for each location, and the results indicated that Duqm has the highest wind and solar energy potential. The system of one wind turbine in Duqm can produce 8335 MWh/year, which could be utilized in the water electrolysis process to produce 88847 kg of hydrogen mass, while a solar system consisting of 2820 solar cells is estimated to produce 1666.223 MWh/ year which is capable of producing 177591 kg of hydrogen mass.

Keywords: green hydrogen, machine learning, wind and solar energies, WEKA, supervised models, random forest

Procedia PDF Downloads 57
3909 Trajectory Tracking of a Redundant Hybrid Manipulator Using a Switching Control Method

Authors: Atilla Bayram

Abstract:

This paper presents the trajectory tracking control of a spatial redundant hybrid manipulator. This manipulator consists of two parallel manipulators which are a variable geometry truss (VGT) module. In fact, each VGT module with 3-degress of freedom (DOF) is a planar parallel manipulator and their operational planes of these VGT modules are arranged to be orthogonal to each other. Also, the manipulator contains a twist motion part attached to the top of the second VGT module to supply the missing orientation of the endeffector. These three modules constitute totally 7-DOF hybrid (parallel-parallel) redundant spatial manipulator. The forward kinematics equations of this manipulator are obtained, then, according to these equations, the inverse kinematics is solved based on an optimization with the joint limit avoidance. The dynamic equations are formed by using virtual work method. In order to test the performance of the redundant manipulator and the controllers presented, two different desired trajectories are followed by using the computed force control method and a switching control method. The switching control method is combined with the computed force control method and genetic algorithm. In the switching control method, the genetic algorithm is only used for fine tuning in the compensation of the trajectory tracking errors.

Keywords: computed force method, genetic algorithm, hybrid manipulator, inverse kinematics of redundant manipulators, variable geometry truss

Procedia PDF Downloads 321
3908 Molecular Detection and Antibiotics Resistance Pattern of Extended-Spectrum Beta-Lactamase Producing Escherichia coli in a Tertiary Hospital in Enugu, Nigeria

Authors: I. N. Nwafia, U. C. Ozumba, M. E. Ohanu, S. O. Ebede

Abstract:

Antibiotic resistance is increasing globally and has become a major health challenge. Extended-spectrum beta-lactamase is clinically important because the ESBL gene are mostly plasmid encoded and these plasmids frequently carry genes encoding resistance to other classes of antimicrobials thereby limiting antibiotic options in the treatment of infections caused by these organisms. The specific objectives of this study were to determine the prevalence of ESBLs production in Escherichia coli, to determine the antibiotic susceptibility pattern of ESBLs producing Escherichia coli, to detect TEM, SHV and CTX-M genes and the risk factors to acquisition of ESBL producing Escherichia coli. The protocol of the study was approved by Health Research and Ethics committee of the University of Nigeria Teaching Hospital (UNTH), Enugu. It was a descriptive cross-sectional study that involved all hospitalized patients in UNTH from whose specimens Escherichia coli was isolated during the period of the study. The samples analysed were urine, wound swabs, blood and cerebrospinal fluid. These samples were cultured in 5% sheep Blood agar and MacConkey agar (Oxoid Laboratories, Cambridge UK) and incubated at 35-370C for 24 hours. Escherichia coli was identified with standard biochemical tests and confirmed using API 20E auxanogram (bioMerieux, Marcy 1'Etoile, France). The antibiotic susceptibility testing was done by disc diffusion method and interpreted according to the Clinical and Laboratory Standard Institute guideline. ESBL production was confirmed using ESBL Epsilometer test strips (Liofilchem srl, Italy). The ESBL bla genes were detected with polymerase chain reaction, after extraction of DNA with plasmid mini-prep kit (Jena Bioscience, Jena, Germany). Data analysis was with appropriate descriptive and inferential statistics. One hundred and six isolates (53.00%) out of the 200 were from urine, followed by isolates from different swabs specimens 53(26.50%) and the least number of the isolates 4(2.00) were from blood (P value = 0.096). Seventy (35.00%) out of the 200 isolates, were confirmed positive for ESBL production. Forty-two (60.00%) of the isolates were from female patients while 28(40.00%) were from male patients (P value = 0.13). Sixty-eight (97.14%) of the isolates were susceptible to imipenem while all of the isolates were resistant to ampicillin, chloramphenicol and tetracycline. From the 70 positive isolates the ESBL genes detected with polymerase chain reaction were blaCTX-M (n=26; 37.14%), blaTEM (n=7; 10.00%), blaSHV (n=2; 2.86%), blaCTX-M/TEM (n=7; 10.0%), blaCTX-M/SHV (n=14; 20.0%) and blaCTX-M/TEM/SHV (n=10; 14.29%). There was no gene detected in 4(5.71%) of the isolates. The most associated risk factors to infections caused by ESBL producing Escherichia coli was previous antibiotics use for the past 3 months followed by admission in the intensive care unit, recent surgery, and urinary catheterization. In conclusion, ESBLs was detected in 4 of every 10 Escherichia coli with the predominant gene detected being CTX-M. This knowledge will enable appropriate measures towards improvement of patient health care, antibiotic stewardship, research and infection control in the hospital.

Keywords: antimicrobial, Escherichia coli, extended spectrum beta lactamase, resistance

Procedia PDF Downloads 276
3907 Sparse Principal Component Analysis: A Least Squares Approximation Approach

Authors: Giovanni Merola

Abstract:

Sparse Principal Components Analysis aims to find principal components with few non-zero loadings. We derive such sparse solutions by adding a genuine sparsity requirement to the original Principal Components Analysis (PCA) objective function. This approach differs from others because it preserves PCA's original optimality: uncorrelatedness of the components and least squares approximation of the data. To identify the best subset of non-zero loadings we propose a branch-and-bound search and an iterative elimination algorithm. This last algorithm finds sparse solutions with large loadings and can be run without specifying the cardinality of the loadings and the number of components to compute in advance. We give thorough comparisons with the existing sparse PCA methods and several examples on real datasets.

Keywords: SPCA, uncorrelated components, branch-and-bound, backward elimination

Procedia PDF Downloads 355
3906 A Distributed Cryptographically Generated Address Computing Algorithm for Secure Neighbor Discovery Protocol in IPv6

Authors: M. Moslehpour, S. Khorsandi

Abstract:

Due to shortage in IPv4 addresses, transition to IPv6 has gained significant momentum in recent years. Like Address Resolution Protocol (ARP) in IPv4, Neighbor Discovery Protocol (NDP) provides some functions like address resolution in IPv6. Besides functionality of NDP, it is vulnerable to some attacks. To mitigate these attacks, Internet Protocol Security (IPsec) was introduced, but it was not efficient due to its limitation. Therefore, SEND protocol is proposed to automatic protection of auto-configuration process. It is secure neighbor discovery and address resolution process. To defend against threats on NDP’s integrity and identity, Cryptographically Generated Address (CGA) and asymmetric cryptography are used by SEND. Besides advantages of SEND, its disadvantages like the computation process of CGA algorithm and sequentially of CGA generation algorithm are considerable. In this paper, we parallel this process between network resources in order to improve it. In addition, we compare the CGA generation time in self-computing and distributed-computing process. We focus on the impact of the malicious nodes on the CGA generation time in the network. According to the result, although malicious nodes participate in the generation process, CGA generation time is less than when it is computed in a one-way. By Trust Management System, detecting and insulating malicious nodes is easier.

Keywords: NDP, IPsec, SEND, CGA, modifier, malicious node, self-computing, distributed-computing

Procedia PDF Downloads 268
3905 Considering Effect of Wind Turbines in the Distribution System

Authors: Majed Ahmadi

Abstract:

In recent years, the high penetration of different types of renewable energy sources (RESs) has affected most of the available strategies. The main motivations behind the high penetration of RESs are clean energy, modular system and easy installation. Among different types of RESs, wind turbine (WT) is an interesting choice referring to the availability of wind in almost any area. The new technologies of WT can provide energy from residential applications to wide grid connected applications. Regarding the WT, advantages such as reducing the dependence on fossil fuels and enhancing the independence and flexibility of large power grid are the most prominent. Nevertheless, the high volatile nature of wind speed injects much uncertainty in the grid that if not managed optimally can put the analyses far from the reality.the aim of this project is scrutiny and to offer proper ways for renewing distribution networks with envisage the effects of wind power plants and uncertainties related to distribution systems including wind power generating plants output rate and consumers consuming rate and also decrease the incidents of the whole network losses, amount of pollution, voltage refraction and cost extent.to solve this problem we use dual point estimate method.And algorithm used in this paper is reformed bat algorithm, which will be under exact research furthermore the results.

Keywords: order renewal, wind turbines, bat algorithm, outspread production, uncertainty

Procedia PDF Downloads 264
3904 Fuzzy-Genetic Algorithm Multi-Objective Optimization Methodology for Cylindrical Stiffened Tanks Conceptual Design

Authors: H. Naseh, M. Mirshams, M. Mirdamadian, H. R. Fazeley

Abstract:

This paper presents an extension of fuzzy-genetic algorithm multi-objective optimization methodology that could effectively be used to find the overall satisfaction of objective functions (selecting the design variables) in the early stages of design process. The coupling of objective functions due to design variables in an engineering design process will result in difficulties in design optimization problems. In many cases, decision making on design variables conflicts with more than one discipline in system design. In space launch system conceptual design, decision making on some design variable (e.g. oxidizer to fuel mass flow rate O/F) in early stages of the design process is related to objective of liquid propellant engine (specific impulse) and Tanks (structure weight). Then, the primary application of this methodology is the design of a liquid propellant engine with the maximum specific impulse and cylindrical stiffened tank with the minimum weight. To this end, the design problem is established the fuzzy rule set based on designer's expert knowledge with a holistic approach. The independent design variables in this model are oxidizer to fuel mass flow rate, thickness of stringers, thickness of rings, shell thickness. To handle the mentioned problems, a fuzzy-genetic algorithm multi-objective optimization methodology is developed based on Pareto optimal set. Consequently, this methodology is modeled with the one stage of space launch system to illustrate accuracy and efficiency of proposed methodology.

Keywords: cylindrical stiffened tanks, multi-objective, genetic algorithm, fuzzy approach

Procedia PDF Downloads 635
3903 A Mini Radar System for Low Altitude Targets Detection

Authors: Kangkang Wu, Kaizhi Wang, Zhijun Yuan

Abstract:

This paper deals with a mini radar system aimed at detecting small targets at the low latitude. The radar operates at Ku-band in the frequency modulated continuous wave (FMCW) mode with two receiving channels. The radar system has the characteristics of compactness, mobility, and low power consumption. This paper focuses on the implementation of the radar system, and the Block least mean square (Block LMS) algorithm is applied to minimize the fortuitous distortion. It is validated from a series of experiments that the track of the unmanned aerial vehicle (UAV) can be easily distinguished with the radar system.

Keywords: unmanned aerial vehicle (UAV), interference, Block Least Mean Square (Block LMS) Algorithm, Frequency Modulated Continuous Wave (FMCW)

Procedia PDF Downloads 307
3902 Monte Carlo Pathwise Sensitivities for Barrier Options with Application to Coco-Bond Calibration

Authors: Thomas Gerstner, Bastian von Harrach, Daniel Roth

Abstract:

The Monte Carlo pathwise sensitivities approach is well established for smooth payoff functions. In this work, we present a new Monte Carlo algorithm that is able to calculate the pathwise sensitivities for discontinuous payoff functions. Our main tool is the one-step survival idea of Glasserman and Staum. Although this technique yields to new terms per observation, while differentiating, the algorithm is still efficient. As an application, we use the results for a two-dimensional calibration of a Coco-Bond, which we model with different types of discretely monitored barrier options.

Keywords: Monte Carlo, discretely monitored barrier options, pathwise sensitivities, Coco-Bond

Procedia PDF Downloads 338
3901 Paddy/Rice Singulation for Determination of Husking Efficiency and Damage Using Machine Vision

Authors: M. Shaker, S. Minaei, M. H. Khoshtaghaza, A. Banakar, A. Jafari

Abstract:

In this study a system of machine vision and singulation was developed to separate paddy from rice and determine paddy husking and rice breakage percentages. The machine vision system consists of three main components including an imaging chamber, a digital camera, a computer equipped with image processing software. The singulation device consists of a kernel holding surface, a motor with vacuum fan, and a dimmer. For separation of paddy from rice (in the image), it was necessary to set a threshold. Therefore, some images of paddy and rice were sampled and the RGB values of the images were extracted using MATLAB software. Then mean and standard deviation of the data were determined. An Image processing algorithm was developed using MATLAB to determine paddy/rice separation and rice breakage and paddy husking percentages, using blue to red ratio. Tests showed that, a threshold of 0.75 is suitable for separating paddy from rice kernels. Results from the evaluation of the image processing algorithm showed that the accuracies obtained with the algorithm were 98.36% and 91.81% for paddy husking and rice breakage percentage, respectively. Analysis also showed that a suction of 45 mmHg to 50 mmHg yielding 81.3% separation efficiency is appropriate for operation of the kernel singulation system.

Keywords: breakage, computer vision, husking, rice kernel

Procedia PDF Downloads 353
3900 Parallel Gripper Modelling and Design Optimization Using Multi-Objective Grey Wolf Optimizer

Authors: Golak Bihari Mahanta, Bibhuti Bhusan Biswal, B. B. V. L. Deepak, Amruta Rout, Gunji Balamurali

Abstract:

Robots are widely used in the manufacturing industry for rapid production with higher accuracy and precision. With the help of End-of-Arm Tools (EOATs), robots are interacting with the environment. Robotic grippers are such EOATs which help to grasp the object in an automation system for improving the efficiency. As the robotic gripper directly influence the quality of the product due to the contact between the gripper surface and the object to be grasped, it is necessary to design and optimize the gripper mechanism configuration. In this study, geometric and kinematic modeling of the parallel gripper is proposed. Grey wolf optimizer algorithm is introduced for solving the proposed multiobjective gripper optimization problem. Two objective functions developed from the geometric and kinematic modeling along with several nonlinear constraints of the proposed gripper mechanism is used to optimize the design variables of the systems. Finally, the proposed methodology compared with a previously proposed method such as Teaching Learning Based Optimization (TLBO) algorithm, NSGA II, MODE and it was seen that the proposed method is more efficient compared to the earlier proposed methodology.

Keywords: gripper optimization, metaheuristics, , teaching learning based algorithm, multi-objective optimization, optimal gripper design

Procedia PDF Downloads 172
3899 An Enhanced Hybrid Backoff Technique for Minimizing the Occurrence of Collision in Mobile Ad Hoc Networks

Authors: N. Sabiyath Fatima, R. K. Shanmugasundaram

Abstract:

In Mobile Ad-hoc Networks (MANETS), every node performs both as transmitter and receiver. The existing backoff models do not exactly forecast the performance of the wireless network. Also, the existing models experience elevated packet collisions. Every time a collision happens, the station’s contention window (CW) is doubled till it arrives at the utmost value. The main objective of this paper is to diminish collision by means of contention window Multiplicative Increase Decrease Backoff (CWMIDB) scheme. The intention of rising CW is to shrink the collision possibility by distributing the traffic into an outsized point in time. Within wireless Ad hoc networks, the CWMIDB algorithm dynamically controls the contention window of the nodes experiencing collisions. During packet communication, the backoff counter is evenly selected from the given choice of [0, CW-1]. At this point, CW is recognized as contention window and its significance lies on the amount of unsuccessful transmission that had happened for the packet. On the initial transmission endeavour, CW is put to least amount value (C min), if transmission effort fails, subsequently the value gets doubled, and once more the value is set to least amount on victorious broadcast. CWMIDB is simulated inside NS2 environment and its performance is compared with Binary Exponential Backoff Algorithm. The simulation results show improvement in transmission probability compared to that of the existing backoff algorithm.

Keywords: backoff, contention window, CWMIDB, MANET

Procedia PDF Downloads 259
3898 Writing a Parametric Design Algorithm Based on Recreation and Structural Analysis of Patkane Model: The Case Study of Oshtorjan Mosque

Authors: Behnoush Moghiminia, Jesus Anaya Diaz

Abstract:

The current study attempts to present the relationship between the structure development and Patkaneh as one of the Iranian geometric patterns and parametric algorithms by introducing two practical methods. While having a structural function, Patkaneh is also used as an ornamental element. It can be helpful in the scientific and practical review of Patkaneh. The current study aims to use Patkaneh as a parametric form generator based on the algorithm. The current paper attempts to express how can a more complete algorithm of this covering be obtained based on the parametric study and analysis of a sample of a Patkaneh and also investigate the relationship between the development of the geometrical pattern of Patkaneh as a structural-decorative element of Iranian architecture and digital design. In this regard, to achieve the research purposes, researchers investigated the oldest type of Patkaneh in the architecture history of Iran, such as the Northern Entrance Patkaneh of Oshtorjan Jame’ Mosque. An accurate investigation was done on the history of the background to answer the questions. Then, by investigating the structural behavior of Patkaneh, the decorative or structural-decorative role of Patkaneh was investigated to eliminate the ambiguity. Then, the geometrical structure of Patkaneh was analyzed by introducing two practical methods. The first method is based on the constituent units of Patkaneh (Square and diamond) and investigating the interactive relationships between them in 2D and 3D. This method is appropriate for cases where there are rational and regular geometrical relationships. The second method is based on the separation of the floors and the investigation of their interrelation. It is practical when the constituent units are not geometrically regular and have numerous diversity. Finally, the parametric form algorithm of these methods was codified.

Keywords: geometric properties, parametric design, Patkaneh, structural analysis

Procedia PDF Downloads 134
3897 The Feasibility of Anaerobic Digestion at 45⁰C

Authors: Nuruol S. Mohd, Safia Ahmed, Rumana Riffat, Baoqiang Li

Abstract:

Anaerobic digestion at mesophilic and thermophilic temperatures have been widely studied and evaluated by numerous researchers. Limited extensive research has been conducted on anaerobic digestion in the intermediate zone of 45°C, mainly due to the notion that limited microbial activity occurs within this zone. The objectives of this research were to evaluate the performance and the capability of anaerobic digestion at 45°C in producing class A biosolids, in comparison to a mesophilic and thermophilic anaerobic digestion system operated at 35°C and 55°C, respectively. In addition to that, the investigation on the possible inhibition factors affecting the performance of the digestion system at this temperature will be conducted as well. The 45°C anaerobic digestion systems were not able to achieve comparable methane yield and high-quality effluent compared to the mesophilic system, even though the systems produced biogas with about 62-67% methane. The 45°C digesters suffered from high acetate accumulation, but sufficient buffering capacity was observed as the pH, alkalinity and volatile fatty acids (VFA)-to-alkalinity ratio were within recommended values. The accumulation of acetate observed in 45°C systems were presumably due to the high temperature which contributed to high hydrolysis rate. Consequently, it produced a large amount of toxic salts that combined with the substrate making them not readily available to be consumed by methanogens. Acetate accumulation, even though contributed to 52 to 71% reduction in acetate degradation process, could not be considered as completely inhibitory. Additionally, at 45°C, no ammonia inhibition was observed and the digesters were able to achieve volatile solids (VS) reduction of 47.94±4.17%. The pathogen counts were less than 1,000 MPN/g total solids, thus, producing Class A biosolids.

Keywords: 45°C anaerobic digestion, acetate accumulation, class A biosolids, salt toxicity

Procedia PDF Downloads 289
3896 Insider Theft Detection in Organizations Using Keylogger and Machine Learning

Authors: Shamatha Shetty, Sakshi Dhabadi, Prerana M., Indushree B.

Abstract:

About 66% of firms claim that insider attacks are more likely to happen. The frequency of insider incidents has increased by 47% in the last two years. The goal of this work is to prevent dangerous employee behavior by using keyloggers and the Machine Learning (ML) model. Every keystroke that the user enters is recorded by the keylogging program, also known as keystroke logging. Keyloggers are used to stop improper use of the system. This enables us to collect all textual data, save it in a CSV file, and analyze it using an ML algorithm and the VirusTotal API. Many large companies use it to methodically monitor how their employees use computers, the internet, and email. We are utilizing the SVM algorithm and the VirusTotal API to improve overall efficiency and accuracy in identifying specific patterns and words to automate and offer the report for improved monitoring.

Keywords: cyber security, machine learning, cyclic process, email notification

Procedia PDF Downloads 40