Search results for: Inverse Distance Weighting.
795 Palmprint Recognition by Wavelet Transform with Competitive Index and PCA
Authors: Deepti Tamrakar, Pritee Khanna
Abstract:
This manuscript presents, palmprint recognition by combining different texture extraction approaches with high accuracy. The Region of Interest (ROI) is decomposed into different frequencytime sub-bands by wavelet transform up-to two levels and only the approximate image of two levels is selected, which is known as Approximate Image ROI (AIROI). This AIROI has information of principal lines of the palm. The Competitive Index is used as the features of the palmprint, in which six Gabor filters of different orientations convolve with the palmprint image to extract the orientation information from the image. The winner-take-all strategy is used to select dominant orientation for each pixel, which is known as Competitive Index. Further, PCA is applied to select highly uncorrelated Competitive Index features, to reduce the dimensions of the feature vector, and to project the features on Eigen space. The similarity of two palmprints is measured by the Euclidean distance metrics. The algorithm is tested on Hong Kong PolyU palmprint database. Different AIROI of different wavelet filter families are also tested with the Competitive Index and PCA. AIROI of db7 wavelet filter achievs Equal Error Rate (EER) of 0.0152% and Genuine Acceptance Rate (GAR) of 99.67% on the palm database of Hong Kong PolyU.Keywords: DWT, EER, Euclidean Distance, Gabor filter, PCA, ROI.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1741794 Enabling the Physical Elements of a Pedestrian Friendly District around a Rail Station for Supporting Transit Oriented Development
Authors: Dyah Titisari Widyastuti
Abstract:
Rail-station area development that is based on the concept of TOD (Transit Oriented Development) is principally oriented to pedestrian accessibility for daily mobility. The aim of this research is elaborating how far the existing physical elements of a rail-station district could facilitate pedestrian mobility and establish a pedestrian friendly district toward implementation of a TOD concept. This research was conducted through some steps: (i) mapping the rail-station area pedestrian sidewalk and pedestrian network as well as activity nodes and transit nodes, (ii) assessing the level of pedestrian sidewalk connectivity joining trip origin and destination. The research area coverage in this case is limited to walking distance of the rail station (around 500 meters or 10-15 minutes walking). The findings of this research on the current condition of the street and pedestrian sidewalk network and connectivity, show good preference for the foot modal share (more than 50%) is achieved. Nevertheless, it depends on the distance from the trip origin to destination.
Keywords: Accessibility of daily mobility, pedestrian friendly district, rail-station district, Transit Oriented Development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 882793 Effect of Support Distance on Damage of Drilled Thin CFRP Laminates
Authors: Jean François Chatelain, Imed Zaghbani, Gilbert Lebrun, Kaml Hasni
Abstract:
Severe damages may occur during the drilling of carbon fiber reinforced plastics (CFRP). In practice, this damage is limited by adding a backup support to the drilled parts. For some aeronautical parts with curvatures, backing up parts is a demanding process. In order to simplify the operation, this research studies the effect of using a configurable setup to support parts on the resulting quality of drilled holes. The test coupons referenced in this study are twenty four-plies unidirectional laminates made of carbon fibers and epoxy resin. Different signals were measured during the drilling process for these laminates, including the thrust force, the displacement and the acceleration. The processing of these signals demonstrated that the damage is due to the combination of two main factors: the spring-back of the thin part and the thrust force. The results found were confirmed for different feeds and speeds. When the distance between supports is increased, it is observed that the spring-back increases but the thrust force decreases. The study proves the feasibility of unsupported drilling of thin CFRP laminates without creating any observable damage.
Keywords: CFRP, Damage, Drilling, Flexible setup.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1796792 The Influence of the Geogrid Layers on the Bearing Capacity of Layered Soils
Authors: S. A. Naeini, H. R. Rahmani, M. Hossein Zade
Abstract:
Many classical bearing capacity theories assume that the natural soil's layers are homogenous for determining the bearing capacity of the soil. But, in many practical projects, we encounter multi-layer soils. Geosynthetic as reinforcement materials have been extensively used in the construction of various structures. In this paper, numerical analysis of the Plate Load Test (PLT) using of ABAQUS software in double-layered soils with different thicknesses of sandy and gravelly layers reinforced with geogrid was considered. The PLT is one of the common filed methods to calculate parameters such as soil bearing capacity, the evaluation of the compressibility and the determination of the Subgrade Reaction module. In fact, the influence of the geogrid layers on the bearing capacity of the layered soils is investigated. Finally, the most appropriate mode for the distance and number of reinforcement layers is determined. Results show that using three layers of geogrid with a distance of 0.3 times the width of the loading plate has the highest efficiency in bearing capacity of double-layer (sand and gravel) soils. Also, the significant increase in bearing capacity between unreinforced and reinforced soil with three layers of geogrid is caused by the condition that the upper layer (gravel) thickness is equal to the loading plate width.
Keywords: Bearing capacity, reinforcement, geogrid, plate load test, layered soils.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 844791 Springback Property and Texture Distribution of Grained Pure Copper
Authors: Takashi Sakai, Hitoshi Omata, Jun-Ichi Koyama
Abstract:
To improve the material characteristics of single- and poly-crystals of pure copper, the respective relationships between crystallographic orientations and microstructures, and the bending and mechanical properties were examined. And texture distribution is also analyzed. A grain refinement procedure was performed to obtain a grained structure. Furthermore, some analytical results related to crystal direction maps, inverse pole figures, and textures were obtained from SEM-EBSD analyses. Results showed that these grained metallic materials have peculiar springback characteristics with various bending angles.Keywords: Pure Copper, Grain Refinement, Environmental Materials, SEM-EBSD Analysis, Texture, Microstructure
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2169790 On Discretization of Second-order Derivatives in Smoothed Particle Hydrodynamics
Authors: R. Fatehi, M.A. Fayazbakhsh, M.T. Manzari
Abstract:
Discretization of spatial derivatives is an important issue in meshfree methods especially when the derivative terms contain non-linear coefficients. In this paper, various methods used for discretization of second-order spatial derivatives are investigated in the context of Smoothed Particle Hydrodynamics. Three popular forms (i.e. "double summation", "second-order kernel derivation", and "difference scheme") are studied using one-dimensional unsteady heat conduction equation. To assess these schemes, transient response to a step function initial condition is considered. Due to parabolic nature of the heat equation, one can expect smooth and monotone solutions. It is shown, however in this paper, that regardless of the type of kernel function used and the size of smoothing radius, the double summation discretization form leads to non-physical oscillations which persist in the solution. Also, results show that when a second-order kernel derivative is used, a high-order kernel function shall be employed in such a way that the distance of inflection point from origin in the kernel function be less than the nearest particle distance. Otherwise, solutions may exhibit oscillations near discontinuities unlike the "difference scheme" which unconditionally produces monotone results.Keywords: Heat conduction, Meshfree methods, Smoothed ParticleHydrodynamics (SPH), Second-order derivatives.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3089789 Fuzzy Numbers and MCDM Methods for Portfolio Optimization
Authors: Thi T. Nguyen, Lee N. Gordon-Brown
Abstract:
A new deployment of the multiple criteria decision making (MCDM) techniques: the Simple Additive Weighting (SAW), and the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) for portfolio allocation, is demonstrated in this paper. Rather than exclusive reference to mean and variance as in the traditional mean-variance method, the criteria used in this demonstration are the first four moments of the portfolio distribution. Each asset is evaluated based on its marginal impacts to portfolio higher moments that are characterized by trapezoidal fuzzy numbers. Then centroid-based defuzzification is applied to convert fuzzy numbers to the crisp numbers by which SAW and TOPSIS can be deployed. Experimental results suggest the similar efficiency of these MCDM approaches to selecting dominant assets for an optimal portfolio under higher moments. The proposed approaches allow investors flexibly adjust their risk preferences regarding higher moments via different schemes adapting to various (from conservative to risky) kinds of investors. The other significant advantage is that, compared to the mean-variance analysis, the portfolio weights obtained by SAW and TOPSIS are consistently well-diversified.Keywords: Fuzzy numbers, SAW, TOPSIS, portfolio optimization, higher moments, risk management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2144788 E-Education in Multicultural Setting: The Success of Mobile Learning
Authors: Subramaniam Chandran
Abstract:
This paper explains how mobile learning assures sustainable e-education for multicultural group of students. This paper reports the impact of mobile learning on distance education in multicultural environment. The emergence of learning technologies through CD, internet, and mobile is increasingly adopted by distance institutes for quick delivery and cost-effective purposes. Their sustainability is conditioned by the structure of learners as well as the teaching community. The experimental study was conducted among the distant learners of Vinayaka Missions University located at Salem in India. Students were drawn from multicultural environment based on different languages, religions, class and communities. During the mobile learning sessions, the students, who are divided on language, religion, class and community, were dominated by play impulse rather than study anxiety or cultural inhibitions. This study confirmed that mobile learning improved the performance of the students despite their division based on region, language or culture. In other words, technology was able to transcend the relative deprivation in the multicultural groups. It also confirms sustainable e-education through mobile learning and cost-effective system of instruction. Mobile learning appropriates the self-motivation and play impulse of the young learners in providing sustainable e-education to multicultural social groups of students.
Keywords: E-Education, mobile learning, multiculturalism.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2050787 Piecewise Interpolation Filter for Effective Processing of Large Signal Sets
Authors: Anatoli Torokhti, Stanley Miklavcic
Abstract:
Suppose KY and KX are large sets of observed and reference signals, respectively, each containing N signals. Is it possible to construct a filter F : KY → KX that requires a priori information only on few signals, p N, from KX but performs better than the known filters based on a priori information on every reference signal from KX? It is shown that the positive answer is achievable under quite unrestrictive assumptions. The device behind the proposed method is based on a special extension of the piecewise linear interpolation technique to the case of random signal sets. The proposed technique provides a single filter to process any signal from the arbitrarily large signal set. The filter is determined in terms of pseudo-inverse matrices so that it always exists.Keywords: Wiener filter, filtering of stochastic signals.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1412786 Transient Currents in a Double Conductor Line above a Conducting Half-Space
Authors: Valentina Koliskina, Inta Volodko
Abstract:
Transient eddy current problem is solved in the present paper by the method of the Laplace transform for the case of a double conductor line located parallel to a conducting half-space. The Fourier sine and cosine integral transforms are used in order to find the Laplace transform of the solution. The inverse Laplace transform of the solution is found in closed form. The integrated electromotive force per unit length of the double conductor line is calculated in the form of an improper integral.Keywords: Transient eddy currents, Laplace transform, double conductor line.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1433785 Identification of Disease Causing DNA Motifs in Human DNA Using Clustering Approach
Authors: G. Tamilpavai, C. Vishnuppriya
Abstract:
Studying DNA (deoxyribonucleic acid) sequence is useful in biological processes and it is applied in the fields such as diagnostic and forensic research. DNA is the hereditary information in human and almost all other organisms. It is passed to their generations. Earlier stage detection of defective DNA sequence may lead to many developments in the field of Bioinformatics. Nowadays various tedious techniques are used to identify defective DNA. The proposed work is to analyze and identify the cancer-causing DNA motif in a given sequence. Initially the human DNA sequence is separated as k-mers using k-mer separation rule. The separated k-mers are clustered using Self Organizing Map (SOM). Using Levenshtein distance measure, cancer associated DNA motif is identified from the k-mer clusters. Experimental results of this work indicate the presence or absence of cancer causing DNA motif. If the cancer associated DNA motif is found in DNA, it is declared as the cancer disease causing DNA sequence. Otherwise the input human DNA is declared as normal sequence. Finally, elapsed time is calculated for finding the presence of cancer causing DNA motif using clustering formation. It is compared with normal process of finding cancer causing DNA motif. Locating cancer associated motif is easier in cluster formation process than the other one. The proposed work will be an initiative aid for finding genetic disease related research.
Keywords: Bioinformatics, cancer motif, DNA, k-mers, Levenshtein distance, SOM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1386784 A Real-time Computer Vision System for VehicleTracking and Collision Detection
Authors: Mustafa Kisa, Fatih Mehmet Botsali
Abstract:
Recent developments in automotive technology are focused on economy, comfort and safety. Vehicle tracking and collision detection systems are attracting attention of many investigators focused on safety of driving in the field of automotive mechatronics. In this paper, a vision-based vehicle detection system is presented. Developed system is intended to be used in collision detection and driver alert. The system uses RGB images captured by a camera in a car driven in the highway. Images captured by the moving camera are used to detect the moving vehicles in the image. A vehicle ahead of the camera is detected in daylight conditions. The proposed method detects moving vehicles by subtracting successive images. Plate height of the vehicle is determined by using a plate recognition algorithm. Distance of the moving object is calculated by using the plate height. After determination of the distance of the moving vehicle relative speed of the vehicle and Time-to-Collision are calculated by using distances measured in successive images. Results obtained in road tests are discussed in order to validate the use of the proposed method.
Keywords: Image possessing, vehicle tracking, license plate detection, computer vision.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3100783 Ordinary Differential Equations with Inverted Functions
Authors: Thomas Kampke
Abstract:
Equations with differentials relating to the inverse of an unknown function rather than to the unknown function itself are solved exactly for some special cases and numerically for the general case. Invertibility combined with differentiability over connected domains forces solutions always to be monotone. Numerical function inversion is key to all solution algorithms which either are of a forward type or a fixed point type considering whole approximate solution functions in each iteration. The given considerations are restricted to ordinary differential equations with inverted functions (ODEIs) of first order. Forward type computations, if applicable, admit consistency of order one and, under an additional accuracy condition, convergence of order one.
Keywords: Euler method, fixed points, golden section, multi-step procedures, Runge Kutta methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1449782 Multi-Focus Image Fusion Using SFM and Wavelet Packet
Authors: Somkait Udomhunsakul
Abstract:
In this paper, a multi-focus image fusion method using Spatial Frequency Measurements (SFM) and Wavelet Packet was proposed. The proposed fusion approach, firstly, the two fused images were transformed and decomposed into sixteen subbands using Wavelet packet. Next, each subband was partitioned into sub-blocks and each block was identified the clearer regions by using the Spatial Frequency Measurement (SFM). Finally, the recovered fused image was reconstructed by performing the Inverse Wavelet Transform. From the experimental results, it was found that the proposed method outperformed the traditional SFM based methods in terms of objective and subjective assessments.
Keywords: Multi-focus image fusion, Wavelet Packet, Spatial Frequency Measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1615781 Neighbour Cell List Reduction in Multi-Tier Heterogeneous Networks
Authors: Mohanad Alhabo, Naveed Nawaz
Abstract:
The ongoing call or data session must be maintained to ensure a good quality of service. This can be accomplished by performing handover procedure while the user is on the move. However, dense deployment of small cells in 5G networks is a challenging issue due to the extensive number of handovers. In this paper, a neighbour cell list method is proposed to reduce the number of target small cells and hence minimizing the number of handovers. The neighbour cell list is built by omitting cells that could cause an unnecessary handover and/or handover failure because of short time of stay of a user in these cells. A multi-attribute decision making technique, simple additive weighting, is then applied to the optimized neighbour cell list. The performance of the proposed method is analysed and compared with that of the existing methods. Results disclose that our method decreases the candidate small cell list, unnecessary handovers, handover failure and short time of stay cells compared to the competitive method.
Keywords: Handover, HetNets, MADM, small cells.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 541780 Effect of Size of the Step in the Response Surface Methodology using Nonlinear Test Functions
Authors: Jesús Everardo Olguín Tiznado, Rafael García Martínez, Claudia Camargo Wilson, Juan Andrés López Barreras, Everardo Inzunza González, Javier Ordorica Villalvazo
Abstract:
The response surface methodology (RSM) is a collection of mathematical and statistical techniques useful in the modeling and analysis of problems in which the dependent variable receives the influence of several independent variables, in order to determine which are the conditions under which should operate these variables to optimize a production process. The RSM estimated a regression model of first order, and sets the search direction using the method of maximum / minimum slope up / down MMS U/D. However, this method selects the step size intuitively, which can affect the efficiency of the RSM. This paper assesses how the step size affects the efficiency of this methodology. The numerical examples are carried out through Monte Carlo experiments, evaluating three response variables: efficiency gain function, the optimum distance and the number of iterations. The results in the simulation experiments showed that in response variables efficiency and gain function at the optimum distance were not affected by the step size, while the number of iterations is found that the efficiency if it is affected by the size of the step and function type of test used.Keywords: RSM, dependent variable, independent variables, efficiency, simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1989779 Transient Analysis and Mitigation of Capacitor Bank Switching on a Standalone Wind Farm
Authors: Ajibola O. Akinrinde, Andrew Swanson, Remy Tiako
Abstract:
There exist significant losses on transmission lines due to distance, as power generating stations could be located far from some isolated settlements. Standalone wind farms could be a good choice of alternative power generation for such settlements that are far from the grid due to factors of long distance or socio-economic problems. However, uncompensated wind farms consume reactive power since wind turbines are induction generators. Therefore, capacitor banks are used to compensate reactive power, which in turn improves the voltage profile of the network. Although capacitor banks help improving voltage profile, they also undergo switching actions due to its compensating response to the variation of various types of load at the consumer’s end. These switching activities could cause transient overvoltage on the network, jeopardizing the end-life of other equipment on the system. In this paper, the overvoltage caused by these switching activities is investigated using the IEEE bus 14-network to represent a standalone wind farm, and the simulation is done using ATP/EMTP software. Scenarios involving the use of pre-insertion resistor and pre-insertion inductor, as well as controlled switching was also carried out in order to decide the best mitigation option to reduce the overvoltage.
Keywords: Capacitor banks, IEEE bus 14-network, Pre-insertion resistor, Standalone wind farm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2293778 A Text Clustering System based on k-means Type Subspace Clustering and Ontology
Authors: Liping Jing, Michael K. Ng, Xinhua Yang, Joshua Zhexue Huang
Abstract:
This paper presents a text clustering system developed based on a k-means type subspace clustering algorithm to cluster large, high dimensional and sparse text data. In this algorithm, a new step is added in the k-means clustering process to automatically calculate the weights of keywords in each cluster so that the important words of a cluster can be identified by the weight values. For understanding and interpretation of clustering results, a few keywords that can best represent the semantic topic are extracted from each cluster. Two methods are used to extract the representative words. The candidate words are first selected according to their weights calculated by our new algorithm. Then, the candidates are fed to the WordNet to identify the set of noun words and consolidate the synonymy and hyponymy words. Experimental results have shown that the clustering algorithm is superior to the other subspace clustering algorithms, such as PROCLUS and HARP and kmeans type algorithm, e.g., Bisecting-KMeans. Furthermore, the word extraction method is effective in selection of the words to represent the topics of the clusters.
Keywords: Subspace Clustering, Text Mining, Feature Weighting, Cluster Interpretation, Ontology
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2462777 A Practical and Efficient Evaluation Function for 3D Model Based Vehicle Matching
Authors: Yuan Zheng
Abstract:
3D model-based vehicle matching provides a new way for vehicle recognition, localization and tracking. Its key is to construct an evaluation function, also called fitness function, to measure the degree of vehicle matching. The existing fitness functions often poorly perform when the clutter and occlusion exist in traffic scenarios. In this paper, we present a practical and efficient fitness function. Unlike the existing evaluation functions, the proposed fitness function is to study the vehicle matching problem from both local and global perspectives, which exploits the pixel gradient information as well as the silhouette information. In view of the discrepancy between 3D vehicle model and real vehicle, a weighting strategy is introduced to differently treat the fitting of the model’s wireframes. Additionally, a normalization operation for the model’s projection is performed to improve the accuracy of the matching. Experimental results on real traffic videos reveal that the proposed fitness function is efficient and robust to the cluttered background and partial occlusion.Keywords: 3D-2D matching, fitness function, 3D vehicle model, local image gradient, silhouette information.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1634776 Aerial Firefighting Aircraft Selection with Standard Fuzzy Sets using Multiple Criteria Group Decision Making Analysis
Authors: C. Ardil
Abstract:
Aircraft selection decisions can be challenging due to their multidimensional and interdisciplinary nature. They involve multiple stakeholders with conflicting objectives and numerous alternative options with uncertain outcomes. This study focuses on the analysis of aerial firefighting aircraft that can be chosen for the Air Fire Service to extinguish forest fires. To make such a selection, the characteristics of the fire zones must be considered, and the capability to manage the logistics involved in such operations, as well as the purchase and maintenance of the aircraft, must be determined. The selection of firefighting aircraft is particularly complex because they have longer fleet lives and require more demanding operation and maintenance than scheduled passenger air service. This paper aims to use the fuzzy proximity measure method to select the most appropriate aerial firefighting aircraft based on decision criteria using multiple attribute decision making analysis. Following fuzzy decision analysis, the most suitable aerial firefighting aircraft is ranked and determined for the Air Fire Service.
Keywords: Aerial firefighting aircraft selection, multiple criteria decision making, fuzzy sets, standard fuzzy sets, determinate fuzzy sets, indeterminate fuzzy sets, proximity measure method, Minkowski distance family function, Hausdorff distance function, MCDM, PMM, PMM-F
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 400775 Hippocampus Segmentation using a Local Prior Model on its Boundary
Authors: Dimitrios Zarpalas, Anastasios Zafeiropoulos, Petros Daras, Nicos Maglaveras
Abstract:
Segmentation techniques based on Active Contour Models have been strongly benefited from the use of prior information during their evolution. Shape prior information is captured from a training set and is introduced in the optimization procedure to restrict the evolution into allowable shapes. In this way, the evolution converges onto regions even with weak boundaries. Although significant effort has been devoted on different ways of capturing and analyzing prior information, very little thought has been devoted on the way of combining image information with prior information. This paper focuses on a more natural way of incorporating the prior information in the level set framework. For proof of concept the method is applied on hippocampus segmentation in T1-MR images. Hippocampus segmentation is a very challenging task, due to the multivariate surrounding region and the missing boundary with the neighboring amygdala, whose intensities are identical. The proposed method, mimics the human segmentation way and thus shows enhancements in the segmentation accuracy.Keywords: Medical imaging & processing, Brain MRI segmentation, hippocampus segmentation, hippocampus-amygdala missingboundary, weak boundary segmentation, region based segmentation, prior information, local weighting scheme in level sets, spatialdistribution of labels, gradient distribution on boundary.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1752774 NSGA Based Optimal Volt / Var Control in Distribution System with Dispersed Generation
Authors: P. N. Hrisheekesha, Jaydev Sharma
Abstract:
In this paper, a method based on Non-Dominated Sorting Genetic Algorithm (NSGA) has been presented for the Volt / Var control in power distribution systems with dispersed generation (DG). Genetic algorithm approach is used due to its broad applicability, ease of use and high accuracy. The proposed method is better suited for volt/var control problems. A multi-objective optimization problem has been formulated for the volt/var control of the distribution system. The non-dominated sorting genetic algorithm based method proposed in this paper, alleviates the problem of tuning the weighting factors required in solving the multi-objective volt/var control optimization problems. Based on the simulation studies carried out on the distribution system, the proposed scheme has been found to be simple, accurate and easy to apply to solve the multiobjective volt/var control optimization problem of the distribution system with dispersed generation.Keywords: Dispersed Generation, Distribution System, Non-Dominated Sorting Genetic Algorithm, Voltage / Reactive powercontrol.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1630773 Ontology-based Concept Weighting for Text Documents
Authors: Hmway Hmway Tar, Thi Thi Soe Nyaunt
Abstract:
Documents clustering become an essential technology with the popularity of the Internet. That also means that fast and high-quality document clustering technique play core topics. Text clustering or shortly clustering is about discovering semantically related groups in an unstructured collection of documents. Clustering has been very popular for a long time because it provides unique ways of digesting and generalizing large amounts of information. One of the issues of clustering is to extract proper feature (concept) of a problem domain. The existing clustering technology mainly focuses on term weight calculation. To achieve more accurate document clustering, more informative features including concept weight are important. Feature Selection is important for clustering process because some of the irrelevant or redundant feature may misguide the clustering results. To counteract this issue, the proposed system presents the concept weight for text clustering system developed based on a k-means algorithm in accordance with the principles of ontology so that the important of words of a cluster can be identified by the weight values. To a certain extent, it has resolved the semantic problem in specific areas.Keywords: Clustering, Concept Weight, Document clustering, Feature Selection, Ontology
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2406772 Using Different Aspects of the Signings for Appearance-based Sign Language Recognition
Authors: Morteza Zahedi, Philippe Dreuw, Thomas Deselaers, Hermann Ney
Abstract:
Sign language is used by the deaf and hard of hearing people for communication. Automatic sign language recognition is a challenging research area since sign language often is the only way of communication for the deaf people. Sign language includes different components of visual actions made by the signer using the hands, the face, and the torso, to convey his/her meaning. To use different aspects of signs, we combine the different groups of features which have been extracted from the image frames recorded directly by a stationary camera. We combine the features in two levels by employing three techniques. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, or by concatenating feature groups over time and using LDA to choose the most discriminant elements. At the model level, a late fusion of differently trained models can be carried out by a log-linear model combination. In this paper, we investigate these three combination techniques in an automatic sign language recognition system and show that the recognition rate can be significantly improved.
Keywords: American sign language, appearance-based features, Feature combination, Sign language recognition
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1399771 Swarmed Discriminant Analysis for Multifunction Prosthesis Control
Authors: Rami N. Khushaba, Ahmed Al-Ani, Adel Al-Jumaily
Abstract:
One of the approaches enabling people with amputated limbs to establish some sort of interface with the real world includes the utilization of the myoelectric signal (MES) from the remaining muscles of those limbs. The MES can be used as a control input to a multifunction prosthetic device. In this control scheme, known as the myoelectric control, a pattern recognition approach is usually utilized to discriminate between the MES signals that belong to different classes of the forearm movements. Since the MES is recorded using multiple channels, the feature vector size can become very large. In order to reduce the computational cost and enhance the generalization capability of the classifier, a dimensionality reduction method is needed to identify an informative yet moderate size feature set. This paper proposes a new fuzzy version of the well known Fisher-s Linear Discriminant Analysis (LDA) feature projection technique. Furthermore, based on the fact that certain muscles might contribute more to the discrimination process, a novel feature weighting scheme is also presented by employing Particle Swarm Optimization (PSO) for estimating the weight of each feature. The new method, called PSOFLDA, is tested on real MES datasets and compared with other techniques to prove its superiority.Keywords: Discriminant Analysis, Pattern Recognition, SignalProcessing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1556770 Curvature Ductility Factor of Rectangular Sections Reinforced Concrete Beams
Authors: Y. Si Youcef, M. Chemrouk
Abstract:
The present work presents a method of calculating the ductility of rectangular sections of beams considering nonlinear behavior of concrete and steel. This calculation procedure allows us to trace the curvature of the section according to the bending moment, and consequently deduce ductility. It also allowed us to study the various parameters that affect the value of the ductility. A comparison of the effect of maximum rates of tension steel, adopted by the codes, ACI [1], EC8 [2] and RPA [3] on the value of the ductility was made. It was concluded that the maximum rate of steels permitted by the ACI [1] codes and RPA [3] are almost similar in their effect on the ductility and too high. Therefore, the ductility mobilized in case of an earthquake is low, the inverse of code EC8 [2]. Recommendations have been made in this direction.Keywords: Ductility, beam, reinforced concrete, seismic code, relationship, time bending, resistance, non-linear behavior.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6194769 The Empirical Survey on the Effect of Using Media in Explosive Forming of Tubular Shells
Authors: V. Hadavi, J. Zamani, R. Hosseini
Abstract:
The special and unique advantages of explosive forming, has developed its use in different industries. Considering the important influence of improving the current explosive forming techniques on increasing the efficiency and control over the explosive forming procedure, the effects of air and water as the energy-conveying medium, and also their differences will be illustrated in this paper. Hence, a large number of explosive forming tests have been conducted on two sizes of thin walled cylindrical shells by using air and water as the working medium. Comparative diagrams of the maximum radial deflection of work-pieces of the same size, as a function of the scaled distance, show that for the points with the same values of scaled distance, the maximum radial deformation caused by the under water explosive loading is 4 to 5 times more than the deflection of the shells under explosive forming, while using air. Results of this experimental research have also been compared with other studies which show that using water as the energy conveying media increases the efficiency up to 4.8 times. The effect of the media on failure modes of the shells, and the necking mechanism of the walls of the specimens, while being explosively loaded, are also discussed in this issue. Measuring the tested specimens shows that, the increase in the internal volume has been accompanied by necking of the walls, which finally results in the radial rupture of the structure.Keywords: Explosive Forming, Energy Conveying Medium, Tubular Shell
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1348768 Ranking of Inventory Policies Using Distance Based Approach Method
Authors: Gupta Amit, Kumar Ramesh, Tewari P. C.
Abstract:
Globalization is putting enormous pressure on the business organizations specially manufacturing one to rethink the supply chain in innovative manners. Inventory consumes major portion of total sale revenue. Effective and efficient inventory management plays a vital role for the successful functioning of any organization. Selection of inventory policy is one of the important purchasing activities. This paper focuses on selection and ranking of alternative inventory policies. A deterministic quantitative model based on Distance Based Approach (DBA) method has been developed for evaluation and ranking of inventory policies. We have employed this concept first time for this type of the selection problem. Four inventory policies economic order quantity (EOQ), just in time (JIT), vendor managed inventory (VMI) and monthly policy are considered. Improper selection could affect a company’s competitiveness in terms of the productivity of its facilities and quality of its products. The ranking of inventory policies is a multi-criteria problem. There is a need to first identify the selection criteria and then processes the information with reference to relative importance of attributes for comparison. Criteria values for each inventory policy can be obtained either analytically or by using a simulation technique or they are linguistic subjective judgments defined by fuzzy sets, like, for example, the values of criteria. A methodology is developed and applied to rank the inventory policies.
Keywords: Inventory Policy, Ranking, DBA, Selection criteria.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1827767 Optimization of the Characteristic Straight Line Method by a “Best Estimate“ of Observed, Normal Orthometric Elevation Differences
Authors: Mahmoud M. S. Albattah
Abstract:
In this paper, to optimize the “Characteristic Straight Line Method" which is used in the soil displacement analysis, a “best estimate" of the geodetic leveling observations has been achieved by taking in account the concept of 'Height systems'. This concept has been discussed in detail and consequently the concept of “height". In landslides dynamic analysis, the soil is considered as a mosaic of rigid blocks. The soil displacement has been monitored and analyzed by using the “Characteristic Straight Line Method". Its characteristic components have been defined constructed from a “best estimate" of the topometric observations. In the measurement of elevation differences, we have used the most modern leveling equipment available. Observational procedures have also been designed to provide the most effective method to acquire data. In addition systematic errors which cannot be sufficiently controlled by instrumentation or observational techniques are minimized by applying appropriate corrections to the observed data: the level collimation correction minimizes the error caused by nonhorizontality of the leveling instrument's line of sight for unequal sight lengths, the refraction correction is modeled to minimize the refraction error caused by temperature (density) variation of air strata, the rod temperature correction accounts for variation in the length of the leveling rod' s Invar/LO-VAR® strip which results from temperature changes, the rod scale correction ensures a uniform scale which conforms to the international length standard and the introduction of the concept of the 'Height systems' where all types of height (orthometric, dynamic, normal, gravity correction, and equipotential surface) have been investigated. The “Characteristic Straight Line Method" is slightly more convenient than the “Characteristic Circle Method". It permits to evaluate a displacement of very small magnitude even when the displacement is of an infinitesimal quantity. The inclination of the landslide is given by the inverse of the distance reference point O to the “Characteristic Straight Line". Its direction is given by the bearing of the normal directed from point O to the Characteristic Straight Line (Fig..6). A “best estimate" of the topometric observations was used to measure the elevation of points carefully selected, before and after the deformation. Gross errors have been eliminated by statistical analyses and by comparing the heights within local neighborhoods. The results of a test using an area where very interesting land surface deformation occurs are reported. Monitoring with different options and qualitative comparison of results based on a sufficient number of check points are presented.
Keywords: Characteristic straight line method, dynamic height, landslides, orthometric height, systematic errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1567766 Integrated Grey Rational Analysis-Standard Deviation Method for Handover in Heterogeneous Networks
Authors: Mohanad Alhabo, Naveed Nawaz, Mahmoud Al-Faris
Abstract:
The dense deployment of small cells is a promising solution to enhance the coverage and capacity of the heterogeneous networks (HetNets). However, the unplanned deployment could bring new challenges to the network ranging from interference, unnecessary handovers and handover failures. This will cause a degradation in the quality of service (QoS) delivered to the end user. In this paper, we propose an integrated Grey Rational Analysis Standard Deviation based handover method (GRA-SD) for HetNet. The proposed method integrates the Standard Deviation (SD) technique to acquire the weight of the handover metrics and the GRA method to select the best handover base station. The performance of the GRA-SD method is evaluated and compared with the traditional Multiple Attribute Decision Making (MADM) methods including Simple Additive Weighting (SAW) and VIKOR methods. Results reveal that the proposed method has outperformed the other methods in terms of minimizing the number of frequent unnecessary handovers and handover failures, in addition to improving the energy efficiency.Keywords: Energy efficiency, handover, HetNets, MADM, small cells.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 498