Search results for: Multi-objective decision making techniques
1146 The Relationship between Fugacity and Stress Intensity Factor for Corrosive Environment in Presence of Hydrogen Embrittlement
Authors: A. R. Shahani, E. Mahdavi, M. Amidpour
Abstract:
Hydrogen diffusion is the main problem for corrosion fatigue in corrosive environment. In order to analyze the phenomenon, it is needed to understand their behaviors specially the hydrogen behavior during the diffusion. So, Hydrogen embrittlement and prediction its behavior as a main corrosive part of the fractions, needed to solve combinations of different equations mathematically. The main point to obtain the equation, having knowledge about the source of causing diffusion and running the atoms into materials, called driving force. This is produced by either gradient of electrical or chemical potential. In this work, we consider the gradient of chemical potential to obtain the property equation. In diffusion of atoms, some of them may be trapped but, it could be ignorable in some conditions. According to the phenomenon of hydrogen embrittlement, the thermodynamic and chemical properties of hydrogen are considered to justify and relate them to fracture mechanics. It is very important to get a stress intensity factor by using fugacity as a property of hydrogen or other gases. Although, the diffusive behavior and embrittlement event are common and the same for other gases but, for making it more clear, we describe it for hydrogen. This considering on the definite gas and describing it helps us to understand better the importance of this relation.
Keywords: Hydrogen embrittlement, Fracture mechanics, Thermodynamic, Stress intensity factor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19601145 Fault Detection of Drinking Water Treatment Process Using PCA and Hotelling's T2 Chart
Authors: Joval P George, Dr. Zheng Chen, Philip Shaw
Abstract:
This paper deals with the application of Principal Component Analysis (PCA) and the Hotelling-s T2 Chart, using data collected from a drinking water treatment process. PCA is applied primarily for the dimensional reduction of the collected data. The Hotelling-s T2 control chart was used for the fault detection of the process. The data was taken from a United Utilities Multistage Water Treatment Works downloaded from an Integrated Program Management (IPM) dashboard system. The analysis of the results show that Multivariate Statistical Process Control (MSPC) techniques such as PCA, and control charts such as Hotelling-s T2, can be effectively applied for the early fault detection of continuous multivariable processes such as Drinking Water Treatment. The software package SIMCA-P was used to develop the MSPC models and Hotelling-s T2 Chart from the collected data.
Keywords: Principal component analysis, hotelling's t2 chart, multivariate statistical process control, drinking water treatment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27841144 Design of Encoding Calculator Software for Huffman and Shannon-Fano Algorithms
Authors: Wilson Chanhemo, Henry. R. Mgombelo, Omar F Hamad, T. Marwala
Abstract:
This paper presents a design of source encoding calculator software which applies the two famous algorithms in the field of information theory- the Shannon-Fano and the Huffman schemes. This design helps to easily realize the algorithms without going into a cumbersome, tedious and prone to error manual mechanism of encoding the signals during the transmission. The work describes the design of the software, how it works, comparison with related works, its efficiency, its usefulness in the field of information technology studies and the future prospects of the software to engineers, students, technicians and alike. The designed “Encodia" software has been developed, tested and found to meet the intended requirements. It is expected that this application will help students and teaching staff in their daily doing of information theory related tasks. The process is ongoing to modify this tool so that it can also be more intensely useful in research activities on source coding.Keywords: Coding techniques, Coding algorithms, Codingefficiency, Encodia, Encoding software.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34711143 Emotional Analysis for Text Search Queries on Internet
Authors: Gemma García López
Abstract:
The goal of this study is to analyze if search queries carried out in search engines such as Google, can offer emotional information about the user that performs them. Knowing the emotional state in which the Internet user is located can be a key to achieve the maximum personalization of content and the detection of worrying behaviors. For this, two studies were carried out using tools with advanced natural language processing techniques. The first study determines if a query can be classified as positive, negative or neutral, while the second study extracts emotional content from words and applies the categorical and dimensional models for the representation of emotions. In addition, we use search queries in Spanish and English to establish similarities and differences between two languages. The results revealed that text search queries performed by users on the Internet can be classified emotionally. This allows us to better understand the emotional state of the user at the time of the search, which could involve adapting the technology and personalizing the responses to different emotional states.Keywords: Emotion classification, text search queries, emotional analysis, sentiment analysis in text, natural language processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7131142 Microstructure and Mechanical Behaviuor of Rotary Friction Welded Titanium Alloys
Authors: M. Avinash, G. V. K. Chaitanya, Dhananjay Kumar Giri, Sarala Upadhya, B. K. Muralidhara
Abstract:
Ti-6Al-4V alloy has demonstrated a high strength to weight ratio as well as good properties at high temperature. The successful application of the alloy in some important areas depends on suitable joining techniques. Friction welding has many advantageous features to be chosen for joining Titanium alloys. The present work investigates the feasibility of producing similar metal joints of this Titanium alloy by rotary friction welding method. The joints are produced at three different speeds and the performances of the welded joints are evaluated by conducting microstructure studies, Vickers Hardness and tensile tests at the joints. It is found that the weld joints produced are sound and the ductile fractures in the tensile weld specimens occur at locations away from the welded joints. It is also found that a rotational speed of 1500 RPM can produce a very good weld, with other parameters kept constant.Keywords: Rotary friction weld, rotational speed, Ti-6Al-4V, weld structures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26041141 Music-Inspired Harmony Search Algorithm for Fixed Outline Non-Slicing VLSI Floorplanning
Authors: K. Sivasubramanian, K. B. Jayanthi
Abstract:
Floorplanning plays a vital role in the physical design process of Very Large Scale Integrated (VLSI) chips. It is an essential design step to estimate the chip area prior to the optimized placement of digital blocks and their interconnections. Since VLSI floorplanning is an NP-hard problem, many optimization techniques were adopted in the literature. In this work, a music-inspired Harmony Search (HS) algorithm is used for the fixed die outline constrained floorplanning, with the aim of reducing the total chip area. HS draws inspiration from the musical improvisation process of searching for a perfect state of harmony. Initially, B*-tree is used to generate the primary floorplan for the given rectangular hard modules and then HS algorithm is applied to obtain an optimal solution for the efficient floorplan. The experimental results of the HS algorithm are obtained for the MCNC benchmark circuits.Keywords: Floor planning, harmony search, non-slicing floorplan, very large scale integrated circuits.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19551140 Proxisch: An Optimization Approach of Large-Scale Unstable Proxy Servers Scheduling
Authors: Xiaoming Jiang, Jinqiao Shi, Qingfeng Tan, Wentao Zhang, Xuebin Wang, Muqian Chen
Abstract:
Nowadays, big companies such as Google, Microsoft, which have adequate proxy servers, have perfectly implemented their web crawlers for a certain website in parallel. But due to lack of expensive proxy servers, it is still a puzzle for researchers to crawl large amounts of information from a single website in parallel. In this case, it is a good choice for researchers to use free public proxy servers which are crawled from the Internet. In order to improve efficiency of web crawler, the following two issues should be considered primarily: (1) Tasks may fail owing to the instability of free proxy servers; (2) A proxy server will be blocked if it visits a single website frequently. In this paper, we propose Proxisch, an optimization approach of large-scale unstable proxy servers scheduling, which allow anyone with extremely low cost to run a web crawler efficiently. Proxisch is designed to work efficiently by making maximum use of reliable proxy servers. To solve second problem, it establishes a frequency control mechanism which can ensure the visiting frequency of any chosen proxy server below the website’s limit. The results show that our approach performs better than the other scheduling algorithms.Keywords: Proxy server, priority queue, optimization approach, distributed web crawling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28911139 Instant Location Detection of Objects Moving at High-Speedin C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev
Abstract:
The practical efficient approach is suggested to estimate the high-speed objects instant bounds in C-OTDR monitoring systems. In case of super-dynamic objects (trains, cars) is difficult to obtain the adequate estimate of the instantaneous object localization because of estimation lag. In other words, reliable estimation coordinates of monitored object requires taking some time for data observation collection by means of C-OTDR system, and only if the required sample volume will be collected the final decision could be issued. But it is contrary to requirements of many real applications. For example, in rail traffic management systems we need to get data of the dynamic objects localization in real time. The way to solve this problem is to use the set of statistical independent parameters of C-OTDR signals for obtaining the most reliable solution in real time. The parameters of this type we can call as «signaling parameters» (SP). There are several the SP’s which carry information about dynamic objects instant localization for each of COTDR channels. The problem is that some of these parameters are very sensitive to dynamics of seismoacoustic emission sources, but are non-stable. On the other hand, in case the SP is very stable it becomes insensitive as rule. This report contains describing of the method for SP’s co-processing which is designed to get the most effective dynamic objects localization estimates in the C-OTDR monitoring system framework.
Keywords: C-OTDR-system, co-processing of signaling parameters, high-speed objects localization, multichannel monitoring systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19061138 Induction Motor Speed Control Using Fuzzy Logic Controller
Authors: V. Chitra, R. S. Prabhakar
Abstract:
Because of the low maintenance and robustness induction motors have many applications in the industries. The speed control of induction motor is more important to achieve maximum torque and efficiency. Various speed control techniques like, Direct Torque Control, Sensorless Vector Control and Field Oriented Control are discussed in this paper. Soft computing technique – Fuzzy logic is applied in this paper for the speed control of induction motor to achieve maximum torque with minimum loss. The fuzzy logic controller is implemented using the Field Oriented Control technique as it provides better control of motor torque with high dynamic performance. The motor model is designed and membership functions are chosen according to the parameters of the motor model. The simulated design is tested using various tool boxes in MATLAB. The result concludes that the efficiency and reliability of the proposed speed controller is good.
Keywords: Induction motor, Field Oriented Control, Fuzzy logic controller, Maximum torque, Membership function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32321137 A Usability Testing Approach to Evaluate User-Interfaces in Business Administration
Authors: Salaheddin Odeh, Ibrahim O. Adwan
Abstract:
This interdisciplinary study is an investigation to evaluate user-interfaces in business administration. The study is going to be implemented on two computerized business administration systems with two distinctive user-interfaces, so that differences between the two systems can be determined. Both systems, a commercial and a prototype developed for the purpose of this study, deal with ordering of supplies, tendering procedures, issuing purchase orders, controlling the movement of the stocks against their actual balances on the shelves and editing them on their tabulations. In the second suggested system, modern computer graphics and multimedia issues were taken into consideration to cover the drawbacks of the first system. To highlight differences between the two investigated systems regarding some chosen standard quality criteria, the study employs various statistical techniques and methods to evaluate the users- interaction with both systems. The study variables are divided into two divisions: independent representing the interfaces of the two systems, and dependent embracing efficiency, effectiveness, satisfaction, error rate etc.
Keywords: Evaluation and usability testing, software prototyping, statistical methods, user-interface design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14641136 Posture Recognition using Combined Statistical and Geometrical Feature Vectors based on SVM
Authors: Omer Rashid, Ayoub Al-Hamadi, Axel Panning, Bernd Michaelis
Abstract:
It is hard to percept the interaction process with machines when visual information is not available. In this paper, we have addressed this issue to provide interaction through visual techniques. Posture recognition is done for American Sign Language to recognize static alphabets and numbers. 3D information is exploited to obtain segmentation of hands and face using normal Gaussian distribution and depth information. Features for posture recognition are computed using statistical and geometrical properties which are translation, rotation and scale invariant. Hu-Moment as statistical features and; circularity and rectangularity as geometrical features are incorporated to build the feature vectors. These feature vectors are used to train SVM for classification that recognizes static alphabets and numbers. For the alphabets, curvature analysis is carried out to reduce the misclassifications. The experimental results show that proposed system recognizes posture symbols by achieving recognition rate of 98.65% and 98.6% for ASL alphabets and numbers respectively.Keywords: Feature Extraction, Posture Recognition, Pattern Recognition, Application.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15181135 Automatic Enhanced Update Summary Generation System for News Documents
Authors: S. V. Kogilavani, C. S. Kanimozhiselvi, S. Malliga
Abstract:
Fast changing knowledge systems on the Internet can be accessed more efficiently with the help of automatic document summarization and updating techniques. The aim of multi-document update summary generation is to construct a summary unfolding the mainstream of data from a collection of documents based on the hypothesis that the user has already read a set of previous documents. In order to provide a lot of semantic information from the documents, deeper linguistic or semantic analysis of the source documents were used instead of relying only on document word frequencies to select important concepts. In order to produce a responsive summary, meaning oriented structural analysis is needed. To address this issue, the proposed system presents a document summarization approach based on sentence annotation with aspects, prepositions and named entities. Semantic element extraction strategy is used to select important concepts from documents which are used to generate enhanced semantic summary.
Keywords: Aspects, named entities, prepositions, update summary.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21331134 Evolutionary Computation Technique for Solving Riccati Differential Equation of Arbitrary Order
Authors: Raja Muhammad Asif Zahoor, Junaid Ali Khan, I. M. Qureshi
Abstract:
In this article an evolutionary technique has been used for the solution of nonlinear Riccati differential equations of fractional order. In this method, genetic algorithm is used as a tool for the competent global search method hybridized with active-set algorithm for efficient local search. The proposed method has been successfully applied to solve the different forms of Riccati differential equations. The strength of proposed method has in its equal applicability for the integer order case, as well as, fractional order case. Comparison of the method has been made with standard numerical techniques as well as the analytic solutions. It is found that the designed method can provide the solution to the equation with better accuracy than its counterpart deterministic approaches. Another advantage of the given approach is to provide results on entire finite continuous domain unlike other numerical methods which provide solutions only on discrete grid of points.Keywords: Riccati Equation, Non linear ODE, Fractional differential equation, Genetic algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19411133 Photomechanical Analysis of Wooden Testing Bodies under Flexural Loadings
Authors: J. Gazzola, I. M. Dal Fabbro, J. Soriano, M. V. G. Silva, S. Rodrigues
Abstract:
Application of wood in rural construction is diffused all around the world since remote times. However, its inclusion in structural design deserves strong support from broad knowledge of material properties. The pertinent literature reveals the application of optical methods in determining the complete field displacement on bodies exhibiting regular as well as irregular surfaces. The use of moiré techniques in experimental mechanics consists in analyzing the patterns generated on the body surface before and after deformation. The objective of this research work is to study the qualitative deformation behavior of wooden testing specimens under specific loading situations. The experiment setup follows the literature description of shadow moiré methods. Results indicate strong anisotropy influence of the generated displacement field. Important qualitative as well as quantitative stress and strain distribution were obtained wooden members which are applicable to rural constructions.Keywords: Moiré methods, wooden structural material, rural constructions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15271132 Generating Class-Based Test Cases for Interface Classes of Object-Oriented Black Box Frameworks
Authors: Jehad Al Dallal, Paul Sorenson
Abstract:
An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define the Framework Interface Classes (FICs) and their possible specifications, which helps in building reusable test cases for the implementations of these classes. This paper introduces a novel technique called all paths-state to generate state-based test cases to test the FICs at class level. The technique is experimentally evaluated. The empirical evaluation shows that all paths-state technique produces test cases with a high degree of coverage for the specifications of the implemented FICs comparing to test cases generated using round-trip path and all-transition techniques.Keywords: Hooks, object-oriented framework, frameworkinterface classes (FICs), specification-based testing, test casegeneration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13261131 Applying Multiple Intelligences to Teach Buddhist Doctrines in a Classroom
Authors: Phalaunnaphat Siriwongs
Abstract:
The classroom of the 21st century is an ever changing forum for new and innovative thoughts and ideas. With increasing technology and opportunity, students have rapid access to information that only decades ago would have taken weeks to obtain. Unfortunately, new techniques and technology are not the cure for the fundamental problems that have plagued the classroom ever since education was established. Class size has been an issue long debated in academia. While it is difficult to pin point an exact number, it is clear that in this case more does not mean better. By looking into the success and pitfalls of classroom size the true advantages of smaller classes will become clear. Previously, one class was comprised of 50 students. Being seventeen and eighteen- year- old students, sometimes it was quite difficult for them to stay focused. To help them understand and gain much knowledge, a researcher introduced “The Theory of Multiple Intelligence” and this, in fact, enabled students to learn according to their own learning preferences no matter how they were being taught. In this lesson, the researcher designed a cycle of learning activities involving all intelligences so that everyone had equal opportunities to learn.
Keywords: Multiple intelligences, role play, performance assessment, formative assessment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15401130 Fingerprint Verification System Using Minutiae Extraction Technique
Authors: Manvjeet Kaur, Mukhwinder Singh, Akshay Girdhar, Parvinder S. Sandhu
Abstract:
Most fingerprint recognition techniques are based on minutiae matching and have been well studied. However, this technology still suffers from problems associated with the handling of poor quality impressions. One problem besetting fingerprint matching is distortion. Distortion changes both geometric position and orientation, and leads to difficulties in establishing a match among multiple impressions acquired from the same finger tip. Marking all the minutiae accurately as well as rejecting false minutiae is another issue still under research. Our work has combined many methods to build a minutia extractor and a minutia matcher. The combination of multiple methods comes from a wide investigation into research papers. Also some novel changes like segmentation using Morphological operations, improved thinning, false minutiae removal methods, minutia marking with special considering the triple branch counting, minutia unification by decomposing a branch into three terminations, and matching in the unified x-y coordinate system after a two-step transformation are used in the work.
Keywords: Biometrics, Minutiae, Crossing number, False Accept Rate (FAR), False Reject Rate (FRR).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 65721129 Limits Problem Solving in Engineering Careers: Competences and Errors
Authors: Veronica Diaz Quezada
Abstract:
In this article, the performance and errors are featured and analysed in the limit problems solving of a real-valued function, in correspondence to competency-based education in engineering careers, in the south of Chile. The methodological component is contextualised in a qualitative research, with a descriptive and explorative design, with elaboration, content validation and application of quantitative instruments, consisting of two parallel forms of open answer tests, based on limit application problems. The mathematical competences and errors made by students from five engineering careers from a public University are identified and characterized. Results show better performance only to solve routine-context problem-solving competence, thus they are oriented towards a rational solution or they use a suitable problem-solving method, achieving the correct solution. Regarding errors, most of them are related to techniques and the incorrect use of theorems and definitions of real-valued function limits of real variable.
Keywords: Engineering education, errors, limits, mathematics competences, problem solving.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13171128 Range-Free Localization Schemes for Wireless Sensor Networks
Authors: R. Khadim, M. Erritali, A. Maaden
Abstract:
Localization of nodes is one of the key issues of Wireless Sensor Network (WSN) that gained a wide attention in recent years. The existing localization techniques can be generally categorized into two types: range-based and range-free. Compared with rang-based schemes, the range-free schemes are more costeffective, because no additional ranging devices are needed. As a result, we focus our research on the range-free schemes. In this paper we study three types of range-free location algorithms to compare the localization error and energy consumption of each one. Centroid algorithm requires a normal node has at least three neighbor anchors, while DV-hop algorithm doesn’t have this requirement. The third studied algorithm is the amorphous algorithm similar to DV-Hop algorithm, and the idea is to calculate the hop distance between two nodes instead of the linear distance between them. The simulation results show that the localization accuracy of the amorphous algorithm is higher than that of other algorithms and the energy consumption does not increase too much.Keywords: Wireless Sensor Networks, Node Localization, Centroid Algorithm, DV–Hop Algorithm, Amorphous Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26301127 Lower Order Harmonics Minimisation in CHB Inverter Using GA and Decomposition by WT
Authors: V. Joshi Manohar, P. Sujatha, K. S. R. Anjaneyulu
Abstract:
Nowadays Multilevel inverters are widely using in various applications. Modulation strategy at fundamental switching frequency like, SHEPWM is prominent technique to eliminate lower order of harmonics with less switching losses and better harmonic profile. The equations which are formed by SHE are highly nonlinear transcendental in nature, there may exist single, multiple or even no solutions for a particular MI. However, some loads such as electrical drives, it is required to operate in whole range of MI. In order to solve SHE equations for whole range of MI, intelligent techniques are well suited to solve equations so as to produce lest %THDV. Hence, this paper uses Continuous genetic algorithm for minimising harmonics. This paper also presents wavelet based analysis of harmonics. The developed algorithm is simulated and %THD from FFT analysis and Wavelet analysis are compared. MATLAB programming environment and SIMULINK models are used whenever necessary.
Keywords: Cascade H-Bridge Inverter (CHB), Continuous Genetic Algorithm (C-GA), Selective Harmonic Elimination Pulse Width Modulation (SHEPWM), Total Harmonic Distortion (%THDv), Wavelet Transform (WT).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29141126 Methodology for Bioenergy Potential and Assessment for Energy Deployment in Rural Vhembe District Areas
Authors: Clement M. Matasane, Mohamed T. Kahn
Abstract:
Biomass resources such as animal waste, agricultural and acro-industrial residues, forestry and woodland waste, and industrial and municipal solid wastes provide alternative means to utilize its untapped potential for biomass/biofuel renewable energy systems. In addition, crop residues (i.e., grain, starch, and energy crops) are commonly available in the district and play an essential role in community farming activities. The remote sensing technology (mappings) and geographic information systems tool will be used to determine the biomass potential in the Vhembe District Municipality. The detailed assessment, estimation, and modeling in quantifying their distribution, abundance, and quality yield an effective and efficient use of their potential. This paper aims to examine the potential and prospects of deploying bioenergy systems in small or micro-systems in the district for community use and applications. This deployment of the biofuels/biomass systems will help communities for sustainable energy supply from their traditional energy use into innovative and suitable methods that improve their livelihood. The study demonstrates the potential applications of Geographical Information Systems (GIS) in spatial mapping analysis, evaluation, modeling, and decision support for easy access to renewable energy systems.
Keywords: Agricultural crops, waste materials, biomass potentials, bioenergy potentials, GIS mappings, environmental data, renewable energy deployment, sustainable energy supply.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3311125 A File Splitting Technique for Reducing the Entropy of Text Files
Authors: Abdel-Rahman M. Jaradat, , Mansour I. Irshid, Talha T. Nassar
Abstract:
A novel file splitting technique for the reduction of the nth-order entropy of text files is proposed. The technique is based on mapping the original text file into a non-ASCII binary file using a new codeword assignment method and then the resulting binary file is split into several subfiles each contains one or more bits from each codeword of the mapped binary file. The statistical properties of the subfiles are studied and it is found that they reflect the statistical properties of the original text file which is not the case when the ASCII code is used as a mapper. The nth-order entropy of these subfiles are determined and it is found that the sum of their entropies is less than that of the original text file for the same values of extensions. These interesting statistical properties of the resulting subfiles can be used to achieve better compression ratios when conventional compression techniques are applied to these subfiles individually and on a bit-wise basis rather than on character-wise basis.
Keywords: Bit-wise compression, entropy, file splitting, source mapping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14421124 Adoption of Appropriate and Cost Effective Technologies in Housing: Indian Experience
Authors: A. K. Jain, M. C. Paliwal
Abstract:
Construction cost in India is increasing at around 50 per cent over the average inflation levels. It have registered increase of up to 15 per cent every year, primarily due to cost of basic building materials such as steel, cement, bricks, timber and other inputs as well as cost of labour. As a result, the cost of construction using conventional building materials and construction is becoming beyond the affordable limits particularly for low-income groups of population as well as a large cross section of the middle - income groups. Therefore, there is a need to adopt cost-effective construction methods either by up-gradation of traditional technologies using local resources or applying modern construction materials and techniques with efficient inputs leading to economic solutions. This has become the most relevant aspect in the context of the large volume of housing to be constructed in both rural and urban areas and the consideration of limitations in the availability of resources such as building materials and finance. This paper makes an overview of the housing status in India and adoption of appropriate and cost effective technologies in the country.Keywords: Appropriate, Cost Effective, Ekra, Five year plan, Poverty
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 49731123 Bottom Up Text Mining through Hierarchical Document Representation
Authors: Y. Djouadi., F. Souam.
Abstract:
Most of the existing text mining approaches are proposed, keeping in mind, transaction databases model. Thus, the mined dataset is structured using just one concept: the “transaction", whereas the whole dataset is modeled using the “set" abstract type. In such cases, the structure of the whole dataset and the relationships among the transactions themselves are not modeled and consequently, not considered in the mining process. We believe that taking into account structure properties of hierarchically structured information (e.g. textual document, etc ...) in the mining process, can leads to best results. For this purpose, an hierarchical associations rule mining approach for textual documents is proposed in this paper and the classical set-oriented mining approach is reconsidered profits to a Direct Acyclic Graph (DAG) oriented approach. Natural languages processing techniques are used in order to obtain the DAG structure. Based on this graph model, an hierarchical bottom up algorithm is proposed. The main idea is that each node is mined with its parent node.Keywords: Graph based association rules mining, Hierarchical document structure, Text mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20561122 Automatically Driven Vector for Guidewire Segmentation in 2D and Biplane Fluoroscopy
Authors: Simon Lessard, Pascal Bigras, Caroline Lau, Daniel Roy, Gilles Soulez, Jacques A. de Guise
Abstract:
The segmentation of endovascular tools in fluoroscopy images can be accurately performed automatically or by minimum user intervention, using known modern techniques. It has been proven in literature, but no clinical implementation exists so far because the computational time requirements of such technology have not yet been met. A classical segmentation scheme is composed of edge enhancement filtering, line detection, and segmentation. A new method is presented that consists of a vector that propagates in the image to track an edge as it advances. The filtering is performed progressively in the projected path of the vector, whose orientation allows for oriented edge detection, and a minimal image area is globally filtered. Such an algorithm is rapidly computed and can be implemented in real-time applications. It was tested on medical fluoroscopy images from an endovascular cerebral intervention. Ex- periments showed that the 2D tracking was limited to guidewires without intersection crosspoints, while the 3D implementation was able to cope with such planar difficulties.
Keywords: Edge detection, Line Enhancement, Segmentation, Fluoroscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17271121 Adaptive Score Normalization: A Novel Approach for Multimodal Biometric Systems
Authors: Anouar Ben Khalifa, Sami Gazzah, Najoua Essoukri BenAmara
Abstract:
Multimodal biometric systems integrate the data presented by multiple biometric sources, hence offering a better performance than the systems based on a single biometric modality. Although the coupling of biometric systems can be done at different levels, the fusion at the scores level is the most common since it has been proven effective than the rest of the fusion levels. However, the scores from different modalities are generally heterogeneous. A step of normalizing the scores is needed to transform these scores into a common domain before combining them. In this paper, we study the performance of several normalization techniques with various fusion methods in a context relating to the merger of three unimodal systems based on the face, the palmprint and the fingerprint. We also propose a new adaptive normalization method that takes into account the distribution of client scores and impostor scores. Experiments conducted on a database of 100 people show that the performances of a multimodal system depend on the choice of the normalization method and the fusion technique. The proposed normalization method has given the best results.
Keywords: Multibiometrics, Fusion, Score level, Score normalization, Adaptive normalization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35531120 A Neural-Network-Based Fault Diagnosis Approach for Analog Circuits by Using Wavelet Transformation and Fractal Dimension as a Preprocessor
Abstract:
This paper presents a new method of analog fault diagnosis based on back-propagation neural networks (BPNNs) using wavelet decomposition and fractal dimension as preprocessors. The proposed method has the capability to detect and identify faulty components in an analog electronic circuit with tolerance by analyzing its impulse response. Using wavelet decomposition to preprocess the impulse response drastically de-noises the inputs to the neural network. The second preprocessing by fractal dimension can extract unique features, which are the fed to a neural network as inputs for further classification. A comparison of our work with [1] and [6], which also employs back-propagation (BP) neural networks, reveals that our system requires a much smaller network and performs significantly better in fault diagnosis of analog circuits due to our proposed preprocessing techniques.
Keywords: Analog circuits, fault diagnosis, tolerance, wavelettransform, fractal dimension, box dimension.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21981119 Material Parameter Identification of Modified AbdelKarim-Ohno Model
Authors: M. Cermak, T. Karasek, J. Rojicek
Abstract:
The key role in phenomenological modelling of cyclic plasticity is good understanding of stress-strain behaviour of given material. There are many models describing behaviour of materials using numerous parameters and constants. Combination of individual parameters in those material models significantly determines whether observed and predicted results are in compliance. Parameter identification techniques such as random gradient, genetic algorithm and sensitivity analysis are used for identification of parameters using numerical modelling and simulation. In this paper genetic algorithm and sensitivity analysis are used to study effect of 4 parameters of modified AbdelKarim-Ohno cyclic plasticity model. Results predicted by Finite Element (FE) simulation are compared with experimental data from biaxial ratcheting test with semi-elliptical loading path.
Keywords: Genetic algorithm, sensitivity analysis, inverse approach, finite element method, cyclic plasticity, ratcheting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23721118 The Link between Money Market and Economic Growth in Nigeria: Vector Error Correction Model Approach
Authors: Ehigiamusoe, Uyi Kizito
Abstract:
The paper examines the impact of money market on economic growth in Nigeria using data for the period 1980-2012. Econometrics techniques such as Ordinary Least Squares Method, Johanson’s Co-integration Test and Vector Error Correction Model were used to examine both the long-run and short-run relationship. Evidence from the study suggest that though a long-run relationship exists between money market and economic growth, but the present state of the Nigerian money market is significantly and negatively related to economic growth. The link between the money market and the real sector of the economy remains very weak. This implies that the market is not yet developed enough to produce the needed growth that will propel the Nigerian economy because of several challenges. It was therefore recommended that government should create the appropriate macroeconomic policies, legal framework and sustain the present reforms with a view to developing the market so as to promote productive activities, investments, and ultimately economic growth.
Keywords: Economic Growth, Investments, Money Market, Money Market Challenges, Money Market Instruments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 84961117 Solution Economic Power Dispatch Problems by an Ant Colony Optimization Approach
Authors: Navid Mehdizadeh Afroozi, Khodakhast Isapour, Mojtaba Hakimzadeh, Abdolmohammad Davodi
Abstract:
The objective of the Economic Dispatch(ED) Problems of electric power generation is to schedule the committed generating units outputs so as to meet the required load demand at minimum operating cost while satisfying all units and system equality and inequality constraints. This paper presents a new method of ED problems utilizing the Max-Min Ant System Optimization. Historically, traditional optimizations techniques have been used, such as linear and non-linear programming, but within the past decade the focus has shifted on the utilization of Evolutionary Algorithms, as an example Genetic Algorithms, Simulated Annealing and recently Ant Colony Optimization (ACO). In this paper we introduce the Max-Min Ant System based version of the Ant System. This algorithm encourages local searching around the best solution found in each iteration. To show its efficiency and effectiveness, the proposed Max-Min Ant System is applied to sample ED problems composed of 4 generators. Comparison to conventional genetic algorithms is presented.
Keywords: Economic Dispatch (ED), Ant Colony Optimization, Fuel Cost, Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2580