Search results for: Unstructured TextDocuments
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 82

Search results for: Unstructured TextDocuments

52 Robust Control Synthesis for an Unmanned Underwater Vehicle

Authors: A. Budiyono

Abstract:

The control design for unmanned underwater vehicles (UUVs) is challenging due to the uncertainties in the complex dynamic modeling of the vehicle as well as its unstructured operational environment. To cope with these difficulties, a practical robust control is therefore desirable. The paper deals with the application of coefficient diagram method (CDM) for a robust control design of an autonomous underwater vehicle. The CDM is an algebraic approach in which the characteristic polynomial and the controller are synthesized simultaneously. Particularly, a coefficient diagram (comparable to Bode diagram) is used effectively to convey pertinent design information and as a measure of trade-off between stability, response speed and robustness. In the polynomial ring, Kharitonov polynomials are employed to analyze the robustness of the controller due to parametric uncertainties.

Keywords: coefficient diagram method, robust control, Kharitonov polynomials, unmanned underwater vehicles.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2043
51 A New Precautionary Method for Measurement and Improvement the Data Quality

Authors: Seyed Mohammad Hossein Moossavizadeh, Mehran Mohsenzadeh, Nasrin Arshadi

Abstract:

the data quality is a kind of complex and unstructured concept, which is concerned by information systems managers. The reason of this attention is the high amount of Expenses for maintenance and cleaning of the inefficient data. Such a data more than its expenses of lack of quality, cause wrong statistics, analysis and decisions in organizations. Therefor the managers intend to improve the quality of their information systems' data. One of the basic subjects of quality improvement is the evaluation of the amount of it. In this paper, we present a precautionary method, which with its application the data of information systems would have a better quality. Our method would cover different dimensions of data quality; therefor it has necessary integrity. The presented method has tested on three dimensions of accuracy, value-added and believability and the results confirm the improvement and integrity of this method.

Keywords: Data quality, precaution, information system, measurement, improvement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1428
50 A Framework for Vacant City-Owned Land to Be Utilised for Urban Agriculture: The Case of Cape Town, South Africa

Authors: P. S. Van Staden, M. M. Campbell

Abstract:

Vacant City of Cape Town-owned land lying unutilized and -productive could be developed for land uses such as urban agriculture that may improve the livelihoods of low income families. The new City of Cape Town zoning scheme includes an Urban Agriculture zoning for the first time. Unstructured qualitative interviews among town planners revealed their optimism about this inclusion as it will provide low-income residents with opportunities to generate an income. An existing farming community at Philippi, located within the municipal boundary of the city, was approached and empirical data obtained through questionnaires provided proof that urban agriculture could be viable in a coastal metropolitan city such as Cape Town even if farmers only produce for their own households. The lease method proposed for urban agriculture is a usufruct agreement conferring the right to another party, other than the legal owner, to enjoy the use and advantages of the property.

Keywords: Land uses, urban agriculture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1957
49 Information Extraction from Unstructured and Ungrammatical Data Sources for Semantic Annotation

Authors: Quratulain N. Rajput, Sajjad Haider, Nasir Touheed

Abstract:

The internet has become an attractive avenue for global e-business, e-learning, knowledge sharing, etc. Due to continuous increase in the volume of web content, it is not practically possible for a user to extract information by browsing and integrating data from a huge amount of web sources retrieved by the existing search engines. The semantic web technology enables advancement in information extraction by providing a suite of tools to integrate data from different sources. To take full advantage of semantic web, it is necessary to annotate existing web pages into semantic web pages. This research develops a tool, named OWIE (Ontology-based Web Information Extraction), for semantic web annotation using domain specific ontologies. The tool automatically extracts information from html pages with the help of pre-defined ontologies and gives them semantic representation. Two case studies have been conducted to analyze the accuracy of OWIE.

Keywords: Ontology, Semantic Annotation, Wrapper, Information Extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2070
48 A Multi-Phase Methodology for Investigating Localisation Policies within the GCC: The Hotel Industry in the KSA and the UAE

Authors: Areej Azhar, Peter Duncan, David Edgar

Abstract:

Due to a high unemployment rate among local people and a high reliance on expatriate workers, the governments in the Gulf Co-operation Council (GCC) countries have been implementing programmes of localisation (replacing foreign workers with GCC nationals). These programmes have been successful in the public sector but much less so in the private sector. However, there are now insufficient jobs for locals in the public sector and the onus to provide employment has fallen on the private sector. This paper is concerned with a study, which is a work in progress (certain elements are complete but not the whole study), investigating the effective implementation of localisation policies in four- and five-star hotels in the Kingdom of Saudi Arabia (KSA) and the United Arab Emirates (UAE). The purpose of the paper is to identify the research gap, and to present the need for the research. Further, it will explain how this research was conducted. Studies of localisation in the GCC countries are under-represented in scholarly literature. Currently, the hotel sectors in KSA and UAE play an important part in the countries’ economies. However, the total proportion of Saudis working in the hotel sector in KSA is slightly under 8%, and in the UAE, the hotel sector remains highly reliant on expatriates. There is therefore a need for research on strategies to enhance the implementation of the localisation policies in general and in the hotel sector in particular. Further, despite the importance of the hotel sector to their economies, there remains a dearth of research into the implementation of localisation policies in this sector. Indeed, as far as the researchers are aware, there is no study examining localisation in the hotel sector in KSA, and few in the UAE. This represents a considerable research gap. Regarding how the research was carried out, a multiple case study strategy was used. The four- and five-star hotel sector in KSA is one of the cases, while the four- and five-star hotel sector in the UAE is the other case. Four- and five-star hotels in KSA and the UAE were chosen as these countries have the longest established localisation policies of all the GCC states and there are more hotels of these classifications in these countries than in any of the other Gulf countries. A literature review was carried out to underpin the research. The empirical data were gathered in three phases. In order to gain a pre-understanding of the issues pertaining to the research context, Phase I involved eight unstructured interviews with officials from the Saudi Commission for Tourism and Antiquities (three interviewees); the Saudi Human Resources Development Fund (one); the Abu Dhabi Tourism and Culture Authority (three); and the Abu Dhabi Development Fund (one).

In Phase II, a questionnaire was administered to 24 managers and 24 employees in four- and five-star hotels in each country to obtain their beliefs, attitudes, opinions, preferences and practices concerning localisation. Unstructured interviews were carried out in Phase III with six managers in each country in order to allow them to express opinions that may not have been explored in sufficient depth in the questionnaire. The interviews in Phases I and III were analysed using thematic analysis and SPSS will be used to analyse the questionnaire data. It is recommended that future research be undertaken on a larger scale, with a larger sample taken from all over KSA and the UAE rather than from only four cities (i.e., Riyadh and Jeddah in KSA and Abu Dhabi and Sharjah in the UAE), as was the case in this research.

Keywords: KSA, UAE, localisation, hotels, Human Resource Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2526
47 Mining and Visual Management of XML-Based Image Collections

Authors: Khalil Shihab, Nida Al-Chalabi

Abstract:

This article describes Uruk, the virtual museum of Iraq that we developed for visual exploration and retrieval of image collections. The system largely exploits the loosely-structured hierarchy of XML documents that provides a useful representation method to store semi-structured or unstructured data, which does not easily fit into existing database. The system offers users the capability to mine and manage the XML-based image collections through a web-based Graphical User Interface (GUI). Typically, at an interactive session with the system, the user can browse a visual structural summary of the XML database in order to select interesting elements. Using this intermediate result, queries combining structure and textual references can be composed and presented to the system. After query evaluation, the full set of answers is presented in a visual and structured way.

Keywords: Data-centric XML, graphical user interfaces, information retrieval, case-based reasoning, fuzzy sets

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1743
46 Signed Approach for Mining Web Content Outliers

Authors: G. Poonkuzhali, K.Thiagarajan, K.Sarukesi, G.V.Uma

Abstract:

The emergence of the Internet has brewed the revolution of information storage and retrieval. As most of the data in the web is unstructured, and contains a mix of text, video, audio etc, there is a need to mine information to cater to the specific needs of the users without loss of important hidden information. Thus developing user friendly and automated tools for providing relevant information quickly becomes a major challenge in web mining research. Most of the existing web mining algorithms have concentrated on finding frequent patterns while neglecting the less frequent ones that are likely to contain outlying data such as noise, irrelevant and redundant data. This paper mainly focuses on Signed approach and full word matching on the organized domain dictionary for mining web content outliers. This Signed approach gives the relevant web documents as well as outlying web documents. As the dictionary is organized based on the number of characters in a word, searching and retrieval of documents takes less time and less space.

Keywords: Outliers, Relevant document, , Signed Approach, Web content mining, Web documents..

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2309
45 Evolving Knowledge Extraction from Online Resources

Authors: Zhibo Xiao, Tharini Nayanika de Silva, Kezhi Mao

Abstract:

In this paper, we present an evolving knowledge extraction system named AKEOS (Automatic Knowledge Extraction from Online Sources). AKEOS consists of two modules, including a one-time learning module and an evolving learning module. The one-time learning module takes in user input query, and automatically harvests knowledge from online unstructured resources in an unsupervised way. The output of the one-time learning is a structured vector representing the harvested knowledge. The evolving learning module automatically schedules and performs repeated one-time learning to extract the newest information and track the development of an event. In addition, the evolving learning module summarizes the knowledge learned at different time points to produce a final knowledge vector about the event. With the evolving learning, we are able to visualize the key information of the event, discover the trends, and track the development of an event.

Keywords: Evolving learning, knowledge extraction, knowledge graph, text mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 890
44 Development of Improved Three Dimensional Unstructured Tetrahedral Mesh Generator

Authors: Ng Yee Luon, Mohd Zamri Yusoff, Norshah Hafeez Shuaib

Abstract:

Meshing is the process of discretizing problem domain into many sub domains before the numerical calculation can be performed. One of the most popular meshes among many types of meshes is tetrahedral mesh, due to their flexibility to fit into almost any domain shape. In both 2D and 3D domains, triangular and tetrahedral meshes can be generated by using Delaunay triangulation. The quality of mesh is an important factor in performing any Computational Fluid Dynamics (CFD) simulations as the results is highly affected by the mesh quality. Many efforts had been done in order to improve the quality of the mesh. The paper describes a mesh generation routine which has been developed capable of generating high quality tetrahedral cells in arbitrary complex geometry. A few test cases in CFD problems are used for testing the mesh generator. The result of the mesh is compared with the one generated by a commercial software. The results show that no sliver exists for the meshes generated, and the overall quality is acceptable since the percentage of the bad tetrahedral is relatively small. The boundary recovery was also successfully done where all the missing faces are rebuilt.

Keywords: Mesh generation, tetrahedral, CFD, Delaunay.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1468
43 The Emerging Central Business District (CBD) in Lafia Town, Nigeria, and its Related Urban Planning Problems

Authors: Barau Daniel, Bashayi Obadiah

Abstract:

A spatial analysis of a large 20th century urban settlement (town/city) easily presents the celebrated central Business District (CBD). Theories of Urban Land Economics have easily justified and attempted to explain the existence of such a district activity area within the cityscape. This work examines the gradual emergence and development of the CBD in Lafia Town, Nigeria over 20 years and the attended urban problems caused by its emergence. Personal knowledge and observation of land use change are the main sources of data for the work, with unstructured interview with residents. The result are that the absence of a co-ordinate land use plan for the town, multi-nuclei nature, and regional location of surrounding towns have affected the growth pattern, hence the CBD. Traffic congestion, dispersed CBD land uses are some of the urban planning problems. The work concludes by advocating for integrating CBD uses.

Keywords: Urban planning, Central Business District (CBD), downtown.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4015
42 Process Oriented Architecture for Emergency Scenarios in the Czech Republic

Authors: Tomáš Ludík, Josef Navrátil, Alena Langerová

Abstract:

Tackling emergency situations is performed based on emergency scenarios. These scenarios do not have a uniform form in the Czech Republic. They are unstructured and developed primarily in the text form. This does not allow solving emergency situations efficiently. For this reason, the paper aims at defining a Process Oriented Architecture to support and thus to improve tackling emergency situations in the Czech Republic. The innovative Process Oriented Architecture is based on the Workflow Reference Model while taking into account the options of Business Process Management Suites for the implementation of process oriented emergency scenarios. To verify the proposed architecture the Proof of Concept has been used which covers the reception of an emergency event at the district emergency operations centre. Within the particular implementation of the proposed architecture the Bonita Open Solution has been used. The architecture created in this way is suitable not only for emergency management, but also for educational purposes.

Keywords: Business Process Management Suite, Czech Republic, Emergency Scenarios, Process Execution, Process Oriented Architecture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1772
41 Investigation of the Effect of Grid Size on External Store Separation Trajectory Using CFD

Authors: Alaa A. Osman, Amgad M. Bayoumy, Ismail El baialy, Osama E. Abdellatif, Essam E. Khallil

Abstract:

In this paper, a numerical simulation of a finned store separating from a wing-pylon configuration has been studied and validated. A dynamic unstructured tetrahedral mesh approach is accomplished by using three grid sizes to numerically solving the discretized three dimensional, inviscid and compressible Euler equations. The method used for computations of separation of an external store assuming quasi-steady flow condition. Computations of quasi-steady flow have been directly coupled to a six degree-offreedom (6DOF) rigid-body motion code to generate store trajectories. The pressure coefficients at four different angular cuts and time histories of various trajectory parameters and wing pressure distribution during the store separation are compared for every grid size with published experimental data.

Keywords: CFD Modelling, Quasi-steady Flow, Moving-body Trajectories, Transonic Store Separation, Moving-body Trajectories.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2936
40 Developing an Audit Quality Model for an Emerging Market

Authors: Bita Mashayekhi, Azadeh Maddahi, Arash Tahriri

Abstract:

The purpose of this paper is developing a model for audit quality, with regard to the contextual and environmental attributes of the audit profession in Iran. For this purpose, using an exploratory approach, and because of the special attributes of the auditing profession in Iran in terms of the legal environment, regulatory and supervisory mechanisms, audit firms size, and etc., we used grounded theory approach as a qualitative research method. Therefore, we got the opinions of the experts in the auditing and capital market areas through unstructured interviews. As a result, the authors revealed the determinants of audit quality, and by using these determinants, developed an Integrated Audit Quality Model, including causal conditions, intervening conditions, context, as well as action strategies related to AQ and their consequences. In this research, audit quality is studied using a systemic approach. According to this approach, the quality of inputs, processes, and outputs of auditing determines the quality of auditing, therefore, the quality of all different parts of this system is considered.

Keywords: Audit quality, integrated audit quality model, audit supply, demand for audit service, grounded theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1237
39 Text Mining of Twitter Data Using a Latent Dirichlet Allocation Topic Model and Sentiment Analysis

Authors: Sidi Yang, Haiyi Zhang

Abstract:

Twitter is a microblogging platform, where millions of users daily share their attitudes, views, and opinions. Using a probabilistic Latent Dirichlet Allocation (LDA) topic model to discern the most popular topics in the Twitter data is an effective way to analyze a large set of tweets to find a set of topics in a computationally efficient manner. Sentiment analysis provides an effective method to show the emotions and sentiments found in each tweet and an efficient way to summarize the results in a manner that is clearly understood. The primary goal of this paper is to explore text mining, extract and analyze useful information from unstructured text using two approaches: LDA topic modelling and sentiment analysis by examining Twitter plain text data in English. These two methods allow people to dig data more effectively and efficiently. LDA topic model and sentiment analysis can also be applied to provide insight views in business and scientific fields.

Keywords: Text mining, Twitter, topic model, sentiment analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1734
38 Unsupervised Text Mining Approach to Early Warning System

Authors: Ichihan Tai, Bill Olson, Paul Blessner

Abstract:

Traditional early warning systems that alarm against crisis are generally based on structured or numerical data; therefore, a system that can make predictions based on unstructured textual data, an uncorrelated data source, is a great complement to the traditional early warning systems. The Chicago Board Options Exchange (CBOE) Volatility Index (VIX), commonly referred to as the fear index, measures the cost of insurance against market crash, and spikes in the event of crisis. In this study, news data is consumed for prediction of whether there will be a market-wide crisis by predicting the movement of the fear index, and the historical references to similar events are presented in an unsupervised manner. Topic modeling-based prediction and representation are made based on daily news data between 1990 and 2015 from The Wall Street Journal against VIX index data from CBOE.

Keywords: Early Warning System, Knowledge Management, Topic Modeling, Market Prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1872
37 Barriers to the Use of Factoring Accounts Receivables: The Ghanaian Contractor’s Perception

Authors: E. Kissi, V. K. Acheamfour, J. J. Gyimah, T. Adjei-Kumi

Abstract:

Factoring accounts receivable is widely accepted as an alternative financing source and utilized in almost every industry that sells business-to-business or business-to-government. However, its patronage in the construction industry is very limited as some barriers hinder its application in the construction industry. This study aims at assessing the barriers to the use of factoring accounts receivables in the Ghanaian construction industry. The study adopted the sequential exploratory research method where structured and unstructured questionnaires were conveniently distributed to D1K1 and D2K2 construction firms in Ghana. Using the one-sample t-test and Kendall’s Coefficient of concordance data were analyzed. The most severe challenge concluded is the high cost of factoring patronage. Other critical challenges identified were low knowledge on factoring processes, inadequate access to information on factoring, and high risks involved in factoring. Hence, it is recommended that contractors should be made aware of the prospects of factoring of accounts receivables in the construction industry. This study serves as basis for further rigorous research into factoring of accounts receivables in the industry.

Keywords: Barriers, contractors, factoring accounts receivables, Ghanaian, perception.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 472
36 Online Forums Hotspot Detection and Analysis Using Aging Theory

Authors: K. Nirmala Devi, V. Murali Bhaskaran

Abstract:

The exponential growth of social media arouses much attention on public opinion information. The online forums, blogs, micro blogs are proving to be extremely valuable resources and are having bulk volume of information. However, most of the social media data is unstructured and semi structured form. So that it is more difficult to decipher automatically. Therefore, it is very much essential to understand and analyze those data for making a right decision. The online forums hotspot detection is a promising research field in the web mining and it guides to motivate the user to take right decision in right time. The proposed system consist of a novel approach to detect a hotspot forum for any given time period. It uses aging theory to find the hot terms and E-K-means for detecting the hotspot forum. Experimental results demonstrate that the proposed approach outperforms k-means for detecting the hotspot forums with the improved accuracy.

Keywords: Hotspot forums, Micro blog, Blog, Sentiment Analysis, Opinion Mining, Social media, Twitter, Web mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2139
35 Feature-Based Summarizing and Ranking from Customer Reviews

Authors: Dim En Nyaung, Thin Lai Lai Thein

Abstract:

Due to the rapid increase of Internet, web opinion sources dynamically emerge which is useful for both potential customers and product manufacturers for prediction and decision purposes. These are the user generated contents written in natural languages and are unstructured-free-texts scheme. Therefore, opinion mining techniques become popular to automatically process customer reviews for extracting product features and user opinions expressed over them. Since customer reviews may contain both opinionated and factual sentences, a supervised machine learning technique applies for subjectivity classification to improve the mining performance. In this paper, we dedicate our work is the task of opinion summarization. Therefore, product feature and opinion extraction is critical to opinion summarization, because its effectiveness significantly affects the identification of semantic relationships. The polarity and numeric score of all the features are determined by Senti-WordNet Lexicon. The problem of opinion summarization refers how to relate the opinion words with respect to a certain feature. Probabilistic based model of supervised learning will improve the result that is more flexible and effective.

Keywords: Opinion Mining, Opinion Summarization, Sentiment Analysis, Text Mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2894
34 Powerful Tool to Expand Business Intelligence: Text Mining

Authors: Li Gao, Elizabeth Chang, Song Han

Abstract:

With the extensive inclusion of document, especially text, in the business systems, data mining does not cover the full scope of Business Intelligence. Data mining cannot deliver its impact on extracting useful details from the large collection of unstructured and semi-structured written materials based on natural languages. The most pressing issue is to draw the potential business intelligence from text. In order to gain competitive advantages for the business, it is necessary to develop the new powerful tool, text mining, to expand the scope of business intelligence. In this paper, we will work out the strong points of text mining in extracting business intelligence from huge amount of textual information sources within business systems. We will apply text mining to each stage of Business Intelligence systems to prove that text mining is the powerful tool to expand the scope of BI. After reviewing basic definitions and some related technologies, we will discuss the relationship and the benefits of these to text mining. Some examples and applications of text mining will also be given. The motivation behind is to develop new approach to effective and efficient textual information analysis. Thus we can expand the scope of Business Intelligence using the powerful tool, text mining.

Keywords: Business intelligence, document warehouse, text mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2612
33 Enhance the Power of Sentiment Analysis

Authors: Yu Zhang, Pedro Desouza

Abstract:

Since big data has become substantially more accessible and manageable due to the development of powerful tools for dealing with unstructured data, people are eager to mine information from social media resources that could not be handled in the past. Sentiment analysis, as a novel branch of text mining, has in the last decade become increasingly important in marketing analysis, customer risk prediction and other fields. Scientists and researchers have undertaken significant work in creating and improving their sentiment models. In this paper, we present a concept of selecting appropriate classifiers based on the features and qualities of data sources by comparing the performances of five classifiers with three popular social media data sources: Twitter, Amazon Customer Reviews, and Movie Reviews. We introduced a couple of innovative models that outperform traditional sentiment classifiers for these data sources, and provide insights on how to further improve the predictive power of sentiment analysis. The modeling and testing work was done in R and Greenplum in-database analytic tools.

Keywords: Sentiment Analysis, Social Media, Twitter, Amazon, Data Mining, Machine Learning, Text Mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3474
32 Ontology-based Concept Weighting for Text Documents

Authors: Hmway Hmway Tar, Thi Thi Soe Nyaunt

Abstract:

Documents clustering become an essential technology with the popularity of the Internet. That also means that fast and high-quality document clustering technique play core topics. Text clustering or shortly clustering is about discovering semantically related groups in an unstructured collection of documents. Clustering has been very popular for a long time because it provides unique ways of digesting and generalizing large amounts of information. One of the issues of clustering is to extract proper feature (concept) of a problem domain. The existing clustering technology mainly focuses on term weight calculation. To achieve more accurate document clustering, more informative features including concept weight are important. Feature Selection is important for clustering process because some of the irrelevant or redundant feature may misguide the clustering results. To counteract this issue, the proposed system presents the concept weight for text clustering system developed based on a k-means algorithm in accordance with the principles of ontology so that the important of words of a cluster can be identified by the weight values. To a certain extent, it has resolved the semantic problem in specific areas.

Keywords: Clustering, Concept Weight, Document clustering, Feature Selection, Ontology

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2362
31 Comparison of GSA, SA and PSO Based Intelligent Controllers for Path Planning of Mobile Robot in Unknown Environment

Authors: P. K. Panigrahi, Saradindu Ghosh, Dayal R. Parhi

Abstract:

Now-a-days autonomous mobile robots have found applications in diverse fields. An autonomous robot system must be able to behave in an intelligent manner to deal with complex and changing environment. This work proposes the performance of path planning and navigation of autonomous mobile robot using Gravitational Search Algorithm (GSA), Simulated Annealing (SA) and Particle Swarm optimization (PSO) based intelligent controllers in an unstructured environment. The approach not only finds a valid collision free path but also optimal one. The main aim of the work is to minimize the length of the path and duration of travel from a starting point to a target while moving in an unknown environment with obstacles without collision. Finally, a comparison is made between the three controllers, it is found that the path length and time duration made by the robot using GSA is better than SA and PSO based controllers for the same work.

Keywords: Autonomous Mobile Robot, Gravitational Search Algorithm, Particle Swarm Optimization, Simulated Annealing Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3076
30 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm

Authors: Ameur Abdelkader, Abed Bouarfa Hafida

Abstract:

Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.

Keywords: Predictive analysis, big data, predictive analysis algorithms. CART algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1018
29 An Ontology Based Question Answering System on Software Test Document Domain

Authors: Meltem Serhatli, Ferda N. Alpaslan

Abstract:

Processing the data by computers and performing reasoning tasks is an important aim in Computer Science. Semantic Web is one step towards it. The use of ontologies to enhance the information by semantically is the current trend. Huge amount of domain specific, unstructured on-line data needs to be expressed in machine understandable and semantically searchable format. Currently users are often forced to search manually in the results returned by the keyword-based search services. They also want to use their native languages to express what they search. In this paper, an ontology-based automated question answering system on software test documents domain is presented. The system allows users to enter a question about the domain by means of natural language and returns exact answer of the questions. Conversion of the natural language question into the ontology based query is the challenging part of the system. To be able to achieve this, a new algorithm regarding free text to ontology based search engine query conversion is proposed. The algorithm is based on investigation of suitable question type and parsing the words of the question sentence.

Keywords: Description Logics, ontology, question answering, reasoning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2109
28 A Numerical Framework to Investigate Intake Aerodynamics Behavior in Icing Conditions

Authors: Ali Mirmohammadi, Arash Taheri, Meysam Mohammadi-Amin

Abstract:

One of the major parts of a jet engine is air intake, which provides proper and required amount of air for the engine to operate. There are several aerodynamic parameters which should be considered in design, such as distortion, pressure recovery, etc. In this research, the effects of lip ice accretion on pitot intake performance are investigated. For ice accretion phenomenon, two supervised multilayer neural networks (ANN) are designed, one for ice shape prediction and another one for ice roughness estimation based on experimental data. The Fourier coefficients of transformed ice shape and parameters include velocity, liquid water content (LWC), median volumetric diameter (MVD), spray time and temperature are used in neural network training. Then, the subsonic intake flow field is simulated numerically using 2D Navier-Stokes equations and Finite Volume approach with Hybrid mesh includes structured and unstructured meshes. The results are obtained in different angles of attack and the variations of intake aerodynamic parameters due to icing phenomenon are discussed. The results show noticeable effects of ice accretion phenomenon on intake behavior.

Keywords: Artificial Neural Network, Ice Accretion, IntakeAerodynamics, Design Parameters, Finite Volume Method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2160
27 The Urban Expansion Characterization of the Bir El Djir Municipality Using Remote Sensing and GIS

Authors: Zakaria Smahi, Khadidja Remaoun, Fatima Achouri

Abstract:

Bir El Djir is an important coastal township in Oran department, located at 450 Km far away from Algiers on northwest of Algeria. In this coastal area, the urban sprawl is one of the main problems that reduce the limited highly fertile land. So, using the remote sensing and GIS technologies have shown their great capabilities to solve many earth resources issues. The aim of this study is to produce land use and cover map for the studied area at varied periods to monitor possible changes that may occurred, particularly in the urban areas and subsequently predict likely changes. For this, two spatial images SPOT and Landsat satellites from 1987 and 2014 respectively were used to assess the changes of urban expansion and encroachment during this period with photo-interpretation and GIS approach. The results revealed that the town of Bir El Djir has shown a highest growth rate in the period 1987-2014 which is 1201.5 hectares in terms of area. These expansions largely concern the new real estate constructions falling within the social and promotional housing programs launched by the government. The most urban expansion is characterized by the new construction in the form of spontaneous or peripheral precarious habitat, but also unstructured slums settled especially in the southeastern part of town.

Keywords: Urban expansion, Remote Sensing, Photointerpretation, Spatial dynamics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2088
26 The Potential Benefits of Multimedia Information Representation in Enhancing Students’ Critical Thinking and History Reasoning

Authors: Ang Ling Weay, Mona Masood

Abstract:

This paper discusses the potential benefits of an interactive multimedia information representation in enhancing students’ critical thinking aligned with history reasoning in learning history amongst Secondary School students in Malaysia. Two modes of multimedia information representation were implemented; chronologic and thematic information representations. A qualitative study of an unstructured interview was conducted among two history teachers, one history education lecturer, two i-think experts, and five students from Form Four secondary school. The interview was to elicit their opinions on the implementation of thinking maps and interactive multimedia information representation in history learning. The key elements of the interactive multimedia (e.g. multiple media, user control, interactivity and use of timelines and concept maps) were then considered to improve the learning process. Findings of the preliminary investigation reveal that the interactive multimedia information representations have the potential benefits to be implemented as an instructional resource in enhancing students’ higher order thinking skills (HOTs). This paper concludes by giving suggestions for future work.

Keywords: Multimedia Information Representation, Critical Thinking, History Reasoning, Chronological and Thematic Information Representation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2349
25 Nonlinear Control of a Continuous Bioreactor Based on Cell Population Model

Authors: Mahdi Sharifian, Mohammad Ali Fanaei

Abstract:

Saccharomyces cerevisiae (baker-s yeast) can exhibit sustained oscillations during the operation in a continuous bioreactor that adversely affects its stability and productivity. Because of heterogeneous nature of cell populations, the cell population balance models can be used to capture the dynamic behavior of such cultures. In this paper an unstructured, segregated model is used which is based on population balance equation(PBE) and then in order to simulation, the 4th order Rung-Kutta is used for time dimension and three methods, finite difference, orthogonal collocation on finite elements and Galerkin finite element are used for discretization of the cell mass domain. The results indicate that the orthogonal collocation on finite element not only is able to predict the oscillating behavior of the cell culture but also needs much little time for calculations. Therefore this method is preferred in comparison with other methods. In the next step two controllers, a globally linearizing control (GLC) and a conventional proportional-integral (PI) controller are designed for controlling the total cell mass per unit volume, and performances of these controllers are compared through simulation. The results show that although the PI controller has simpler structure, the GLC has better performance.

Keywords: Bioreactor, cell population balance, finite difference, orthogonal collocation on finite elements, Galerkin finite element, feedback linearization, PI controller.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1840
24 Classifier Based Text Mining for Neural Network

Authors: M. Govindarajan, R. M. Chandrasekaran

Abstract:

Text Mining is around applying knowledge discovery techniques to unstructured text is termed knowledge discovery in text (KDT), or Text data mining or Text Mining. In Neural Network that address classification problems, training set, testing set, learning rate are considered as key tasks. That is collection of input/output patterns that are used to train the network and used to assess the network performance, set the rate of adjustments. This paper describes a proposed back propagation neural net classifier that performs cross validation for original Neural Network. In order to reduce the optimization of classification accuracy, training time. The feasibility the benefits of the proposed approach are demonstrated by means of five data sets like contact-lenses, cpu, weather symbolic, Weather, labor-nega-data. It is shown that , compared to exiting neural network, the training time is reduced by more than 10 times faster when the dataset is larger than CPU or the network has many hidden units while accuracy ('percent correct') was the same for all datasets but contact-lences, which is the only one with missing attributes. For contact-lences the accuracy with Proposed Neural Network was in average around 0.3 % less than with the original Neural Network. This algorithm is independent of specify data sets so that many ideas and solutions can be transferred to other classifier paradigms.

Keywords: Back propagation, classification accuracy, textmining, time complexity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4164
23 A Methodology for Automatic Diversification of Document Categories

Authors: Dasom Kim, Chen Liu, Myungsu Lim, Soo-Hyeon Jeon, Byeoung Kug Jeon, Kee-Young Kwahk, Namgyu Kim

Abstract:

Recently, numerous documents including large volumes of unstructured data and text have been created because of the rapid increase in the use of social media and the Internet. Usually, these documents are categorized for the convenience of users. Because the accuracy of manual categorization is not guaranteed, and such categorization requires a large amount of time and incurs huge costs. Many studies on automatic categorization have been conducted to help mitigate the limitations of manual categorization. Unfortunately, most of these methods cannot be applied to categorize complex documents with multiple topics because they work on the assumption that individual documents can be categorized into single categories only. Therefore, to overcome this limitation, some studies have attempted to categorize each document into multiple categories. However, the learning process employed in these studies involves training using a multi-categorized document set. These methods therefore cannot be applied to the multi-categorization of most documents unless multi-categorized training sets using traditional multi-categorization algorithms are provided. To overcome this limitation, in this study, we review our novel methodology for extending the category of a single-categorized document to multiple categorizes, and then introduce a survey-based verification scenario for estimating the accuracy of our automatic categorization methodology.

Keywords: Big Data Analysis, Document Classification, Text Mining, Topic Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1708