Search results for: Web usage mining.
196 Image Classification and Accuracy Assessment Using the Confusion Matrix, Contingency Matrix, and Kappa Coefficient
Authors: F. F. Howard, C. B. Boye, I. Yakubu, J. S. Y. Kuma
Abstract:
One of the ways that could be used for the production of land use and land cover maps by a procedure known as image classification is the use of the remote sensing technique. Numerous elements ought to be taken into consideration, including the availability of highly satisfactory Landsat imagery, secondary data and a precise classification process. The goal of this study was to classify and map the land use and land cover of the study area using remote sensing and Geospatial Information System (GIS) analysis. The classification was done using Landsat 8 satellite images acquired in December 2020 covering the study area. The Landsat image was downloaded from the USGS. The Landsat image with 30 m resolution was geo-referenced to the WGS_84 datum and Universal Transverse Mercator (UTM) Zone 30N coordinate projection system. A radiometric correction was applied to the image to reduce the noise in the image. This study consists of two sections: the Land Use/Land Cover (LULC) and Accuracy Assessments using the confusion and contingency matrix and the Kappa coefficient. The LULC classifications were vegetation (agriculture) (67.87%), water bodies (0.01%), mining areas (5.24%), forest (26.02%), and settlement (0.88%). The overall accuracy of 97.87% and the kappa coefficient (K) of 97.3% were obtained for the confusion matrix. While an overall accuracy of 95.7% and a Kappa coefficient of 0.947 were obtained for the contingency matrix, the kappa coefficients were rated as substantial; hence, the classified image is fit for further research.
Keywords: Confusion Matrix, contingency matrix, kappa coefficient, land used/ land cover, accuracy assessment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 265195 Result of Fatty Acid Content in Meat of Selenge Breed Younger Cattle
Authors: Myagmarsuren Soronzonjav, N. Togtokhbayar, L. Davaahuu, B. Minjigdorj, Seong Gu Hwang
Abstract:
The number of natural or organic product consumers is increased in recent years and this healthy demand pushes to increase usage of healthy meat. At the same time, consumers pay more attention on the healthy fat, especially on unsaturated fatty acids. These long chain carbohydrates reduce heart diseases, improve memory and eye sight and activate the immune system. One of the important issues to be solved for our Mongolia’s food security is to provide healthy, fresh, widely available and cheap meat for the population. Thus, an importance of the Selenge breed meat production is increasing in order to supply the quality meat food security since the Selenge breed cattle are rapidly multiplied, beneficial in term of income, the same quality as Mongolian breed, and well digested for human body. We researched the lipid, unsaturated and saturated fatty acid contents of meat of Selenge breed younger cattle by their muscle types. Result of our research reveals that 11 saturated fatty acids are detected. For the content of palmitic acid among saturated fatty acids, 23.61% was in the sirloin meat, 24.01% was in the round and chuck meat, and 24.83% was in the short loin meat.
Keywords: Chromatogram, gas chromatography, organic resolving, saturated and unsaturated fatty acids.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1066194 Fuzzy Wavelet Packet based Feature Extraction Method for Multifunction Myoelectric Control
Authors: Rami N. Khushaba, Adel Al-Jumaily
Abstract:
The myoelectric signal (MES) is one of the Biosignals utilized in helping humans to control equipments. Recent approaches in MES classification to control prosthetic devices employing pattern recognition techniques revealed two problems, first, the classification performance of the system starts degrading when the number of motion classes to be classified increases, second, in order to solve the first problem, additional complicated methods were utilized which increase the computational cost of a multifunction myoelectric control system. In an effort to solve these problems and to achieve a feasible design for real time implementation with high overall accuracy, this paper presents a new method for feature extraction in MES recognition systems. The method works by extracting features using Wavelet Packet Transform (WPT) applied on the MES from multiple channels, and then employs Fuzzy c-means (FCM) algorithm to generate a measure that judges on features suitability for classification. Finally, Principle Component Analysis (PCA) is utilized to reduce the size of the data before computing the classification accuracy with a multilayer perceptron neural network. The proposed system produces powerful classification results (99% accuracy) by using only a small portion of the original feature set.Keywords: Biomedical Signal Processing, Data mining andInformation Extraction, Machine Learning, Rehabilitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1738193 Comparison of Welding Fumes Exposure during Standing and Sitting Welder’s Position
Authors: Azian Hariri, M. Z. M Yusof, A. M. Leman
Abstract:
Experimental study was conducted to assess personal welding fumes exposure toward welders during an aluminum metal inert gas (MIG) process. The welding process was carried out by a welding machine attached to a Computer Numerical Control (CNC) workbench. A dummy welder was used to replicate welder during welding works and was attached with sampling pumps and filter cassettes for welding fumes sampling. Direct reading instruments to measure air velocity, humidity, temperature and particulate matter with diameter size 10µm or less (PM10) were located behind the dummy welder and parallel to the neck collar level to make sure the measured welding fumes exposure were not being influenced by other factors. Welding fumes exposure during standing and sitting position with and without the usage of local exhaust ventilation (LEV) was investigated. Welding fume samples were then digested and analyzed by using inductively coupled plasma mass spectroscopy (ICP-MS) according to ASTM D7439-08 method. The results of the study showed the welding fume exposure during sitting was lower compared to standing position. LEV helped reduce aluminum and lead exposure to acceptable levels during standing position. However during sitting position reduction of exposure was smaller. It can be concluded that welder position and the correct positioning of LEV should be implemented for effective exposure reduction.
Keywords: ICP-MS, MIG process, personal sampling, welding fumes exposure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2611192 Simulation Study on the Indoor Thermal Comfort with Insulation on Interior Structural Components of Super High-Rise Residences
Authors: Y. Wang, H. Fukuda, A. Ozaki, H. Sato
Abstract:
In this study, we discussed the effects on the thermal comfort of super high-rise residences that how effected by the high thermal capacity structural components. We considered different building orientations, structures, and insulation methods. We used the dynamic simulation software THERB (simulation of the thermal environment of residential buildings). It can estimate the temperature, humidity, sensible temperature, and heating/cooling load for multiple buildings. In the past studies, we examined the impact of air-conditioning loads (hereinafter referred to as AC loads) on the interior structural parts and the AC-usage patterns of super-high-rise residences. Super-high-rise residences have more structural components such as pillars and beams than do ordinary apartment buildings. The skeleton is generally made of concrete and steel, which have high thermal-storage capacities. The thermal-storage capacity of super-high-rise residences is considered to have a larger impact on the AC load and thermal comfort than that of ordinary residences. We show that the AC load of super-high-rise units would be reduced by installing insulation on the surfaces of interior walls that are not usually insulated in Japan.Keywords: High-rise Residences, AC Load, Thermal Comfort, Thermal Storage, Insulation Patterns
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1544191 Context Aware Anomaly Behavior Analysis for Smart Home Systems
Authors: Zhiwen Pan, Jesus Pacheco, Salim Hariri, Yiqiang Chen, Bozhi Liu
Abstract:
The Internet of Things (IoT) will lead to the development of advanced Smart Home services that are pervasive, cost-effective, and can be accessed by home occupants from anywhere and at any time. However, advanced smart home applications will introduce grand security challenges due to the increase in the attack surface. Current approaches do not handle cybersecurity from a holistic point of view; hence, a systematic cybersecurity mechanism needs to be adopted when designing smart home applications. In this paper, we present a generic intrusion detection methodology to detect and mitigate the anomaly behaviors happened in Smart Home Systems (SHS). By utilizing our Smart Home Context Data Structure, the heterogeneous information and services acquired from SHS are mapped in context attributes which can describe the context of smart home operation precisely and accurately. Runtime models for describing usage patterns of home assets are developed based on characterization functions. A threat-aware action management methodology, used to efficiently mitigate anomaly behaviors, is proposed at the end. Our preliminary experimental results show that our methodology can be used to detect and mitigate known and unknown threats, as well as to protect SHS premises and services.
Keywords: Internet of Things, network security, context awareness, intrusion detection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1282190 Supplementary Cementitious Materials as Sustainable Partial Replacement for Cement in the Building Industry
Authors: Nwakaego C. Onyenokporo
Abstract:
Cement is the most extensively used construction material due to its strength and versatility of use. However, the production of Portland cement has become unsustainable because of high energy usage, reduction of natural non-renewable resources and emissions of greenhouse gases. Production of cement contributes to anthropogenic greenhouse gases emissions annually. The growing concerns for the environment resulting from this constant and excessive use of cement has therefore raised the need for more green materials and technology. The use of supplementary cementitious materials (SCMs) is considered as one of the many alternatives suited to address this issue and serve as a sustainable partial replacement for cement in construction. This paper will examine the reuse of these waste materials to partially replace Portland cement. It provides a critical review of literature analysing various supplementary cementitious materials which are applicable in the building industry as either partial replacement for cement or aggregates. These materials have been grouped based on source into industrial wastes, domestic/general wastes, and agricultural wastes. The reuse of these waste materials could potentially reduce the negative effects of cement production and reduce landfills which constitute an environmental nuisance. This paper seeks to inform building industry professionals and researchers in the field on the applicability of these waste materials in construction.
Keywords: cement, greenhouse gases, landfills, sustainable, waste materials
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 743189 Supplementary Cementitious Materials as Sustainable Partial Replacement for Cement in the Building Industry
Authors: Nwakaego C. Onyenokporo
Abstract:
Cement is the most extensively used construction material due to its strength and versatility of use. However, the production of Portland cement has become unsustainable because of high energy usage, reduction of natural non-renewable resources and emissions of greenhouse gases. Production of cement contributes to anthropogenic greenhouse gases emissions annually. The growing concerns for the environment resulting from this constant and excessive use of cement has therefore raised the need for more green materials and technology. The use of supplementary cementitious materials (SCMs) is considered as one of the many alternatives suited to address this issue and serve as a sustainable partial replacement for cement in construction. This paper will examine the reuse of these waste materials to partially replace Portland cement. It provides a critical review of literature analysing various supplementary cementitious materials which are applicable in the building industry as either partial replacement for cement or aggregates. These materials have been grouped based on source into industrial wastes, domestic/general wastes, and agricultural wastes. The reuse of these waste materials could potentially reduce the negative effects of cement production and reduce landfills which constitute an environmental nuisance. This paper seeks to inform building industry professionals and researchers in the field on the applicability of these waste materials in construction.
Keywords: Cement, greenhouse gases, landfills, sustainable, waste materials.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 684188 An Efficient Watermarking Method for MP3 Audio Files
Authors: Dimitrios Koukopoulos, Yiannis Stamatiou
Abstract:
In this work, we present for the first time in our perception an efficient digital watermarking scheme for mpeg audio layer 3 files that operates directly in the compressed data domain, while manipulating the time and subband/channel domain. In addition, it does not need the original signal to detect the watermark. Our scheme was implemented taking special care for the efficient usage of the two limited resources of computer systems: time and space. It offers to the industrial user the capability of watermark embedding and detection in time immediately comparable to the real music time of the original audio file that depends on the mpeg compression, while the end user/audience does not face any artifacts or delays hearing the watermarked audio file. Furthermore, it overcomes the disadvantage of algorithms operating in the PCMData domain to be vulnerable to compression/recompression attacks, as it places the watermark in the scale factors domain and not in the digitized sound audio data. The strength of our scheme, that allows it to be used with success in both authentication and copyright protection, relies on the fact that it gives to the users the enhanced capability their ownership of the audio file not to be accomplished simply by detecting the bit pattern that comprises the watermark itself, but by showing that the legal owner knows a hard to compute property of the watermark.
Keywords: Audio watermarking, mpeg audio layer 3, hard instance generation, NP-completeness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1836187 Fuzzy Relatives of the CLARANS Algorithm With Application to Text Clustering
Authors: Mohamed A. Mahfouz, M. A. Ismail
Abstract:
This paper introduces new algorithms (Fuzzy relative of the CLARANS algorithm FCLARANS and Fuzzy c Medoids based on randomized search FCMRANS) for fuzzy clustering of relational data. Unlike existing fuzzy c-medoids algorithm (FCMdd) in which the within cluster dissimilarity of each cluster is minimized in each iteration by recomputing new medoids given current memberships, FCLARANS minimizes the same objective function minimized by FCMdd by changing current medoids in such away that that the sum of the within cluster dissimilarities is minimized. Computing new medoids may be effected by noise because outliers may join the computation of medoids while the choice of medoids in FCLARANS is dictated by the location of a predominant fraction of points inside a cluster and, therefore, it is less sensitive to the presence of outliers. In FCMRANS the step of computing new medoids in FCMdd is modified to be based on randomized search. Furthermore, a new initialization procedure is developed that add randomness to the initialization procedure used with FCMdd. Both FCLARANS and FCMRANS are compared with the robust and linearized version of fuzzy c-medoids (RFCMdd). Experimental results with different samples of the Reuter-21578, Newsgroups (20NG) and generated datasets with noise show that FCLARANS is more robust than both RFCMdd and FCMRANS. Finally, both FCMRANS and FCLARANS are more efficient and their outputs are almost the same as that of RFCMdd in terms of classification rate.Keywords: Data Mining, Fuzzy Clustering, Relational Clustering, Medoid-Based Clustering, Cluster Analysis, Unsupervised Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2404186 Integrating AI Visualization Tools to Enhance Student Engagement and Understanding in AI Education
Authors: Yong W. Foo, Lai M. Tang
Abstract:
Artificial Intelligence (AI), particularly the usage of deep neural networks for hierarchical representations from data, has found numerous complex applications across various domains, including computer vision, robotics, autonomous vehicles, and other scientific fields. However, their inherent “black box” nature can sometimes make it challenging for early researchers or school students of various levels to comprehend and trust the results they produce. Consequently, there has been a growing demand for reliable visualization tools in engineering and science education to help learners understand, trust, and explain a deep learning network. This has led to a notable emphasis on the visualization of AI in the research community in recent years. AI visualization tools are increasingly being adopted to significantly improve the comprehension of complex topics in deep learning. This paper presents an approach to empower students to actively explore the inner workings of deep neural networks by integrating the student-centered learning approach of flipped classroom models with the investigative capabilities of AI visualization tools, namely, the TensorFlow Playground, the Local Interpretable Model-agnostic Explanations (LIME), and the SHapley Additive exPlanations (SHAP), for delivering an AI education curriculum. Integrating these two factors is crucial for fostering ownership, responsibility, and critical thinking skills in the age of AI.
Keywords: Deep Learning, Explainable AI, AI Visualization, Representation Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42185 The Development of Online Lessons in Integration Model
Authors: Chalermpol Tapsai
Abstract:
The objectives of this research were to develop and find the efficiency of integrated online lessons by investigating the usage of online lessons, the relationship between learners’ background knowledge, and the achievement after learning with online lessons. The sample group in this study consisted of 97 students randomly selected from 121 students registering in 1/2012 at Trimitwittayaram Learning Center. The sample technique employed stratified sample technique of 4 groups according to their proficiency, i.e. high, moderate, low, and non-knowledge. The research instrument included online lessons in integration model on the topic of Java Programming, test after each lesson, the achievement test at the end of the course, and the questionnaires to find learners’ satisfaction. The results showed that the efficiency of online lessons was 90.20/89.18 with the achievement of after learning with the lessons higher than that before the lessons at the statistically significant level of 0.05. Moreover, the background knowledge of the learners on the programming showed the positive relationship with the achievement learning at the statistically significant level at 0.05. Learners with high background knowledge employed less exercises and samples than those with lower background knowledge. While learners with different background in the group of moderate and low did not show the significant difference in employing samples and exercises.
Keywords: Integration model, Online lessons.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1490184 GIS-based Non-point Sources of Pollution Simulation in Cameron Highlands, Malaysia
Authors: M. Eisakhani, A. Pauzi, O. Karim, A. Malakahmad, S.R. Mohamed Kutty, M. H. Isa
Abstract:
Cameron Highlands is a mountainous area subjected to torrential tropical showers. It extracts 5.8 million liters of water per day for drinking supply from its rivers at several intake points. The water quality of rivers in Cameron Highlands, however, has deteriorated significantly due to land clearing for agriculture, excessive usage of pesticides and fertilizers as well as construction activities in rapidly developing urban areas. On the other hand, these pollution sources known as non-point pollution sources are diverse and hard to identify and therefore they are difficult to estimate. Hence, Geographical Information Systems (GIS) was used to provide an extensive approach to evaluate landuse and other mapping characteristics to explain the spatial distribution of non-point sources of contamination in Cameron Highlands. The method to assess pollution sources has been developed by using Cameron Highlands Master Plan (2006-2010) for integrating GIS, databases, as well as pollution loads in the area of study. The results show highest annual runoff is created by forest, 3.56 × 108 m3/yr followed by urban development, 1.46 × 108 m3/yr. Furthermore, urban development causes highest BOD load (1.31 × 106 kgBOD/yr) while agricultural activities and forest contribute the highest annual loads for phosphorus (6.91 × 104 kgP/yr) and nitrogen (2.50 × 105 kgN/yr), respectively. Therefore, best management practices (BMPs) are suggested to be applied to reduce pollution level in the area.Keywords: Cameron Highlands, Land use, Non-point Sources of Pollution
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2879183 Intensifier as Changed from the Impolite Word in Thai
Authors: Methawee Yuttapongtada
Abstract:
Intensifier is the linguistic term and device that is generally found in different languages in order to enhance and give additional quantity, quality or emotion to the words of each language. In fact, each language in the world has both of the similar and dissimilar intensifying device. More specially, the wide variety of intensifying device is used for Thai language and one of those is usage of the impolite word or the word that used to mean something negative as intensifier. The data collection in this study was done throughout the spoken language style by collecting from intensifiers regarded as impolite words because these words as employed in the other contexts will be held as the rude, swear words or the words with negative meaning. Then, backward study to the past was done in order to consider the historical change. Explanation of the original meaning and the contexts of words use from the past till the present time were done by use of both textual documents and dictionaries available in different periods. It was found that regarding the semantics and pragmatic aspects, subjectification also is the significant motivation that changed the impolite words to intensifiers. At last, it can explain pathway of the semantic change of these very words undoubtedly. Moreover, it is found that use tendency in the impolite word or the word that used to mean something negative will more be increased and this phenomenon is commonly found in many languages in the world and results of this research may support to the belief that human language in the world is universal and the same still reflected that human has the fundamental thought as the same to each other basically.
Keywords: Impolite word, intensifier, Thai, semantic change.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1312182 Learning Classifier Systems Approach for Automated Discovery of Crisp and Fuzzy Hierarchical Production Rules
Authors: Suraiya Jabin, Kamal K. Bharadwaj
Abstract:
This research presents a system for post processing of data that takes mined flat rules as input and discovers crisp as well as fuzzy hierarchical structures using Learning Classifier System approach. Learning Classifier System (LCS) is basically a machine learning technique that combines evolutionary computing, reinforcement learning, supervised or unsupervised learning and heuristics to produce adaptive systems. A LCS learns by interacting with an environment from which it receives feedback in the form of numerical reward. Learning is achieved by trying to maximize the amount of reward received. Crisp description for a concept usually cannot represent human knowledge completely and practically. In the proposed Learning Classifier System initial population is constructed as a random collection of HPR–trees (related production rules) and crisp / fuzzy hierarchies are evolved. A fuzzy subsumption relation is suggested for the proposed system and based on Subsumption Matrix (SM), a suitable fitness function is proposed. Suitable genetic operators are proposed for the chosen chromosome representation method. For implementing reinforcement a suitable reward and punishment scheme is also proposed. Experimental results are presented to demonstrate the performance of the proposed system.Keywords: Hierarchical Production Rule, Data Mining, Learning Classifier System, Fuzzy Subsumption Relation, Subsumption matrix, Reinforcement Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1456181 Active Intra-ONU Scheduling with Cooperative Prediction Mechanism in EPONs
Authors: Chuan-Ching Sue, Shi-Zhou Chen, Ting-Yu Huang
Abstract:
Dynamic bandwidth allocation in EPONs can be generally separated into inter-ONU scheduling and intra-ONU scheduling. In our previous work, the active intra-ONU scheduling (AS) utilizes multiple queue reports (QRs) in each report message to cooperate with the inter-ONU scheduling and makes the granted bandwidth fully utilized without leaving unused slot remainder (USR). This scheme successfully solves the USR problem originating from the inseparability of Ethernet frame. However, without proper setting of threshold value in AS, the number of QRs constrained by the IEEE 802.3ah standard is not enough, especially in the unbalanced traffic environment. This limitation may be solved by enlarging the threshold value. The large threshold implies the large gap between the adjacent QRs, thus resulting in the large difference between the best granted bandwidth and the real granted bandwidth. In this paper, we integrate AS with a cooperative prediction mechanism and distribute multiple QRs to reduce the penalty brought by the prediction error. Furthermore, to improve the QoS and save the usage of queue reports, the highest priority (EF) traffic which comes during the waiting time is granted automatically by OLT and is not considered in the requested bandwidth of ONU. The simulation results show that the proposed scheme has better performance metrics in terms of bandwidth utilization and average delay for different classes of packets.Keywords: EPON, Inter-ONU and Intra-ONU scheduling, Prediction, Unused slot remainder
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1596180 Visual Text Analytics Technologies for Real-Time Big Data: Chronological Evolution and Issues
Authors: Siti Azrina B. A. Aziz, Siti Hafizah A. Hamid
Abstract:
New approaches to analyze and visualize data stream in real-time basis is important in making a prompt decision by the decision maker. Financial market trading and surveillance, large-scale emergency response and crowd control are some example scenarios that require real-time analytic and data visualization. This situation has led to the development of techniques and tools that support humans in analyzing the source data. With the emergence of Big Data and social media, new techniques and tools are required in order to process the streaming data. Today, ranges of tools which implement some of these functionalities are available. In this paper, we present chronological evolution evaluation of technologies for supporting of real-time analytic and visualization of the data stream. Based on the past research papers published from 2002 to 2014, we gathered the general information, main techniques, challenges and open issues. The techniques for streaming text visualization are identified based on Text Visualization Browser in chronological order. This paper aims to review the evolution of streaming text visualization techniques and tools, as well as to discuss the problems and challenges for each of identified tools.Keywords: Information visualization, visual analytics, text mining, visual text analytics tools, big data visualization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1003179 ANN Based Currency Recognition System using Compressed Gray Scale and Application for Sri Lankan Currency Notes - SLCRec
Authors: D. A. K. S. Gunaratna, N. D. Kodikara, H. L. Premaratne
Abstract:
Automatic currency note recognition invariably depends on the currency note characteristics of a particular country and the extraction of features directly affects the recognition ability. Sri Lanka has not been involved in any kind of research or implementation of this kind. The proposed system “SLCRec" comes up with a solution focusing on minimizing false rejection of notes. Sri Lankan currency notes undergo severe changes in image quality in usage. Hence a special linear transformation function is adapted to wipe out noise patterns from backgrounds without affecting the notes- characteristic images and re-appear images of interest. The transformation maps the original gray scale range into a smaller range of 0 to 125. Applying Edge detection after the transformation provided better robustness for noise and fair representation of edges for new and old damaged notes. A three layer back propagation neural network is presented with the number of edges detected in row order of the notes and classification is accepted in four classes of interest which are 100, 500, 1000 and 2000 rupee notes. The experiments showed good classification results and proved that the proposed methodology has the capability of separating classes properly in varying image conditions.Keywords: Artificial intelligence, linear transformation and pattern recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2834178 The Effectiveness of Banks’ Web Sites: A Study of Turkish Banking Sector
Authors: Raif Parlakkaya, Huseyin Cetin, Duygu Irdiren
Abstract:
By the development of World Wide Web, the usage rate of Internet has rapidly grown globally; and provided a basis for the emergence of electronic business. As well as other sectors, the banking sector has adopted the use of internet with the developments in information and communication technologies. Due to the public disclosure and transparency principle of Corporate Governance, the importance of information disclosure of banks on their web sites has increased significantly. For the purpose of this study, a Bank Disclosure Attribute Index (BDAI) in Turkey has been constructed through classifying the information disclosure on banks’ web sites into general, financial, investors and corporate governance attributes. All 47 banks in Turkish Banking System have been evaluated according to the index with the aim of providing a comparison between banks. By Chi Square Test, Pearson Correlation, T-Test, and ANOVA statistical tools, it has been concluded that the majority of banks in Turkey have shared information on their web sites adequately with respect to their total index score. Although there is a positive correlation between various types of information on banks’ web sites, there is no uniformity among them. Also, no significant difference between various types of information disclosure and bank types has been observed. Compared with the total index score averages of the five largest banks in Turkey, there are some banks that need to improve the content of their web sites.
Keywords: Banking sector, public disclosure, Turkey, web site evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1425177 An Evaluation of Kahoot Application and Its Environment as a Learning Tool
Authors: Muhammad Yasir Babar, Ebrahim Panah
Abstract:
Over the past 20 years, internet has seen continual advancement and with the advent of online technology, various types of web-based games have been developed. Games are frequently being used among different age groups from baby boomers to generation Z. Games are not only used for entertainment but also utilized as a learning approach transmitting education to a level that is more interesting and effective for students. One of the popular web-based education games is Kahoot with growing popularity and usage, which is being used in different fields of studies. However, little knowledge is available on university students’ perception of Kahoot environment and application for learning subjects. Hence, the objective of the current study is to investigate students’ perceptions of Kahoot application and environment as a learning tool. The study employed a survey approach by distributing Google Forms –created questionnaire, with high level of reliability index, to 62 students (11 males and 51 females). The findings show that students have positive attitudes towards Kahoot application and its environment for learning. Regarding Kahoot application, it was indicated that activities created using Kahoot are more interesting for students, Kahoot is useful for collaborative learning, and Kahoot enhances interest in learning lesson. In terms of Kahoot environment, it was found that using this application through mobile is easy for students, its design is simple and useful, Kahoot-created activities can easily be shared, and the application can easily be used on any platform. The findings of the study have implications for instructors, policymakers and curriculum developers.
Keywords: Application, environment, Kahoot, learning tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 801176 Intelligent Recognition of Diabetes Disease via FCM Based Attribute Weighting
Authors: Kemal Polat
Abstract:
In this paper, an attribute weighting method called fuzzy C-means clustering based attribute weighting (FCMAW) for classification of Diabetes disease dataset has been used. The aims of this study are to reduce the variance within attributes of diabetes dataset and to improve the classification accuracy of classifier algorithm transforming from non-linear separable datasets to linearly separable datasets. Pima Indians Diabetes dataset has two classes including normal subjects (500 instances) and diabetes subjects (268 instances). Fuzzy C-means clustering is an improved version of K-means clustering method and is one of most used clustering methods in data mining and machine learning applications. In this study, as the first stage, fuzzy C-means clustering process has been used for finding the centers of attributes in Pima Indians diabetes dataset and then weighted the dataset according to the ratios of the means of attributes to centers of theirs. Secondly, after weighting process, the classifier algorithms including support vector machine (SVM) and k-NN (k- nearest neighbor) classifiers have been used for classifying weighted Pima Indians diabetes dataset. Experimental results show that the proposed attribute weighting method (FCMAW) has obtained very promising results in the classification of Pima Indians diabetes dataset.
Keywords: Fuzzy C-means clustering, Fuzzy C-means clustering based attribute weighting, Pima Indians diabetes dataset, SVM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1767175 A Fuzzy-Rough Feature Selection Based on Binary Shuffled Frog Leaping Algorithm
Authors: Javad Rahimipour Anaraki, Saeed Samet, Mahdi Eftekhari, Chang Wook Ahn
Abstract:
Feature selection and attribute reduction are crucial problems, and widely used techniques in the field of machine learning, data mining and pattern recognition to overcome the well-known phenomenon of the Curse of Dimensionality. This paper presents a feature selection method that efficiently carries out attribute reduction, thereby selecting the most informative features of a dataset. It consists of two components: 1) a measure for feature subset evaluation, and 2) a search strategy. For the evaluation measure, we have employed the fuzzy-rough dependency degree (FRFDD) of the lower approximation-based fuzzy-rough feature selection (L-FRFS) due to its effectiveness in feature selection. As for the search strategy, a modified version of a binary shuffled frog leaping algorithm is proposed (B-SFLA). The proposed feature selection method is obtained by hybridizing the B-SFLA with the FRDD. Nine classifiers have been employed to compare the proposed approach with several existing methods over twenty two datasets, including nine high dimensional and large ones, from the UCI repository. The experimental results demonstrate that the B-SFLA approach significantly outperforms other metaheuristic methods in terms of the number of selected features and the classification accuracy.Keywords: Binary shuffled frog leaping algorithm, feature selection, fuzzy-rough set, minimal reduct.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 731174 Relevance Feedback within CBIR Systems
Authors: Mawloud Mosbah, Bachir Boucheham
Abstract:
We present here the results for a comparative study of some techniques, available in the literature, related to the relevance feedback mechanism in the case of a short-term learning. Only one method among those considered here is belonging to the data mining field which is the K-nearest neighbors algorithm (KNN) while the rest of the methods is related purely to the information retrieval field and they fall under the purview of the following three major axes: Shifting query, Feature Weighting and the optimization of the parameters of similarity metric. As a contribution, and in addition to the comparative purpose, we propose a new version of the KNN algorithm referred to as an incremental KNN which is distinct from the original version in the sense that besides the influence of the seeds, the rate of the actual target image is influenced also by the images already rated. The results presented here have been obtained after experiments conducted on the Wang database for one iteration and utilizing color moments on the RGB space. This compact descriptor, Color Moments, is adequate for the efficiency purposes needed in the case of interactive systems. The results obtained allow us to claim that the proposed algorithm proves good results; it even outperforms a wide range of techniques available in the literature.
Keywords: CBIR, Category Search, Relevance Feedback (RFB), Query Point Movement, Standard Rocchio’s Formula, Adaptive Shifting Query, Feature Weighting, Optimization of the Parameters of Similarity Metric, Original KNN, Incremental KNN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2342173 ParkedGuard: An Efficient and Accurate Parked Domain Detection System Using Graphical Locality Analysis and Coarse-To-Fine Strategy
Authors: Chia-Min Lai, Wan-Ching Lin, Hahn-Ming Lee, Ching-Hao Mao
Abstract:
As world wild internet has non-stop developments, making profit by lending registered domain names emerges as a new business in recent years. Unfortunately, the larger the market scale of domain lending service becomes, the riskier that there exist malicious behaviors or malwares hiding behind parked domains will be. Also, previous work for differentiating parked domain suffers two main defects: 1) too much data-collecting effort and CPU latency needed for features engineering and 2) ineffectiveness when detecting parked domains containing external links that are usually abused by hackers, e.g., drive-by download attack. Aiming for alleviating above defects without sacrificing practical usability, this paper proposes ParkedGuard as an efficient and accurate parked domain detector. Several scripting behavioral features were analyzed, while those with special statistical significance are adopted in ParkedGuard to make feature engineering much more cost-efficient. On the other hand, finding memberships between external links and parked domains was modeled as a graph mining problem, and a coarse-to-fine strategy was elaborately designed by leverage the graphical locality such that ParkedGuard outperforms the state-of-the-art in terms of both recall and precision rates.Keywords: Coarse-to-fine strategy, domain parking service, graphical locality analysis, parked domain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1253172 Efficiency in Urban Governance towards Sustainability and Competitiveness of City : A Case Study of Kuala Lumpur
Authors: Hamzah Jusoh, Azmizam Abdul Rashid
Abstract:
Malaysia has successfully applied economic planning to guide the development of the country from an economy of agriculture and mining to a largely industrialised one. Now, with its sights set on attaining the economic level of a fully developed nation by 2020, the planning system must be made even more efficient and focused. It must ensure that every investment made in the country, contribute towards creating the desirable objective of a strong, modern, internationally competitive, technologically advanced, post-industrial economy. Cities in Malaysia must also be fully aware of the enormous competition it faces in a region with rapidly expanding and modernising economies, all contending for the same pool of potential international investments. Efficiency of urban governance is also fundamental issue in development characterized by sustainability, subsidiarity, equity, transparency and accountability, civic engagement and citizenship, and security. As described above, city competitiveness is harnessed through 'city marketing and city management'. High technology and high skilled industries, together with finance, transportation, tourism, business, information and professional services shopping and other commercial activities, are the principal components of the nation-s economy, which must be developed to a level well beyond where it is now. In this respect, Kuala Lumpur being the premier city must play the leading role.Keywords: Economic planning, sustainability, efficiency, urban governance and city competitiveness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2242171 Spreading Japan's National Image through China during the Era of Mass Tourism: The Japan National Tourism Organization’s Use of Sina Weibo
Authors: Abigail Qian Zhou
Abstract:
Since China has entered an era of mass tourism, there has been a fundamental change in the way Chinese people approach and perceive the image of other countries. With the advent of the new media era, social networking sites such as Sina Weibo have become a tool for many foreign governmental organizations to spread and promote their national image. Among them, the Japan National Tourism Organization (JNTO) was one of the first foreign official tourism agencies to register with Sina Weibo and actively implement communication activities. Due to historical and political reasons, cognition of Japan's national image by the Chinese has always been complicated and contradictory. However, since 2015, China has become the largest source of tourists visiting Japan. This clearly indicates that the broadening of Japan's national image in China has been effective and has value worthy of reference in promoting a positive Chinese perception of Japan and encouraging Japanese tourism. Within this context and using the method of content analysis in media studies through content mining software, this study analyzed how JNTO’s Sina Weibo accounts have constructed and spread Japan's national image. This study also summarized the characteristics of its content and form, and finally revealed the strategy of JNTO in building its international image. The findings of this study not only add a tourism-based perspective to traditional national image communications research, but also provide some reference for the effective international dissemination of national image in the future.Keywords: National image, tourism, international communication, Japan, China.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1001170 Chemical and Sensory Properties of Chardonnay Wines Produced in Different Oak Barrels
Authors: Valentina Obradović, Josip Mesić, Maja Ergović Ravančić, Kamila Mijowska, Brankica Svitlica
Abstract:
French oak and American oak barrels are most famous all over the world, but barrels of different origin can also be used for obtaining high quality wines. The aim of this research was to compare the influence of different Slovenian (Croatian) and French oak barrels on the quality of Chardonnay wine. Grapes were grown in the Croatian wine growing region of Kutjevo in 2015. Chardonnay wines were tested for basic oenological parameters (alcohol, extract, reducing sugar, SO2, acidity), total polyphenols content (Folin-Ciocalteu method), antioxidant activity (ABTS and DPPH method) and colour density. Sensory evaluation was performed by students of viticulture/oenology. Samples produced by classical fermentation and ageing in French oak barrels had better results for polyphenols and sensory evaluation (especially low toasting level) than samples in Slovenian barrels. All tested samples were scored as a “quality” or “premium quality” wines. Sur lie method of fermentation and ageing in Slovenian oak barrel had very good extraction of polyphenols and high antioxidant activity with the usage of authentic yeasts, while commercial yeast strain resulted in worse chemical and sensory parameters.
Keywords: Chardonnay, French oak, Slovenian oak, sur lie.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 947169 An Archetype to Sustain Knowledge Management Systems through Intranet
Authors: B. T. Sayed, Nafaâ Jabeur, M. Aref
Abstract:
Creation and maintenance of knowledge management systems has been recognized as an important research area. Consecutively lack of accurate results from knowledge management systems limits the organization to apply their knowledge management processes. This leads to a failure in getting the right information to the right people at the right time thus followed by a deficiency in decision making processes. An Intranet offers a powerful tool for communication and collaboration, presenting data and information, and the means that creates and shares knowledge, all in one easily accessible place. This paper proposes an archetype describing how a knowledge management system, with the support of intranet capabilities, could very much increase the accuracy of capturing, storing and retrieving knowledge based processes thereby increasing the efficiency of the system. This system will expect a critical mass of usage, by the users, for intranet to function as knowledge management systems. This prototype would lead to a design of an application that would impose creation and maintenance of an effective knowledge management system through intranet. The aim of this paper is to introduce an effective system to handle capture, store and distribute knowledge management in a form that may not lead to any failure which exists in most of the systems. The methodology used in the system would require all the employees, in the organization, to contribute the maximum to deliver the system to a successful arena. The system is still in its initial mode and thereby the authors are under the process to practically implement the ideas, as mentioned in the system, to produce satisfactory results.Keywords: Knowledge Management Systems, Intranet, Methodology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2001168 Requirements Driven Multiple View Paradigm for Developing Security Architecture
Authors: K. Chandra Sekaran
Abstract:
This paper describes a paradigmatic approach to develop architecture of secure systems by describing the requirements from four different points of view: that of the owner, the administrator, the user, and the network. Deriving requirements and developing architecture implies the joint elicitation and describing the problem and the structure of the solution. The view points proposed in this paper are those we consider as requirements towards their contributions as major parties in the design, implementation, usage and maintenance of secure systems. The dramatic growth of the technology of Internet and the applications deployed in World Wide Web have lead to the situation where the security has become a very important concern in the development of secure systems. Many security approaches are currently being used in organizations. In spite of the widespread use of many different security solutions, the security remains a problem. It is argued that the approach that is described in this paper for the development of secure architecture is practical by all means. The models representing these multiple points of view are termed the requirements model (views of owner and administrator) and the operations model (views of user and network). In this paper, this multiple view paradigm is explained by first describing the specific requirements and or characteristics of secure systems (particularly in the domain of networks) and the secure architecture / system development methodology.
Keywords: Multiple view paradigms, requirements model, operations model, secure system, owner, administrator, user, network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1372167 Object Identification with Color, Texture, and Object-Correlation in CBIR System
Authors: Awais Adnan, Muhammad Nawaz, Sajid Anwar, Tamleek Ali, Muhammad Ali
Abstract:
Needs of an efficient information retrieval in recent years in increased more then ever because of the frequent use of digital information in our life. We see a lot of work in the area of textual information but in multimedia information, we cannot find much progress. In text based information, new technology of data mining and data marts are now in working that were started from the basic concept of database some where in 1960. In image search and especially in image identification, computerized system at very initial stages. Even in the area of image search we cannot see much progress as in the case of text based search techniques. One main reason for this is the wide spread roots of image search where many area like artificial intelligence, statistics, image processing, pattern recognition play their role. Even human psychology and perception and cultural diversity also have their share for the design of a good and efficient image recognition and retrieval system. A new object based search technique is presented in this paper where object in the image are identified on the basis of their geometrical shapes and other features like color and texture where object-co-relation augments this search process. To be more focused on objects identification, simple images are selected for the work to reduce the role of segmentation in overall process however same technique can also be applied for other images.Keywords: Object correlation, Geometrical shape, Color, texture, features, contents.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2031