Search results for: similarity measures
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4100

Search results for: similarity measures

4010 Genetic Characterization of Barley Genotypes via Inter-Simple Sequence Repeat

Authors: Mustafa Yorgancılar, Emine Atalay, Necdet Akgün, Ali Topal

Abstract:

In this study, polymerase chain reaction based Inter-simple sequence repeat (ISSR) from DNA fingerprinting techniques were used to investigate the genetic relationships among barley crossbreed genotypes in Turkey. It is important that selection based on the genetic base in breeding programs via ISSR, in terms of breeding time. 14 ISSR primers generated a total of 97 bands, of which 81 (83.35%) were polymorphic. The highest total resolution power (RP) value was obtained from the F2 (0.53) and M16 (0.51) primers. According to the ISSR result, the genetic similarity index changed between 0.64–095; Lane 3 with Line 6 genotypes were the closest, while Line 36 were the most distant ones. The ISSR markers were found to be promising for assessing genetic diversity in barley crossbreed genotypes.

Keywords: barley, crossbreed, genetic similarity, ISSR

Procedia PDF Downloads 311
4009 An Integrated Fuzzy Inference System and Technique for Order of Preference by Similarity to Ideal Solution Approach for Evaluation of Lean Healthcare Systems

Authors: Aydin M. Torkabadi, Ehsan Pourjavad

Abstract:

A decade after the introduction of Lean in Saskatchewan’s public healthcare system, its effectiveness remains a controversial subject among health researchers, workers, managers, and politicians. Therefore, developing a framework to quantitatively assess the Lean achievements is significant. This study investigates the success of initiatives across Saskatchewan health regions by recognizing the Lean healthcare criteria, measuring the success levels, comparing the regions, and identifying the areas for improvements. This study proposes an integrated intelligent computing approach by applying Fuzzy Inference System (FIS) and Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS). FIS is used as an efficient approach to assess the Lean healthcare criteria, and TOPSIS is applied for ranking the values in regards to the level of leanness. Due to the innate uncertainty in decision maker judgments on criteria, principals of the fuzzy theory are applied. Finally, FIS-TOPSIS was established as an efficient technique in determining the lean merit in healthcare systems.

Keywords: lean healthcare, intelligent computing, fuzzy inference system, healthcare evaluation, technique for order of preference by similarity to ideal solution, multi-criteria decision making, MCDM

Procedia PDF Downloads 135
4008 Monitoring Systemic Risk in the Hedge Fund Sector

Authors: Frank Hespeler, Giuseppe Loiacono

Abstract:

We propose measures for systemic risk generated through intra-sectorial interdependencies in the hedge fund sector. These measures are based on variations in the average cross-effects of funds showing significant interdependency between their individual returns and the moments of the sector’s return distribution. The proposed measures display a high ability to identify periods of financial distress, are robust to modifications in the underlying econometric model and are consistent with intuitive interpretation of the results.

Keywords: hedge funds, systemic risk, vector autoregressive model, risk monitoring

Procedia PDF Downloads 295
4007 Risk Based on Computer Auditing and Measures of ‎Prevention

Authors: Mohammad Hadi Khorashadi Zadeh, Amin Karkon, Seyd Mohammad Reza Mashhoori

Abstract:

The technology of computer audit played a major role in the progress and prospects of a proper application to improve the quality and efficiency of audit work. But due to the technical complexity and the specific risks of computer audit, it should be shown effective in audit and preventive action. Mainly through research in this paper, we propose the causes of audit risk in a computer environment and the risk of further proposals for measures to control, to some extent reduce the risk of computer audit and improve the audit quality.

Keywords: computer auditing, risk, measures to prevent, information management

Procedia PDF Downloads 484
4006 A System Dynamics Model for Assessment of Alternative Energy Policy Measures: A Case of Energy Management System as an Energy Efficiency Policy Tool

Authors: Andra Blumberga, Uldis Bariss, Anna Kubule, Dagnija Blumberga

Abstract:

European Union Energy Efficiency Directive provides a set of binding energy efficiency measures to reach. Each of the member states can use either energy efficiency obligation scheme or alternative policy measures or combination of both. Latvian government has decided to divide savings among obligation scheme (65%) and alternative measures (35%). This decision might lead to significant energy tariff increase hence impact on the national economy. To assess impact of alternative policy measures focusing on energy management scheme based on ISO 50001 and ability to decrease share of obligation scheme a System Dynamics modeling was used. Simulation results show that energy efficiency goal can be met with alternative policy measure to large energy consumers in industrial, tertiary and public sectors by applying the energy tax exemption for implementers of energy management system. A delay in applying alternative policy measures plays very important role in reaching the energy efficiency goal. One year delay in implementation of this policy measure reduces cumulative energy savings from 2016 to 2017 from 5200 GWh to 3000 GWh in 2020.

Keywords: system dynamics, energy efficiency, policy measure, energy management system, obligation scheme

Procedia PDF Downloads 252
4005 Semantic-Based Collaborative Filtering to Improve Visitor Cold Start in Recommender Systems

Authors: Baba Mbaye

Abstract:

In collaborative filtering recommendation systems, a user receives suggested items based on the opinions and evaluations of a community of users. This type of recommendation system uses only the information (notes in numerical values) contained in a usage matrix as input data. This matrix can be constructed based on users' behaviors or by offering users to declare their opinions on the items they know. The cold start problem leads to very poor performance for new users. It is a phenomenon that occurs at the beginning of use, in the situation where the system lacks data to make recommendations. There are three types of cold start problems: cold start for a new item, a new system, and a new user. We are interested in this article at the cold start for a new user. When the system welcomes a new user, the profile exists but does not have enough data, and its communities with other users profiles are still unknown. This leads to recommendations not adapted to the profile of the new user. In this paper, we propose an approach that improves cold start by using the notions of similarity and semantic proximity between users profiles during cold start. We will use the cold-metadata available (metadata extracted from the new user's data) useful in positioning the new user within a community. The aim is to look for similarities and semantic proximities with the old and current user profiles of the system. Proximity is represented by close concepts considered to belong to the same group, while similarity groups together elements that appear similar. Similarity and proximity are two close but not similar concepts. This similarity leads us to the construction of similarity which is based on: a) the concepts (properties, terms, instances) independent of ontology structure and, b) the simultaneous representation of the two concepts (relations, presence of terms in a document, simultaneous presence of the authorities). We propose an ontology, OIVCSRS (Ontology of Improvement Visitor Cold Start in Recommender Systems), in order to structure the terms and concepts representing the meaning of an information field, whether by the metadata of a namespace, or the elements of a knowledge domain. This approach allows us to automatically attach the new user to a user community, partially compensate for the data that was not initially provided and ultimately to associate a better first profile with the cold start. Thus, the aim of this paper is to propose an approach to improving cold start using semantic technologies.

Keywords: visitor cold start, recommender systems, collaborative filtering, semantic filtering

Procedia PDF Downloads 195
4004 Resources-Based Ontology Matching to Access Learning Resources

Authors: A. Elbyed

Abstract:

Nowadays, ontologies are used for achieving a common understanding within a user community and for sharing domain knowledge. However, the de-centralized nature of the web makes indeed inevitable that small communities will use their own ontologies to describe their data and to index their own resources. Certainly, accessing to resources from various ontologies created independently is an important challenge for answering end user queries. Ontology mapping is thus required for combining ontologies. However, mapping complete ontologies at run time is a computationally expensive task. This paper proposes a system in which mappings between concepts may be generated dynamically as the concepts are encountered during user queries. In this way, the interaction itself defines the context in which small and relevant portions of ontologies are mapped. We illustrate application of the proposed system in the context of Technology Enhanced Learning (TEL) where learners need to access to learning resources covering specific concepts.

Keywords: resources query, ontologies, ontology mapping, similarity measures, semantic web, e-learning

Procedia PDF Downloads 284
4003 Human Rights and Counter-Terrorism in Nigeria: A Systematic Review

Authors: Tarela J. Ike

Abstract:

Over the years, the hemorrhagic acts of Boko Haram have led to the adoption of counter-terrorism measures which mostly takes the form of military repressive measures. These measures have wrought flagrant violation of human rights worthy of concern. Hence, the need to examine the efficacy of the counter-terrorism measures adopted by the Nigeria government in combatting terrorism. This article addresses this issue by relying on a systematic literature review which examines the impact of Nigeria counter-terrorism measures from 2009 to 2016 in combating terrorism. The review of literature includes 42 article. Of the 42 articles, 14 met the peer-reviewed requirement which finds that most of Nigeria’s counter-terrorism policies are geared toward the use of state repressive military approach which violates the human right. Thus, the study concludes that to effectively address the terrorist uprising; Nigeria should adopt a non-aggressive counter-terrorism approach which incorporates religious clerics, and community active engagement strategy in combatting terrorism as opposed to military retaliation which violates human right and so far proved ineffective.

Keywords: Boko Haram, counter-terrorism, human rights, military retaliation

Procedia PDF Downloads 383
4002 Using of the Fractal Dimensions for the Analysis of Hyperkinetic Movements in the Parkinson's Disease

Authors: Sadegh Marzban, Mohamad Sobhan Sheikh Andalibi, Farnaz Ghassemi, Farzad Towhidkhah

Abstract:

Parkinson's disease (PD), which is characterized by the tremor at rest, rigidity, akinesia or bradykinesia and postural instability, affects the quality of life of involved individuals. The concept of a fractal is most often associated with irregular geometric objects that display self-similarity. Fractal dimension (FD) can be used to quantify the complexity and the self-similarity of an object such as tremor. In this work, we are aimed to propose a new method for evaluating hyperkinetic movements such as tremor, by using the FD and other correlated parameters in patients who are suffered from PD. In this study, we used 'the tremor data of Physionet'. The database consists of fourteen participants, diagnosed with PD including six patients with high amplitude tremor and eight patients with low amplitude. We tried to extract features from data, which can distinguish between patients before and after medication. We have selected fractal dimensions, including correlation dimension, box dimension, and information dimension. Lilliefors test has been used for normality test. Paired t-test or Wilcoxon signed rank test were also done to find differences between patients before and after medication, depending on whether the normality is detected or not. In addition, two-way ANOVA was used to investigate the possible association between the therapeutic effects and features extracted from the tremor. Just one of the extracted features showed significant differences between patients before and after medication. According to the results, correlation dimension was significantly different before and after the patient's medication (p=0.009). Also, two-way ANOVA demonstrates significant differences just in medication effect (p=0.033), and no significant differences were found between subject's differences (p=0.34) and interaction (p=0.97). The most striking result emerged from the data is that correlation dimension could quantify medication treatment based on tremor. This study has provided a technique to evaluate a non-linear measure for quantifying medication, nominally the correlation dimension. Furthermore, this study supports the idea that fractal dimension analysis yields additional information compared with conventional spectral measures in the detection of poor prognosis patients.

Keywords: correlation dimension, non-linear measure, Parkinson’s disease, tremor

Procedia PDF Downloads 219
4001 Improving Topic Quality of Scripts by Using Scene Similarity Based Word Co-Occurrence

Authors: Yunseok Noh, Chang-Uk Kwak, Sun-Joong Kim, Seong-Bae Park

Abstract:

Scripts are one of the basic text resources to understand broadcasting contents. Since broadcast media wields lots of influence over the public, tools for understanding broadcasting contents are more required. Topic modeling is the method to get the summary of the broadcasting contents from its scripts. Generally, scripts represent contents descriptively with directions and speeches. Scripts also provide scene segments that can be seen as semantic units. Therefore, a script can be topic modeled by treating a scene segment as a document. Because scripts consist of speeches mainly, however, relatively small co-occurrences among words in the scene segments are observed. This causes inevitably the bad quality of topics based on statistical learning method. To tackle this problem, we propose a method of learning with additional word co-occurrence information obtained using scene similarities. The main idea of improving topic quality is that the information that two or more texts are topically related can be useful to learn high quality of topics. In addition, by using high quality of topics, we can get information more accurate whether two texts are related or not. In this paper, we regard two scene segments are related if their topical similarity is high enough. We also consider that words are co-occurred if they are in topically related scene segments together. In the experiments, we showed the proposed method generates a higher quality of topics from Korean drama scripts than the baselines.

Keywords: broadcasting contents, scripts, text similarity, topic model

Procedia PDF Downloads 291
4000 Application of KL Divergence for Estimation of Each Metabolic Pathway Genes

Authors: Shohei Maruyama, Yasuo Matsuyama, Sachiyo Aburatani

Abstract:

The development of the method to annotate unknown gene functions is an important task in bioinformatics. One of the approaches for the annotation is The identification of the metabolic pathway that genes are involved in. Gene expression data have been utilized for the identification, since gene expression data reflect various intracellular phenomena. However, it has been difficult to estimate the gene function with high accuracy. It is considered that the low accuracy of the estimation is caused by the difficulty of accurately measuring a gene expression. Even though they are measured under the same condition, the gene expressions will vary usually. In this study, we proposed a feature extraction method focusing on the variability of gene expressions to estimate the genes' metabolic pathway accurately. First, we estimated the distribution of each gene expression from replicate data. Next, we calculated the similarity between all gene pairs by KL divergence, which is a method for calculating the similarity between distributions. Finally, we utilized the similarity vectors as feature vectors and trained the multiclass SVM for identifying the genes' metabolic pathway. To evaluate our developed method, we applied the method to budding yeast and trained the multiclass SVM for identifying the seven metabolic pathways. As a result, the accuracy that calculated by our developed method was higher than the one that calculated from the raw gene expression data. Thus, our developed method combined with KL divergence is useful for identifying the genes' metabolic pathway.

Keywords: metabolic pathways, gene expression data, microarray, Kullback–Leibler divergence, KL divergence, support vector machines, SVM, machine learning

Procedia PDF Downloads 376
3999 Risk-Based Computer Auditing and Measures of Prevention

Authors: Mohammad Hadi Khorashadi Zadeh, Amin Karkon, Seyd Mohammad Reza Mashhoori ‎

Abstract:

the technology of Computer audit played a major role in the progress and ‎prospects of a proper application to improve the quality and efficiency of audit ‎work. But due to the technical complexity and the specific risks of computer ‎audit, it should be shown effective in audit and preventive action. Mainly through ‎research in this paper, we proposes the causes of audit risk in a computer ‎environment and the risk of further proposals for measures to control, to some ‎extent reduce the risk of computer audit and improve the audit quality.‎

Keywords: computer auditing, risk, measures to prevent, information technology

Procedia PDF Downloads 453
3998 Empirical Exploration for the Correlation between Class Object-Oriented Connectivity-Based Cohesion and Coupling

Authors: Jehad Al Dallal

Abstract:

Attributes and methods are the basic contents of an object-oriented class. The connectivity among these class members and the relationship between the class and other classes play an important role in determining the quality of an object-oriented system. Class cohesion evaluates the degree of relatedness of class attributes and methods, whereas class coupling refers to the degree to which a class is related to other classes. Researchers have proposed several class cohesion and class coupling measures. However, the correlation between class coupling and class cohesion measures have not been thoroughly studied. In this paper, using classes of three open-source Java systems, we empirically investigate the correlation between several measures of connectivity-based class cohesion and coupling. Four connectivity-based cohesion measures and eight coupling measures are considered in the empirical study. The empirical study results show that class connectivity-based cohesion and coupling internal quality attributes are inversely correlated. The strength of the correlation depends highly on the cohesion and coupling measurement approaches.

Keywords: object-oriented class, software quality, class cohesion measure, class coupling measure

Procedia PDF Downloads 292
3997 Russian ‘Active Measures’: An Applicable Supporting Tool for Russia`s Foreign Policy Objectives in the 21st Century

Authors: Håkon Riiber

Abstract:

This paper explores the extent to which Russian ‘Active Measures’ play a role in contemporary Russian foreign policy and in what way the legacy of the Soviet Union is still apparent in these practices. The analysis draws on a set of case studies from the 21st century to examine these aspects, showing which ‘Active Measures’ features are old and which are new in the post-Cold War era. The paper highlights that the topic has gained significant academic and political interest in recent years, largely due to the aggressive posture of the Russian Federation on the world stage, exemplified through interventions in Estonia, Georgia, and Ukraine and interference in several democratic elections in the West. However, the paper argues that the long-term impact of these measures may have unintended implications for Russia. While Russia is unlikely to stop using Active Measures, increased awareness of the exploitation of weaknesses, institutions, or other targets may lead to greater security measures and an ability to identify and defend against these activities. The paper contends that Soviet-style ‘Active Measures’ from the Cold War era have been modernized and are now utilized to create an advantageous atmosphere for further exploitation to support contemporary Russian foreign policy. It offers three key points to support this argument: the reenergized legacy of the Cold War era, the use of ‘Active Measures’ in a number of cases in the 21st century, and the applicability of AM to the Russian approach to foreign policy. The analysis reveals that while this is not a new Russian phenomenon, it is still oversimplified and inaccurately understood by the West, which may result in a decreased ability to defend against these activities and limit the unwarranted escalation of the ongoing security situation between the West and Russia. The paper concludes that the legacy of Soviet-era Active Measures continues to influence Russian foreign policy, and modern technological advances have only made them more applicable to the current political climate. Overall, this paper sheds light on the important issue of Russian ‘Active Measures’ and the role they play in contemporary Russian foreign policy. It emphasizes the need for increased awareness, understanding, and security measures to defend against these activities and prevent further escalation of the security situation between the West and Russia.

Keywords: Russian espionage, active measures, disinformation, Russian intelligence

Procedia PDF Downloads 64
3996 A Tool for Assessing Performance and Structural Quality of Business Process

Authors: Mariem Kchaou, Wiem Khlif, Faiez Gargouri

Abstract:

Modeling business processes is an essential task when evaluating, improving, or documenting existing business processes. To be efficient in such tasks, a business process model (BPM) must have high structural quality and high performance. Evidently, evaluating the performance of a business process model is a necessary step to reduce time, cost, while assessing the structural quality aims to improve the understandability and the modifiability of the BPMN model. To achieve these objectives, a set of structural and performance measures have been proposed. Since the diversity of measures, we propose a framework that integrates both structural and performance aspects for classifying them. Our measure classification is based on business process model perspectives (e.g., informational, functional, organizational, behavioral, and temporal), and the elements (activity, event, actor, etc.) involved in computing the measures. Then, we implement this framework in a tool assisting the structural quality and the performance of a business process. The tool helps the designers to select an appropriate subset of measures associated with the corresponding perspective and to calculate and interpret their values in order to improve the structural quality and the performance of the model.

Keywords: performance, structural quality, perspectives, tool, classification framework, measures

Procedia PDF Downloads 122
3995 Production Structures of Energy Based on Water Force, Its Infrastructure Protection, and Possible Causes of Failure

Authors: Gabriela-Andreea Despescu, Mădălina-Elena Mavrodin, Gheorghe Lăzăroiu, Florin Adrian Grădinaru

Abstract:

The purpose of this paper is to contribute to the enhancement of a hydroelectric plant protection by coordinating protection measures and existing security and introducing new measures under a risk management process. Also, the plan identifies key critical elements of a hydroelectric plant, from its level vulnerabilities and threats it is subjected to in order to achieve the necessary protection measures to reduce the level of risk.

Keywords: critical infrastructure, risk analysis, critical infrastructure protection, vulnerability, risk management, turbine, impact analysis

Procedia PDF Downloads 512
3994 Decoding Gender Disparities in AI: An Experimental Exploration Within the Realm of AI and Trust Building

Authors: Alexander Scott English, Yilin Ma, Xiaoying Liu

Abstract:

The widespread use of artificial intelligence in everyday life has triggered a fervent discussion covering a wide range of areas. However, to date, research on the influence of gender in various segments and factors from a social science perspective is still limited. This study aims to explore whether there are gender differences in human trust in AI for its application in basic everyday life and correlates with human perceived similarity, perceived emotions (including competence and warmth), and attractiveness. We conducted a study involving 321 participants using a two-subject experimental design with a two-factor (masculinized vs. feminized voice of the AI) multiplied by a two-factor (pitch level of the AI's voice) between-subject experimental design. Four contexts were created for the study and randomly assigned. The results of the study showed significant gender differences in perceived similarity, trust, and perceived emotion of the AIs, with females rating them significantly higher than males. Trust was higher in relation to AIs presenting the same gender (e.g., human female to female AI, human male to male AI). Mediation modeling tests indicated that emotion perception and similarity played a sufficiently mediating role in trust. Notably, although trust in AIs was strongly correlated with human gender, there was no significant effect on the gender of the AI. In addition, the study discusses the effects of subjects' age, job search experience, and job type on the findings.

Keywords: artificial intelligence, gender differences, human-robot trust, mediation modeling

Procedia PDF Downloads 19
3993 Plagiarism Detection for Flowchart and Figures in Texts

Authors: Ahmadu Maidorawa, Idrissa Djibo, Muhammad Tella

Abstract:

This paper presents a method for detecting flow chart and figure plagiarism based on shape of image processing and multimedia retrieval. The method managed to retrieve flowcharts with ranked similarity according to different matching sets. Plagiarism detection is well known phenomenon in the academic arena. Copying other people is considered as serious offense that needs to be checked. There are many plagiarism detection systems such as turn-it-in that has been developed to provide these checks. Most, if not all, discard the figures and charts before checking for plagiarism. Discarding the figures and charts result in look holes that people can take advantage. That means people can plagiarize figures and charts easily without the current plagiarism systems detecting it. There are very few papers which talks about flowcharts plagiarism detection. Therefore, there is a need to develop a system that will detect plagiarism in figures and charts.

Keywords: flowchart, multimedia retrieval, figures similarity, image comparison, figure retrieval

Procedia PDF Downloads 433
3992 Human-Elephant Conflict and Mitigation Measures in Buffer Zone of Bardia National Park, Nepal

Authors: Rabin Paudel, Dambar Bahadur Mahato, Prabin Poudel, Bijaya Neupane, Sakar Jha

Abstract:

Understanding Human-Elephant Conflict (HEC) is very important in countries like Nepal, where solutions to escalating conflicts are urgently required. However, most of the HEC mitigation measures implemented so far have been done on an ad hoc basis without the detailed understanding of nature and extent of the damage. This study aims to assess the current scenario of HEC in regards to crop and property damages by Wild Asian Elephant and people’s perception towards existing mitigating measures and elephant conservation in Buffer zone area of Bardia National Park. The methods used were a questionnaire survey (N= 178), key-informant interview (N= 18) and focal group discussions (N= 6). Descriptive statistics were used to determine the nature and extent of damage and to understand people’s perception towards HEC, its mitigation measures and elephant conservation. Chi-square test was applied to determine the significance of crop and property damages with respect to distance from the park boundary. Out of all types of damage, crop damage was found to be the highest (51%), followed by house damage (31%) and damage to stored grains (18%) with winter being the season with the greatest elephant damage. Among 178 respondents, the majority of them (82%) were positive towards elephant conservation despite the increment in HEC incidents as perceived by 88% of total respondents. Among the mitigation measures present, the most applied was electric fence (91%) followed by barbed wire fence (5%), reinforced concrete cement wall (3%) and gabion wall (1%). Most effective mitigation measures were reinforced concrete cement wall and gabion wall. To combat increasing crop damage, the insurance policy should be initiated. The efficiency of the mitigation measures should be timely monitored, and corrective measures should be applied as per the need.

Keywords: crop and property damage, elephant conflict, Asiatic wild elephant, mitigation measures

Procedia PDF Downloads 120
3991 Development of a Rating Scale for Elementary EFL Writing

Authors: Mohammed S. Assiri

Abstract:

In EFL programs, rating scales used in writing assessment are often constructed by intuition. Intuition-based scales tend to provide inaccurate and divisive ratings of learners’ writing performance. Hence, following an empirical approach, this study attempted to develop a rating scale for elementary-level writing at an EFL program in Saudi Arabia. Towards this goal, 98 students’ essays were scored and then coded using comprehensive taxonomy of writing constructs and their measures. An automatic linear modeling was run to find out which measures would best predict essay scores. A nonparametric ANOVA, the Kruskal-Wallis test, was then used to determine which measures could best differentiate among scoring levels. Findings indicated that there were certain measures that could serve as either good predictors of essay scores or differentiators among scoring levels, or both. The main conclusion was that a rating scale can be empirically developed using predictive and discriminative statistical tests.

Keywords: analytic scoring, rating scales, writing assessment, writing constructs, writing performance

Procedia PDF Downloads 433
3990 Web Proxy Detection via Bipartite Graphs and One-Mode Projections

Authors: Zhipeng Chen, Peng Zhang, Qingyun Liu, Li Guo

Abstract:

With the Internet becoming the dominant channel for business and life, many IPs are increasingly masked using web proxies for illegal purposes such as propagating malware, impersonate phishing pages to steal sensitive data or redirect victims to other malicious targets. Moreover, as Internet traffic continues to grow in size and complexity, it has become an increasingly challenging task to detect the proxy service due to their dynamic update and high anonymity. In this paper, we present an approach based on behavioral graph analysis to study the behavior similarity of web proxy users. Specifically, we use bipartite graphs to model host communications from network traffic and build one-mode projections of bipartite graphs for discovering social-behavior similarity of web proxy users. Based on the similarity matrices of end-users from the derived one-mode projection graphs, we apply a simple yet effective spectral clustering algorithm to discover the inherent web proxy users behavior clusters. The web proxy URL may vary from time to time. Still, the inherent interest would not. So, based on the intuition, by dint of our private tools implemented by WebDriver, we examine whether the top URLs visited by the web proxy users are web proxies. Our experiment results based on real datasets show that the behavior clusters not only reduce the number of URLs analysis but also provide an effective way to detect the web proxies, especially for the unknown web proxies.

Keywords: bipartite graph, one-mode projection, clustering, web proxy detection

Procedia PDF Downloads 220
3989 Unsupervised Assistive and Adaptative Intelligent Agent in Smart Enviroment

Authors: Sebastião Pais, João Casal, Ricardo Ponciano, Sérgio Lorenço

Abstract:

The adaptation paradigm is a basic defining feature for pervasive computing systems. Adaptation systems must work efficiently in a smart environment while providing suitable information relevant to the user system interaction. The key objective is to deduce the information needed information changes. Therefore relying on fixed operational models would be inappropriate. This paper presents a study on developing an Intelligent Personal Assistant to assist the user in interacting with their Smart Environment. We propose an Unsupervised and Language-Independent Adaptation through Intelligent Speech Interface and a set of methods of Acquiring Knowledge, namely Semantic Similarity and Unsupervised Learning.

Keywords: intelligent personal assistants, intelligent speech interface, unsupervised learning, language-independent, knowledge acquisition, association measures, symmetric word similarities, attributional word similarities

Procedia PDF Downloads 529
3988 Unsupervised Assistive and Adaptive Intelligent Agent in Smart Environment

Authors: Sebastião Pais, João Casal, Ricardo Ponciano, Sérgio Lourenço

Abstract:

The adaptation paradigm is a basic defining feature for pervasive computing systems. Adaptation systems must work efficiently in smart environment while providing suitable information relevant to the user system interaction. The key objective is to deduce the information needed information changes. Therefore, relying on fixed operational models would be inappropriate. This paper presents a study on developing a Intelligent Personal Assistant to assist the user in interacting with their Smart Environment. We propose a Unsupervised and Language-Independent Adaptation through Intelligent Speech Interface and a set of methods of Acquiring Knowledge, namely Semantic Similarity and Unsupervised Learning.

Keywords: intelligent personal assistants, intelligent speech interface, unsupervised learning, language-independent, knowledge acquisition, association measures, symmetric word similarities, attributional word similarities

Procedia PDF Downloads 604
3987 Lacunarity measures on Mammographic Image Applying Fractal Dimension and Lacunarity Measures

Authors: S. Sushma, S. Balasubramanian, K. C. Latha, R. Sridhar

Abstract:

Structural texture measures are used to address the aspect of breast cancer risk assessment in screening mammograms. The current study investigates whether texture properties characterized by local Fractal Dimension (FD) and lacunarity contribute to assess breast cancer risk. Fractal Dimension represents the complexity while the lacunarity characterize the gap of a fractal dimension. In this paper, we present our result confirming that the lacunarity value resulted in algorithm using mammogram images states that level of lacunarity will be low when the Fractal Dimension value will be high.

Keywords: breast cancer, fractal dimension, image analysis, lacunarity, mammogram

Procedia PDF Downloads 351
3986 Comparison of Crossover Types to Obtain Optimal Queries Using Adaptive Genetic Algorithm

Authors: Wafa’ Alma'Aitah, Khaled Almakadmeh

Abstract:

this study presents an information retrieval system of using genetic algorithm to increase information retrieval efficiency. Using vector space model, information retrieval is based on the similarity measurement between query and documents. Documents with high similarity to query are judge more relevant to the query and should be retrieved first. Using genetic algorithms, each query is represented by a chromosome; these chromosomes are fed into genetic operator process: selection, crossover, and mutation until an optimized query chromosome is obtained for document retrieval. Results show that information retrieval with adaptive crossover probability and single point type crossover and roulette wheel as selection type give the highest recall. The proposed approach is verified using (242) proceedings abstracts collected from the Saudi Arabian national conference.

Keywords: genetic algorithm, information retrieval, optimal queries, crossover

Procedia PDF Downloads 261
3985 Experimental Study Analyzing the Similarity Theory Formulations for the Effect of Aerodynamic Roughness Length on Turbulence Length Scales in the Atmospheric Surface Layer

Authors: Matthew J. Emes, Azadeh Jafari, Maziar Arjomandi

Abstract:

Velocity fluctuations of shear-generated turbulence are largest in the atmospheric surface layer (ASL) of nominal 100 m depth, which can lead to dynamic effects such as galloping and flutter on small physical structures on the ground when the turbulence length scales and characteristic length of the physical structure are the same order of magnitude. Turbulence length scales are a measure of the average sizes of the energy-containing eddies that are widely estimated using two-point cross-correlation analysis to convert the temporal lag to a separation distance using Taylor’s hypothesis that the convection velocity is equal to the mean velocity at the corresponding height. Profiles of turbulence length scales in the neutrally-stratified ASL, as predicted by Monin-Obukhov similarity theory in Engineering Sciences Data Unit (ESDU) 85020 for single-point data and ESDU 86010 for two-point correlations, are largely dependent on the aerodynamic roughness length. Field measurements have shown that longitudinal turbulence length scales show significant regional variation, whereas length scales of the vertical component show consistent Obukhov scaling from site to site because of the absence of low-frequency components. Hence, the objective of this experimental study is to compare the similarity theory relationships between the turbulence length scales and aerodynamic roughness length with those calculated using the autocorrelations and cross-correlations of field measurement velocity data at two sites: the Surface Layer Turbulence and Environmental Science Test (SLTEST) facility in a desert ASL in Dugway, Utah, USA and the Commonwealth Scientific and Industrial Research Organisation (CSIRO) wind tower in a rural ASL in Jemalong, NSW, Australia. The results indicate that the longitudinal turbulence length scales increase with increasing aerodynamic roughness length, as opposed to the relationships derived by similarity theory correlations in ESDU models. However, the ratio of the turbulence length scales in the lateral and vertical directions to the longitudinal length scales is relatively independent of surface roughness, showing consistent inner-scaling between the two sites and the ESDU correlations. Further, the diurnal variation of wind velocity due to changes in atmospheric stability conditions has a significant effect on the turbulence structure of the energy-containing eddies in the lower ASL.

Keywords: aerodynamic roughness length, atmospheric surface layer, similarity theory, turbulence length scales

Procedia PDF Downloads 101
3984 Theoretical Exploration for the Impact of Accounting for Special Methods in Connectivity-Based Cohesion Measurement

Authors: Jehad Al Dallal

Abstract:

Class cohesion is a key object-oriented software quality attribute that is used to evaluate the degree of relatedness of class attributes and methods. Researchers have proposed several class cohesion measures. However, the effect of considering the special methods (i.e., constructors, destructors, and access and delegation methods) in cohesion calculation is not thoroughly theoretically studied for most of them. In this paper, we address this issue for three popular connectivity-based class cohesion measures. For each of the considered measures we theoretically study the impact of including or excluding special methods on the values that are obtained by applying the measure. This study is based on analyzing the definitions and formulas that are proposed for the measures. The results show that including/excluding special methods has a considerable effect on the obtained cohesion values and that this effect varies from one measure to another. For each of the three connectivity-based measures, the proposed theoretical study recommended excluding the special methods in cohesion measurement.

Keywords: object-oriented class, software quality, class cohesion measure, class cohesion, special methods

Procedia PDF Downloads 264
3983 Generation of Photo-Mosaic Images through Block Matching and Color Adjustment

Authors: Hae-Yeoun Lee

Abstract:

Mosaic refers to a technique that makes image by gathering lots of small materials in various colours. This paper presents an automatic algorithm that makes the photomosaic image using photos. The algorithm is composed of four steps: Partition and feature extraction, block matching, redundancy removal and colour adjustment. The input image is partitioned in the small block to extract feature. Each block is matched to find similar photo in database by comparing similarity with Euclidean difference between blocks. The intensity of the block is adjusted to enhance the similarity of image by replacing the value of light and darkness with that of relevant block. Further, the quality of image is improved by minimizing the redundancy of tiles in the adjacent blocks. Experimental results support that the proposed algorithm is excellent in quantitative analysis and qualitative analysis.

Keywords: photomosaic, Euclidean distance, block matching, intensity adjustment

Procedia PDF Downloads 251
3982 Hybrid Reliability-Similarity-Based Approach for Supervised Machine Learning

Authors: Walid Cherif

Abstract:

Data mining has, over recent years, seen big advances because of the spread of internet, which generates everyday a tremendous volume of data, and also the immense advances in technologies which facilitate the analysis of these data. In particular, classification techniques are a subdomain of Data Mining which determines in which group each data instance is related within a given dataset. It is used to classify data into different classes according to desired criteria. Generally, a classification technique is either statistical or machine learning. Each type of these techniques has its own limits. Nowadays, current data are becoming increasingly heterogeneous; consequently, current classification techniques are encountering many difficulties. This paper defines new measure functions to quantify the resemblance between instances and then combines them in a new approach which is different from actual algorithms by its reliability computations. Results of the proposed approach exceeded most common classification techniques with an f-measure exceeding 97% on the IRIS Dataset.

Keywords: data mining, knowledge discovery, machine learning, similarity measurement, supervised classification

Procedia PDF Downloads 428
3981 The Importance of Information in Psychological Operations for Counterterrorism

Authors: Abbas Fazelinia

Abstract:

Terrorism is not a new phenomenon to the world, yet it remains difficult to define and to counter. Countering terrorism requires several measures that must be taken at the same time. Counterterrorism strategies of most countries depend on military measures. However, those strategies should also focus on nonlethal measures, such as economic, political, and social measures. The psychological dimensions of terrorism must be understood, evaluated, and used in countering terrorism. This study suggests that psychological operations, as nonlethal military operations, can be used to influence individuals not to join terrorist organizations and to facilitate defections from terrorist organizations. However, in order to implement effective psychological operations, one has to have appropriate intelligence about terrorist organizations. Examining terrorist organizations help us to identify their vulnerabilities and obtain this intelligence. This article concludes that terrorists’ motivations, terrorist organizations’ radicalization, recruitment, and conversion processes, ideology, goals, strategies, and general structure form the intelligence requirement for psychological operations in counterterrorism. The methodology used in this article is a mixed method.

Keywords: psychological operations, terrorist, counterterrorism, terrorism

Procedia PDF Downloads 301