Search results for: data source
28346 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis
Authors: C. B. Le, V. N. Pham
Abstract:
In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering
Procedia PDF Downloads 18928345 Priority of Goal Over Source in Persian Directional Motion Verbs
Authors: Tahereh Samenian
Abstract:
There is ample evidence that source and goal are disproportionately expressed in languages, and goal usually plays a more prominent role than source. The results show that the mismatch between the goal and the source is not entirely rooted in non-linguistic behaviors, i.e. that linguistic descriptions also show the focus of the goal on the source in events; Non-verbal memory for events, on the other hand, indicates that the focus of the goal is only on events that are purposefully moving and the actor is alive. In the present study, an attempt is made to examine the principle of priority of the goal over the source by focusing on Persian directional motion verbs. For this purpose, 117 Persian directional motion verbs have been selected from the dictionary and data for them have been collected from the body of Bijan Khan and the components of goal and source have been identified in sentences and the prominence of the components of goal and source has been shown in the form of diagrams. As it was obtained from the data, Persian motion-directional verbs also showed the bias of the goal over source in motion events.Keywords: motion-directional verbs, priority of goal over source principle, cognitive factors, linguistic factors
Procedia PDF Downloads 8728344 Semantic Data Schema Recognition
Authors: Aïcha Ben Salem, Faouzi Boufares, Sebastiao Correia
Abstract:
The subject covered in this paper aims at assisting the user in its quality approach. The goal is to better extract, mix, interpret and reuse data. It deals with the semantic schema recognition of a data source. This enables the extraction of data semantics from all the available information, inculding the data and the metadata. Firstly, it consists of categorizing the data by assigning it to a category and possibly a sub-category, and secondly, of establishing relations between columns and possibly discovering the semantics of the manipulated data source. These links detected between columns offer a better understanding of the source and the alternatives for correcting data. This approach allows automatic detection of a large number of syntactic and semantic anomalies.Keywords: schema recognition, semantic data profiling, meta-categorisation, semantic dependencies inter columns
Procedia PDF Downloads 41828343 Multi-Source Data Fusion for Urban Comprehensive Management
Authors: Bolin Hua
Abstract:
In city governance, various data are involved, including city component data, demographic data, housing data and all kinds of business data. These data reflects different aspects of people, events and activities. Data generated from various systems are different in form and data source are different because they may come from different sectors. In order to reflect one or several facets of an event or rule, data from multiple sources need fusion together. Data from different sources using different ways of collection raised several issues which need to be resolved. Problem of data fusion include data update and synchronization, data exchange and sharing, file parsing and entry, duplicate data and its comparison, resource catalogue construction. Governments adopt statistical analysis, time series analysis, extrapolation, monitoring analysis, value mining, scenario prediction in order to achieve pattern discovery, law verification, root cause analysis and public opinion monitoring. The result of Multi-source data fusion is to form a uniform central database, which includes people data, location data, object data, and institution data, business data and space data. We need to use meta data to be referred to and read when application needs to access, manipulate and display the data. A uniform meta data management ensures effectiveness and consistency of data in the process of data exchange, data modeling, data cleansing, data loading, data storing, data analysis, data search and data delivery.Keywords: multi-source data fusion, urban comprehensive management, information fusion, government data
Procedia PDF Downloads 39328342 UNIX Source Code Leak: Evaluation and Feasible Solutions
Authors: Gu Dongxing, Li Yuxuan, Nong Tengxiao, Burra Venkata Durga Kumar
Abstract:
Since computers are widely used in business models, more and more companies choose to store important information in computers to improve productivity. However, this information can be compromised in many cases, such as when it is stored locally on the company's computers or when it is transferred between servers and clients. Of these important information leaks, source code leaks are probably the most costly. Because the source code often represents the core technology of the company, especially for the Internet companies, source code leakage may even lead to the company's core products lose market competitiveness, and then lead to the bankruptcy of the company. In recent years, such as Microsoft, AMD and other large companies have occurred source code leakage events, suffered a huge loss. This reveals to us the importance and necessity of preventing source code leakage. This paper aims to find ways to prevent source code leakage based on the direction of operating system, and based on the fact that most companies use Linux or Linux-like system to realize the interconnection between server and client, to discuss how to reduce the possibility of source code leakage during data transmission.Keywords: data transmission, Linux, source code, operating system
Procedia PDF Downloads 27028341 Topological Sensitivity Analysis for Reconstruction of the Inverse Source Problem from Boundary Measurement
Authors: Maatoug Hassine, Mourad Hrizi
Abstract:
In this paper, we consider a geometric inverse source problem for the heat equation with Dirichlet and Neumann boundary data. We will reconstruct the exact form of the unknown source term from additional boundary conditions. Our motivation is to detect the location, the size and the shape of source support. We present a one-shot algorithm based on the Kohn-Vogelius formulation and the topological gradient method. The geometric inverse source problem is formulated as a topology optimization one. A topological sensitivity analysis is derived from a source function. Then, we present a non-iterative numerical method for the geometric reconstruction of the source term with unknown support using a level curve of the topological gradient. Finally, we give several examples to show the viability of our presented method.Keywords: geometric inverse source problem, heat equation, topological optimization, topological sensitivity, Kohn-Vogelius formulation
Procedia PDF Downloads 30028340 Noise Source Identification on Urban Construction Sites Using Signal Time Delay Analysis
Authors: Balgaisha G. Mukanova, Yelbek B. Utepov, Aida G. Nazarova, Alisher Z. Imanov
Abstract:
The problem of identifying local noise sources on a construction site using a sensor system is considered. Mathematical modeling of detected signals on sensors was carried out, considering signal decay and signal delay time between the source and detector. Recordings of noises produced by construction tools were used as a dependence of noise on time. Synthetic sensor data was constructed based on these data, and a model of the propagation of acoustic waves from a point source in the three-dimensional space was applied. All sensors and sources are assumed to be located in the same plane. A source localization method is checked based on the signal time delay between two adjacent detectors and plotting the direction of the source. Based on the two direct lines' crossline, the noise source's position is determined. Cases of one dominant source and the case of two sources in the presence of several other sources of lower intensity are considered. The number of detectors varies from three to eight detectors. The intensity of the noise field in the assessed area is plotted. The signal of a two-second duration is considered. The source is located for subsequent parts of the signal with a duration above 0.04 sec; the final result is obtained by computing the average value.Keywords: acoustic model, direction of arrival, inverse source problem, sound localization, urban noises
Procedia PDF Downloads 6228339 Requirement Engineering Within Open Source Software Development: A Case Study
Authors: Kars Beek, Remco Groeneveld, Sjaak Brinkkemper
Abstract:
Although there is much literature available on requirement documentation in traditional software development, few studies have been conducted about this topic in open source software development. While open-source software development is becoming more important, the software development processes are often not as structured as corporate software development processes. Papers show that communities, creating open-source software, often lack structure and documentation. However, most recent studies about this topic are often ten or more years old. Therefore, this research has been conducted to determine if the lack of structure and documentation in requirement engineering is currently still the situation in these communities. Three open-source products have been chosen as subjects for conducting this research. The data for this research was gathered based on interviews, observations, and analyses of feature proposals and issue tracking tools. In this paper, we present a comparison and an analysis of the different methods used for requirements documentation to understand the current practices of requirements documentation in open source software development.Keywords: case study, open source software, open source software development, requirement elicitation, requirement engineering
Procedia PDF Downloads 10328338 Analyzing Oil Seeps Manifestations and Petroleum Impregnation in Northwestern Tunisia From Aliphatic Biomarkers and Statistical Data
Authors: Sawsen Jarray, Tahani Hallek, Mabrouk Montacer
Abstract:
The tectonically damaged terrain in Tunisia's Northwest is seen in the country's numerous oil leaks. Finding a genetic link between these oil seeps and the area's putative source rocks is the goal of this investigation. Here, we use aliphatic biomarkers assessed by GC-MS to describe the organic geochemical data of 18 oil seeps samples and 4 source rocks (M'Cherga, Fahdene, Bahloul, and BouDabbous). In order to establish correlations between oil and oil and oil and source rock, terpanes, hopanes, and steranes biomarkers were identified. The source rocks under study were deposited in a marine environment and were suboxic, with minor signs of continental input for the M'Cherga Formation. There is no connection between the Fahdene and Bahloul source rocks and the udied oil seeps. According to the biomarkers C27 18-22,29,30trisnorneohopane (Ts) and C27 17-22,29,30-trisnorhopane (Tm), these source rocks are mature and have reached the oil window. Regarding oil seeps, geochemical data indicate that, with the exception of four samples that showed some continental markings, the bulk of samples were deposited in an open marine environment. These most recent samples from oil seeps have a unique lithology (marl) that distinguishes them from the others (carbonate). There are two classes of oil seeps, according to statistical analysis of relationships between oil and oil and oil and source rocks. The first comprised samples that showed a positive connection with carbonate-lithological and marine-derived BouDabbous black shales. The second is a result of M'Cherga source rock and is made up of oil seeps with remnants of the terrestrial environment and a lithology with a marl trend. The Fahdene and Bahloul source rocks have no connection to the observed oil seeps. There are two different types of hydrocarbon spills depending on their link to tectonic deformations (oil seeps) and outcropping mature source rocks (oil impregnations), in addition to the existence of two generations of hydrocarbon spills in Northwest Tunisia (Lower Cretaceous/Ypresian).Keywords: petroleum seeps, source rocks, biomarkers, statistic, Northern Tunisia
Procedia PDF Downloads 6928337 Source Separation for Global Multispectral Satellite Images Indexing
Authors: Aymen Bouzid, Jihen Ben Smida
Abstract:
In this paper, we propose to prove the importance of the application of blind source separation methods on remote sensing data in order to index multispectral images. The proposed method starts with Gabor Filtering and the application of a Blind Source Separation to get a more effective representation of the information contained on the observation images. After that, a feature vector is extracted from each image in order to index them. Experimental results show the superior performance of this approach.Keywords: blind source separation, content based image retrieval, feature extraction multispectral, satellite images
Procedia PDF Downloads 40328336 A Case Study of Open Source Development Practices within a Large Company Setting
Authors: Alma Orucevic-Alagic, Martin Höst
Abstract:
Open source communities have demonstrated that complex and enterprise grade software can be produced, supported, and maintained by self-organizing groups of developers using primarily electronic form of communication. Due to the inherent nature of open source development, a specific set of open source software development practices has evolved. While there is an ongoing research on the topic of applicability of open source development practices within a company setting, still little is known about their benefits and challenges. The objective of this research is to understand if and to what degree open source development practices observed within a mature open source community are aligned with development practices within a large software and hardware company setting. For the purpose of this case study a set of open source development practices that are present in a mature open source community has been identified. Then, development practices of a large, international, hardware and software company based in Sweden were assessed and compared to the identified open source community practices. It is shown that there are many similarities between a mature open source community and a large company setting in regard to software development practices. We also identify practices that exist in open source communities and that are not standard within a company setting, but whose implementation can result in an improved software development efficiency within the company setting.Keywords: development practices, open source software, innersource, closed open source
Procedia PDF Downloads 55828335 Estimation of Source Parameters Using Source Parameters Imaging Method From Digitised High Resolution Airborne Magnetic Data of a Basement Complex
Authors: O. T. Oluriz, O. D. Akinyemi, J. A.Olowofela, O. A. Idowu, S. A. Ganiyu
Abstract:
This study was carried out using aeromagnetic data which record variation in the magnitude of the earth magnetic field in order to detect local changes in the properties of the underlying geology. The aeromagnetic data (Sheet No. 261) was acquired from the archives of Nigeria Geological Survey Agency of Nigeria, obtained in 2009. The study present estimation of source parameters within an area of about 3,025 square kilometers on geographic latitude to and longitude to within Ibadan and it’s environs in Oyo State, southwestern Nigeria. The area under study belongs to part of basement complex in southwestern Nigeria. Estimation of source parameters of aeromagnetic data was achieve through the application of source imaging parameters (SPI) techniques that provide delineation, depth, dip contact, susceptibility contrast and mineral potentials of magnetic signatures within the region. The depth to the magnetic sources in the area ranges from 0.675 km to 4.48 km. The estimated depth limit to shallow sources is 0.695 km and depth to deep sources is 4.48 km. The apparent susceptibility values of the entire study area obtained ranges from 0.01 to 0.005 [SI]. This study has shown that the magnetic susceptibility within study area is controlled mainly by super paramagnetic minerals.Keywords: aeromagnetic, basement complex, meta-sediment, precambrian
Procedia PDF Downloads 43028334 A Proposal for U-City (Smart City) Service Method Using Real-Time Digital Map
Authors: SangWon Han, MuWook Pyeon, Sujung Moon, DaeKyo Seo
Abstract:
Recently, technologies based on three-dimensional (3D) space information are being developed and quality of life is improving as a result. Research on real-time digital map (RDM) is being conducted now to provide 3D space information. RDM is a service that creates and supplies 3D space information in real time based on location/shape detection. Research subjects on RDM include the construction of 3D space information with matching image data, complementing the weaknesses of image acquisition using multi-source data, and data collection methods using big data. Using RDM will be effective for space analysis using 3D space information in a U-City and for other space information utilization technologies.Keywords: RDM, multi-source data, big data, U-City
Procedia PDF Downloads 43328333 Preliminary Analysis for Oil and Gas Geological Characteristics and Exploration Prospects of Doseo Basin in Central Africa
Authors: Haiqiang Song, Huiqing Liu
Abstract:
The Doseo basin in Chad, Central Africa is one of the most important oil and gas blocks in the world. However, the low degree of oil and gas exploration and the lack of relevant geological data restrict the understanding and resource evaluation of the basin. To further develop the Doseo basin efficiently, it is urgent to deeply analyze the source rock characteristics and hydrocarbon generation potential of the Doseo basin. Based on seismic and drilling data in recent years, this paper systematically evaluates the geochemical characteristics of source rocks and their generated oils in Doseo Basin, explores the development, distribution, and evolution characteristics of source rocks, and evaluates the exploration potential of Doseo Basin according to the hydrocarbon enrichment law. The results show that the Lower Cretaceous Baliemian and Apudian source rocks in Doseo Basin are well developed, with high organic matter abundance (average TOC≥3%) and good organic matter types (type I~II), which are the main development layers of source rocks, but the organic matter maturity is generally low (Ro of the drilled source rocks is mainly between 0.4%~0.8%). The planar structure also shows that the main hydrocarbon accumulation mode in Doseo sag is the forward tectonic reservoirs such as near source anticlines and faulted noses. Finally, it is estimated that the accumulative resources of the main source rocks in the Doseo Basin are about 4.33× 108T in Apudite and Balim terrace layers. The results of this study will help guide the next step of oil and gas exploration, which is expected to drive the next step of oil and gas development.Keywords: Doseo basin, lower cretaceous, source rock characteristics, developmental characteristics, hydrocarbon generation potential
Procedia PDF Downloads 11128332 End to End Monitoring in Oracle Fusion Middleware for Data Verification
Authors: Syed Kashif Ali, Usman Javaid, Abdullah Chohan
Abstract:
In large enterprises multiple departments use different sort of information systems and databases according to their needs. These systems are independent and heterogeneous in nature and sharing information/data between these systems is not an easy task. The usage of middleware technologies have made data sharing between systems very easy. However, monitoring the exchange of data/information for verification purposes between target and source systems is often complex or impossible for maintenance department due to security/access privileges on target and source systems. In this paper, we are intended to present our experience of an end to end data monitoring approach at middle ware level implemented in Oracle BPEL for data verification without any help of monitoring tool.Keywords: service level agreement, SOA, BPEL, oracle fusion middleware, web service monitoring
Procedia PDF Downloads 48028331 The Comparison of Open Source Software for Digital Libraries
Authors: Kanita Beširević
Abstract:
Open-source software development activities highly rely on Internet gathering communities volunteering in software development projects. Additionally, the libraries and cultural institutions share their metadata in the form of linked metadata to enable dissemination and enrichment. The open-source software provides free alternatives to traditional software solutions. The article aims to investigate the ever-increasing options for the digital library open source software adoption. The software available is presented and compared to other software solutions as well as to their previous versions. The top three open-source digital library software solutions are presented and compared. The comparison criteria are adopted from the UNESCO study by Bankier, J., & Gleason, K. Institutional Repository Software Comparison comprising of twelve criteria to appraise software, namely: infrastructure, front-end design, content discovery, publication tools, interoperability, and preservation. This article adopts a descriptive methodology based on data and information collected through selected software websites and the literature review.Keywords: open source software, digital library, DSpace, Fedora, Greenstone
Procedia PDF Downloads 11628330 A Second Law Assessment of Organic Rankine Cycle Depending on Source Temperature
Authors: Kyoung Hoon Kim
Abstract:
Organic Rankine Cycle (ORC) has potential in reducing fossil fuels and relaxing environmental problems. In this work performance analysis of ORC is conducted based on the second law of thermodynamics for recovery of low temperature heat source from 100°C to 140°C using R134a as the working fluid. Effects of system parameters such as turbine inlet pressure or source temperature are theoretically investigated on the exergy destructions (anergies) at various components of the system as well as net work production or exergy efficiency. Results show that the net work or exergy efficiency has a peak with respect to the turbine inlet pressure when the source temperature is low, however, increases monotonically with increasing turbine inlet pressure when the source temperature is high.Keywords: Organic Rankine Cycle (ORC), low temperature heat source, exergy, source temperature
Procedia PDF Downloads 45128329 Transfer Knowledge From Multiple Source Problems to a Target Problem in Genetic Algorithm
Authors: Terence Soule, Tami Al Ghamdi
Abstract:
To study how to transfer knowledge from multiple source problems to the target problem, we modeled the Transfer Learning (TL) process using Genetic Algorithms as the model solver. TL is the process that aims to transfer learned data from one problem to another problem. The TL process aims to help Machine Learning (ML) algorithms find a solution to the problems. The Genetic Algorithms (GA) give researchers access to information that we have about how the old problem is solved. In this paper, we have five different source problems, and we transfer the knowledge to the target problem. We studied different scenarios of the target problem. The results showed combined knowledge from multiple source problems improves the GA performance. Also, the process of combining knowledge from several problems results in promoting diversity of the transferred population.Keywords: transfer learning, genetic algorithm, evolutionary computation, source and target
Procedia PDF Downloads 14028328 Evaluation of a Data Fusion Algorithm for Detecting and Locating a Radioactive Source through Monte Carlo N-Particle Code Simulation and Experimental Measurement
Authors: Hadi Ardiny, Amir Mohammad Beigzadeh
Abstract:
Through the utilization of a combination of various sensors and data fusion methods, the detection of potential nuclear threats can be significantly enhanced by extracting more information from different data. In this research, an experimental and modeling approach was employed to track a radioactive source by combining a surveillance camera and a radiation detector (NaI). To run this experiment, three mobile robots were utilized, with one of them equipped with a radioactive source. An algorithm was developed in identifying the contaminated robot through correlation between camera images and camera data. The computer vision method extracts the movements of all robots in the XY plane coordinate system, and the detector system records the gamma-ray count. The position of the robots and the corresponding count of the moving source were modeled using the MCNPX simulation code while considering the experimental geometry. The results demonstrated a high level of accuracy in finding and locating the target in both the simulation model and experimental measurement. The modeling techniques prove to be valuable in designing different scenarios and intelligent systems before initiating any experiments.Keywords: nuclear threats, radiation detector, MCNPX simulation, modeling techniques, intelligent systems
Procedia PDF Downloads 12328327 Employing a Knime-based and Open-source Tools to Identify AMI and VER Metabolites from UPLC-MS Data
Authors: Nouf Alourfi
Abstract:
This study examines the metabolism of amitriptyline (AMI) and verapamil (VER) using a KNIME-based method. KNIME improved workflow is an open-source data-analytics platform that integrates a number of open-source metabolomics tools such as CFMID and MetFrag to provide standard data visualisations, predict candidate metabolites, assess them against experimental data, and produce reports on identified metabolites. The use of this workflow is demonstrated by employing three types of liver microsomes (human, rat, and Guinea pig) to study the in vitro metabolism of the two drugs (AMI and VER). This workflow is used to create and treat UPLC-MS (Orbitrap) data. The formulas and structures of these drugs' metabolites can be assigned automatically. The key metabolic routes for amitriptyline are hydroxylation, N-dealkylation, N-oxidation, and conjugation, while N-demethylation, O-demethylation and N-dealkylation, and conjugation are the primary metabolic routes for verapamil. The identified metabolites are compatible to the published, clarifying the solidity of the workflow technique and the usage of computational tools like KNIME in supporting the integration and interoperability of emerging novel software packages in the metabolomics area.Keywords: KNIME, CFMID, MetFrag, Data Analysis, Metabolomics
Procedia PDF Downloads 11928326 The Exploration Targets of the Nanpu Sag: Insight from Organic Geochemical Characteristics of Source Rocks and Oils
Authors: Lixin Pei, Zhilong Huang, Wenzhe Gang
Abstract:
Organic geochemistry of source rocks and oils in the Nanpu Sag, Bohai Bay basin was studied on the basis of the results of Rock-Eval and biomarker. The possible source rocks consist of the third member (Es₃) and the first member (Es₁) of Shahejie formation and the third member of Dongying Formation (Ed₃) in the Nanpu Sag. The Es₃, Es₁, and Ed₃ source rock intervals in the Nanpu Sag all have high organic-matter richness and are at hydrocarbon generating stage, which are regarded as effective source rocks. The three possible source rock intervals have different biomarker associations and can be differentiated by gammacerane/αβ C₃₀ hopane, ETR ([C₂₈+C₂₉]/ [C₂₈+C₂₉+Ts]), C₂₇ diasterane/sterane and C₂₇/C₂₉ steranes, which suggests they deposited in different environments. Based on the oil-source rock correlation, the shallow oils mainly originated from the Es₃ and Es₁ source rocks in the Nanpu Sag. Through hydrocarbon generation and expulsion history of the source rocks, trap development history and accumulation history, the shallow oils mainly originated from paleo-reservoirs in the Es₃ and Es₁ during the period of Neotectonism, and the residual paleo-reservoirs in the Es₃ and Es₁ would be the focus targets in the Nanpu Sag; Bohai Bay Basin.Keywords: source rock, biomarker association, Nanpu Sag, Bohai Bay Basin
Procedia PDF Downloads 37328325 Intrabody Communication Using Different Ground Configurations in Digital Door Lock
Authors: Daewook Kim, Gilwon Yoon
Abstract:
Intrabody communication (IBC) is a new way of transferring data using human body as a medium. Minute current can travel though human body without any harm. IBC can remove electrical wires for human area network. IBC can be also a secure communication network system unlike wireless networks which can be accessed by anyone with bad intentions. One of the IBC systems is based on frequency shift keying modulation where individual data are transmitted to the external devices for the purpose of secure access such as digital door lock. It was found that the quality of IBC data transmission was heavily dependent on ground configurations of electronic circuits. Reliable IBC transmissions were not possible when both of the transmitter and receiver used batteries as circuit power source. Transmission was reliable when power supplies were used as power source for both transmitting and receiving sites because the common ground was established through the grounds of instruments such as power supply and oscilloscope. This was due to transmission dipole size and the ground effects of floor and AC power line. If one site used battery as power source and the other site used the AC power as circuit power source, transmission was possible.Keywords: frequency shift keying, ground, intrabody, communication, door lock
Procedia PDF Downloads 41828324 Knowledge-Driven Decision Support System Based on Knowledge Warehouse and Data Mining by Improving Apriori Algorithm with Fuzzy Logic
Authors: Pejman Hosseinioun, Hasan Shakeri, Ghasem Ghorbanirostam
Abstract:
In recent years, we have seen an increasing importance of research and study on knowledge source, decision support systems, data mining and procedure of knowledge discovery in data bases and it is considered that each of these aspects affects the others. In this article, we have merged information source and knowledge source to suggest a knowledge based system within limits of management based on storing and restoring of knowledge to manage information and improve decision making and resources. In this article, we have used method of data mining and Apriori algorithm in procedure of knowledge discovery one of the problems of Apriori algorithm is that, a user should specify the minimum threshold for supporting the regularity. Imagine that a user wants to apply Apriori algorithm for a database with millions of transactions. Definitely, the user does not have necessary knowledge of all existing transactions in that database, and therefore cannot specify a suitable threshold. Our purpose in this article is to improve Apriori algorithm. To achieve our goal, we tried using fuzzy logic to put data in different clusters before applying the Apriori algorithm for existing data in the database and we also try to suggest the most suitable threshold to the user automatically.Keywords: decision support system, data mining, knowledge discovery, data discovery, fuzzy logic
Procedia PDF Downloads 33528323 Using Open Source Data and GIS Techniques to Overcome Data Deficiency and Accuracy Issues in the Construction and Validation of Transportation Network: Case of Kinshasa City
Authors: Christian Kapuku, Seung-Young Kho
Abstract:
An accurate representation of the transportation system serving the region is one of the important aspects of transportation modeling. Such representation often requires developing an abstract model of the system elements, which also requires important amount of data, surveys and time. However, in some cases such as in developing countries, data deficiencies, time and budget constraints do not always allow such accurate representation, leaving opportunities to assumptions that may negatively affect the quality of the analysis. With the emergence of Internet open source data especially in the mapping technologies as well as the advances in Geography Information System, opportunities to tackle these issues have raised. Therefore, the objective of this paper is to demonstrate such application through a practical case of the development of the transportation network for the city of Kinshasa. The GIS geo-referencing was used to construct the digitized map of Transportation Analysis Zones using available scanned images. Centroids were then dynamically placed at the center of activities using an activities density map. Next, the road network with its characteristics was built using OpenStreet data and other official road inventory data by intersecting their layers and cleaning up unnecessary links such as residential streets. The accuracy of the final network was then checked, comparing it with satellite images from Google and Bing. For the validation, the final network was exported into Emme3 to check for potential network coding issues. Results show a high accuracy between the built network and satellite images, which can mostly be attributed to the use of open source data.Keywords: geographic information system (GIS), network construction, transportation database, open source data
Procedia PDF Downloads 16728322 An Introductory Study on Optimization Algorithm for Movable Sensor Network-Based Odor Source Localization
Authors: Yossiri Ariyakul, Piyakiat Insom, Poonyawat Sangiamkulthavorn, Takamichi Nakamoto
Abstract:
In this paper, the method of optimization algorithm for sensor network comprised of movable sensor nodes which can be used for odor source localization was proposed. A sensor node is composed of an odor sensor, an anemometer, and a wireless communication module. The odor intensity measured from the sensor nodes are sent to the processor to perform the localization based on optimization algorithm by which the odor source localization map is obtained as a result. The map can represent the exact position of the odor source or show the direction toward it remotely. The proposed method was experimentally validated by creating the odor source localization map using three, four, and five sensor nodes in which the accuracy to predict the position of the odor source can be observed.Keywords: odor sensor, odor source localization, optimization, sensor network
Procedia PDF Downloads 29928321 Forensic Challenges in Source Device Identification for Digital Videos
Authors: Mustapha Aminu Bagiwa, Ainuddin Wahid Abdul Wahab, Mohd Yamani Idna Idris, Suleman Khan
Abstract:
Video source device identification has become a problem of concern in numerous domains especially in multimedia security and digital investigation. This is because videos are now used as evidence in legal proceedings. Source device identification aim at identifying the source of digital devices using the content they produced. However, due to affordable processing tools and the influx in digital content generating devices, source device identification is still a major problem within the digital forensic community. In this paper, we discuss source device identification for digital videos by identifying techniques that were proposed in the literature for model or specific device identification. This is aimed at identifying salient open challenges for future research.Keywords: video forgery, source camcorder, device identification, forgery detection
Procedia PDF Downloads 63128320 Sensor Data Analysis for a Large Mining Major
Authors: Sudipto Shanker Dasgupta
Abstract:
One of the largest mining companies wanted to look at health analytics for their driverless trucks. These trucks were the key to their supply chain logistics. The automated trucks had multi-level sub-assemblies which would send out sensor information. The use case that was worked on was to capture the sensor signal from the truck subcomponents and analyze the health of the trucks from repair and replacement purview. Open source software was used to stream the data into a clustered Hadoop setup in Amazon Web Services cloud and Apache Spark SQL was used to analyze the data. All of this was achieved through a 10 node amazon 32 core, 64 GB RAM setup real-time analytics was achieved on ‘300 million records’. To check the scalability of the system, the cluster was increased to 100 node setup. This talk will highlight how Open Source software was used to achieve the above use case and the insights on the high data throughput on a cloud set up.Keywords: streaming analytics, data science, big data, Hadoop, high throughput, sensor data
Procedia PDF Downloads 40428319 Optimized Weight Selection of Control Data Based on Quotient Space of Multi-Geometric Features
Authors: Bo Wang
Abstract:
The geometric processing of multi-source remote sensing data using control data of different scale and different accuracy is an important research direction of multi-platform system for earth observation. In the existing block bundle adjustment methods, as the controlling information in the adjustment system, the approach using single observation scale and precision is unable to screen out the control information and to give reasonable and effective corresponding weights, which reduces the convergence and adjustment reliability of the results. Referring to the relevant theory and technology of quotient space, in this project, several subjects are researched. Multi-layer quotient space of multi-geometric features is constructed to describe and filter control data. Normalized granularity merging mechanism of multi-layer control information is studied and based on the normalized scale factor, the strategy to optimize the weight selection of control data which is less relevant to the adjustment system can be realized. At the same time, geometric positioning experiment is conducted using multi-source remote sensing data, aerial images, and multiclass control data to verify the theoretical research results. This research is expected to break through the cliché of the single scale and single accuracy control data in the adjustment process and expand the theory and technology of photogrammetry. Thus the problem to process multi-source remote sensing data will be solved both theoretically and practically.Keywords: multi-source image geometric process, high precision geometric positioning, quotient space of multi-geometric features, optimized weight selection
Procedia PDF Downloads 28428318 Nazca: A Context-Based Matching Method for Searching Heterogeneous Structures
Authors: Karine B. de Oliveira, Carina F. Dorneles
Abstract:
The structure level matching is the problem of combining elements of a structure, which can be represented as entities, classes, XML elements, web forms, and so on. This is a challenge due to large number of distinct representations of semantically similar structures. This paper describes a structure-based matching method applied to search for different representations in data sources, considering the similarity between elements of two structures and the data source context. Using real data sources, we have conducted an experimental study comparing our approach with our baseline implementation and with another important schema matching approach. We demonstrate that our proposal reaches higher precision than the baseline.Keywords: context, data source, index, matching, search, similarity, structure
Procedia PDF Downloads 36428317 Calculation of Detection Efficiency of Horizontal Large Volume Source Using Exvol Code
Authors: M. Y. Kang, Euntaek Yoon, H. D. Choi
Abstract:
To calculate the full energy (FE) absorption peak efficiency for arbitrary volume sample, we developed and verified the EXVol (Efficiency calculator for EXtended Voluminous source) code which is based on effective solid angle method. EXVol is possible to describe the source area as a non-uniform three-dimensional (x, y, z) source. And decompose and set it into several sets of volume units. Users can equally divide (x, y, z) coordinate system to calculate the detection efficiency at a specific position of a cylindrical volume source. By determining the detection efficiency for differential volume units, the total radiative absolute distribution and the correction factor of the detection efficiency can be obtained from the nondestructive measurement of the source. In order to check the performance of the EXVol code, Si ingot of 20 cm in diameter and 50 cm in height were used as a source. The detector was moved at the collimation geometry to calculate the detection efficiency at a specific position and compared with the experimental values. In this study, the performance of the EXVol code was extended to obtain the detection efficiency distribution at a specific position in a large volume source.Keywords: attenuation, EXVol, detection efficiency, volume source
Procedia PDF Downloads 185