Search results for: geological database
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2040

Search results for: geological database

1920 Deep Learning Strategies for Mapping Complex Vegetation Patterns in Mediterranean Environments Undergoing Climate Change

Authors: Matan Cohen, Maxim Shoshany

Abstract:

Climatic, topographic and geological diversity, together with frequent disturbance and recovery cycles, produce highly complex spatial patterns of trees, shrubs, dwarf shrubs and bare ground patches. Assessment of spatial and temporal variations of these life-forms patterns under climate change is of high ecological priority. Here we report on one of the first attempts to discriminate between images of three Mediterranean life-forms patterns at three densities. The development of an extensive database of orthophoto images representing these 9 pattern categories was instrumental for training and testing pre-trained and newly-trained DL models utilizing DenseNet architecture. Both models demonstrated the advantages of using Deep Learning approaches over existing spectral and spatial (pattern or texture) algorithmic methods in differentiation 9 life-form spatial mixtures categories.

Keywords: texture classification, deep learning, desert fringe ecosystems, climate change

Procedia PDF Downloads 74
1919 Change of Endocrine and Exocrine Insufficiency on Non-Diabetes Patients after Distal Pancreatectomy: A Nationwide Database Study

Authors: Jin-Ming Wu, Te-Wei Ho, Yu-Wen Tien

Abstract:

Background: The aim of this population-based study was to determine the occurrence of diabetes and exocrine pancreatic insufficiencies (EPI) on non-diabetes subjects receiving distal pancreatectomy (DP). Method: A nationwide cohort study between 2000 and 2010 was collected from the Taiwan National Health Insurance Research Database. Among 3264 DP patients, we identified 1410 non-diabetes and 966 non-diabetes non-EPI. Results. Of 1410 non-diabetes DP subjects, 312 patients (22.1%) developed newly-diagnosed diabetes after PD. On a multiple logistic regression model, co-morbid hyperlipidemia (odds ratio, 1.640; 95% CI, 1.362–2.763; P < 0.001) and pancreatitis (odds ratio, 2.428; 95% CI, 1.889–3.121; P < 0.001) significantly contributed to higher incidences of diabetes after DP. Moreover, 380 subjects (39.3%) developed EPI, and pancreatic cancer is the statistically significant risk factor (odds ratio, 4.663; 95% CI, 2.108–6.085; P < 0.001). Conclusion: The patients with co-morbid hyperlipidemia and chronic pancreatitis had higher rates of newly-diagnosed diabetes after DP, moreover, pancreatic cancer subjects had higher rates of pancreatic exocrine insufficiency after DP. The clinicians should be alert to follow up glucose metabolism and clinical symptoms of fat intolerance for DP patients.

Keywords: distal pancreatectomy, National database, diabetes, exocrine insufficiency

Procedia PDF Downloads 179
1918 Privacy Preserving in Association Rule Mining on Horizontally Partitioned Database

Authors: Manvar Sagar, Nikul Virpariya

Abstract:

The advancement in data mining techniques plays an important role in many applications. In context of privacy and security issues, the problems caused by association rule mining technique are investigated by many research scholars. It is proved that the misuse of this technique may reveal the database owner’s sensitive and private information to others. Many researchers have put their effort to preserve privacy in Association Rule Mining. Amongst the two basic approaches for privacy preserving data mining, viz. Randomization based and Cryptography based, the later provides high level of privacy but incurs higher computational as well as communication overhead. Hence, it is necessary to explore alternative techniques that improve the over-heads. In this work, we propose an efficient, collusion-resistant cryptography based approach for distributed Association Rule mining using Shamir’s secret sharing scheme. As we show from theoretical and practical analysis, our approach is provably secure and require only one time a trusted third party. We use secret sharing for privately sharing the information and code based identification scheme to add support against malicious adversaries.

Keywords: Privacy, Privacy Preservation in Data Mining (PPDM), horizontally partitioned database, EMHS, MFI, shamir secret sharing

Procedia PDF Downloads 383
1917 SQL Generator Based on MVC Pattern

Authors: Chanchai Supaartagorn

Abstract:

Structured Query Language (SQL) is the standard de facto language to access and manipulate data in a relational database. Although SQL is a language that is simple and powerful, most novice users will have trouble with SQL syntax. Thus, we are presenting SQL generator tool which is capable of translating actions and displaying SQL commands and data sets simultaneously. The tool was developed based on Model-View-Controller (MVC) pattern. The MVC pattern is a widely used software design pattern that enforces the separation between the input, processing, and output of an application. Developers take full advantage of it to reduce the complexity in architectural design and to increase flexibility and reuse of code. In addition, we use White-Box testing for the code verification in the Model module.

Keywords: MVC, relational database, SQL, White-Box testing

Procedia PDF Downloads 405
1916 Delineating Subsurface Linear Features and Faults Under Sedimentary Cover in the Bahira Basin Using Integrated Gravity and Magnetic Data

Authors: M. Lghoul, N. El Goumi, M. Guernouche

Abstract:

In order to predict the structural and tectonic framework of the Bahira basin and to have a 3D geological modeling of the basin, an integrated multidisciplinary work has been conducted using gravity, magnetic and geological data. The objective of the current study is delineating the subsurfacefeatures, faults, and geological limits, using airborne magnetic and gravity data analysis of the Bahira basin. To achieve our goal, we have applied different enhanced techniques on magnetic and gravity data: power spectral analysis techniques, reduction to pole (RTP), upward continuation, analytical signal, tilt derivative, total horizontal derivative, 3D Euler deconvolutionand source parameter imagining. The major lineaments/faults trend are: NE–SW, NW-SE, ENE–WSW, and WNW–ESE. The 3D Euler deconvolution analysis highlighted a number of fault trend, mainly in the ENE-WSW, WNW-ESE directions. The depth tothe top of the basement sources in the study area ranges between 200 m, in the southern and northern part of the Bahira basin, to 5000 m located in the Eastern part of the basin.

Keywords: magnetic, gravity, structural trend, depth to basement

Procedia PDF Downloads 112
1915 Integrating a Universal Forensic DNA Database: Anticipated Deterrent Effects

Authors: Karen Fang

Abstract:

Investigative genetic genealogy has attracted much interest in both the field of ethics and the public eye due to its global application in criminal cases. Arguments have been made regarding privacy and informed consent, especially with law enforcement using consumer genetic testing results to convict individuals. In the case of public interest, DNA databases have the strong potential to significantly reduce crime, which in turn leads to safer communities and better futures. With the advancement of genetic technologies, the integration of a universal forensic DNA database in violent crimes, crimes against children, and missing person cases is expected to deter crime while protecting one’s privacy. Rather than collecting whole genomes from the whole population, STR profiles can be used to identify unrelated individuals without compromising personal information such as physical appearance, disease risk, and geographical origin, and additionally, reduce cost and storage space. STR DNA profiling is already used in the forensic science field and going a step further benefits several areas, including the reduction in recidivism, improved criminal court case turnaround time, and just punishment. Furthermore, adding individuals to the database as early as possible prevents young offenders and first-time offenders from participating in criminal activity. It is important to highlight that DNA databases should be inclusive and tightly governed, and the misconception on the use of DNA based on crime television series and other media sources should be addressed. Nonetheless, deterrent effects have been observed in countries like the US and Denmark with DNA databases that consist of serious violent offenders. Fewer crimes were reported, and fewer people were convicted of those crimes- a favorable outcome, not even the death penalty could provide. Currently, there is no better alternative than a universal forensic DNA database made up of STR profiles. It can open doors for investigative genetic genealogy and fostering better communities. Expanding the appropriate use of DNA databases is ethically acceptable and positively impacts the public.

Keywords: bioethics, deterrent effects, DNA database, investigative genetic genealogy, privacy, public interest

Procedia PDF Downloads 136
1914 System of Quality Automation for Documents (SQAD)

Authors: R. Babi Saraswathi, K. Divya, A. Habeebur Rahman, D. B. Hari Prakash, S. Jayanth, T. Kumar, N. Vijayarangan

Abstract:

Document automation is the design of systems and workflows, assembling repetitive documents to meet the specific business needs. In any organization or institution, documenting employee’s information is very important for both employees as well as management. It shows an individual’s progress to the management. Many documents of the employee are in the form of papers, so it is very difficult to arrange and for future reference we need to spend more time in getting the exact document. Also, it is very tedious to generate reports according to our needs. The process gets even more difficult on getting approvals and hence lacks its security aspects. This project overcomes the above-stated issues. By storing the details in the database and maintaining the e-documents, the automation system reduces the manual work to a large extent. Then the approval process of some important documents can be done in a much-secured manner by using Digital Signature and encryption techniques. Details are maintained in the database and e-documents are stored in specific folders and generation of various kinds of reports is possible. Moreover, an efficient search method is implemented is used in the database. Automation supporting document maintenance in many aspects is useful for minimize data entry, reduce the time spent on proof-reading, avoids duplication, and reduce the risks associated with the manual error, etc.

Keywords: e-documents, automation, digital signature, encryption

Procedia PDF Downloads 372
1913 Modeling Core Flooding Experiments for Co₂ Geological Storage Applications

Authors: Avinoam Rabinovich

Abstract:

CO₂ geological storage is a proven technology for reducing anthropogenic carbon emissions, which is paramount for achieving the ambitious net zero emissions goal. Core flooding experiments are an important step in any CO₂ storage project, allowing us to gain information on the flow of CO₂ and brine in the porous rock extracted from the reservoir. This information is important for understanding basic mechanisms related to CO₂ geological storage as well as for reservoir modeling, which is an integral part of a field project. In this work, a different method for constructing accurate models of CO₂-brine core flooding will be presented. Results for synthetic cases and real experiments will be shown and compared with numerical models to exhibit their predictive capabilities. Furthermore, the various mechanisms which impact the CO₂ distribution and trapping in the rock samples will be discussed, and examples from models and experiments will be provided. The new method entails solving an inverse problem to obtain a three-dimensional permeability distribution which, along with the relative permeability and capillary pressure functions, constitutes a model of the flow experiments. The model is more accurate when data from a number of experiments are combined to solve the inverse problem. This model can then be used to test various other injection flow rates and fluid fractions which have not been tested in experiments. The models can also be used to bridge the gap between small-scale capillary heterogeneity effects (sub-core and core scale) and large-scale (reservoir scale) effects, known as the upscaling problem.

Keywords: CO₂ geological storage, residual trapping, capillary heterogeneity, core flooding, CO₂-brine flow

Procedia PDF Downloads 50
1912 Critical Review of Oceanic and Geological Storage of Carbon Sequestration

Authors: Milad Nooshadi, Alessandro Manzardo

Abstract:

CO₂ emissions in the atmosphere continue to rise, mostly as a result of the combustion of fossil fuels. CO₂ injection into the oceans and geological formation as a process of physical carbon capture are two of the most promising emerging strategies for mitigating climate change and global warming. The purpose of this research is to evaluate the two mentioned methods of CO₂ sequestration and to assess information on previous and current advancements, limitations, and uncertainties associated with carbon sequestration in order to identify possible prospects for ensuring the timely implementation of the technology, such as determining how governments and companies can gain a better understanding of CO₂ storage in terms of which media have the most applicable capacity, which type of injection has the fewer environmental impact, and how much carbon sequestration and storage will cost. The behavior of several forms is characterized as a near field, a far field, and a see-floor in ocean storage, and three medias in geological formations as an oil and gas reservoir, a saline aquifer, and a coal bed. To determine the capacity of various forms of media, an analysis of some models and practical experiments are necessary. Additionally, as a major component of sequestration, the various injection methods into diverse media and their monitoring are associated with a variety of environmental impacts and financial consequences.

Keywords: carbon sequestration, ocean storage, geologic storage, carbon transportation

Procedia PDF Downloads 82
1911 Biimodal Biometrics System Using Fusion of Iris and Fingerprint

Authors: Attallah Bilal, Hendel Fatiha

Abstract:

This paper proposes the bimodal biometrics system for identity verification iris and fingerprint, at matching score level architecture using weighted sum of score technique. The features are extracted from the pre processed images of iris and fingerprint. These features of a query image are compared with those of a database image to obtain matching scores. The individual scores generated after matching are passed to the fusion module. This module consists of three major steps i.e., normalization, generation of similarity score and fusion of weighted scores. The final score is then used to declare the person as genuine or an impostor. The system is tested on CASIA database and gives an overall accuracy of 91.04% with FAR of 2.58% and FRR of 8.34%.

Keywords: iris, fingerprint, sum rule, fusion

Procedia PDF Downloads 344
1910 Regeneration of Geological Models Using Support Vector Machine Assisted by Principal Component Analysis

Authors: H. Jung, N. Kim, B. Kang, J. Choe

Abstract:

History matching is a crucial procedure for predicting reservoir performances and making future decisions. However, it is difficult due to uncertainties of initial reservoir models. Therefore, it is important to have reliable initial models for successful history matching of highly heterogeneous reservoirs such as channel reservoirs. In this paper, we proposed a novel scheme for regenerating geological models using support vector machine (SVM) and principal component analysis (PCA). First, we perform PCA for figuring out main geological characteristics of models. Through the procedure, permeability values of each model are transformed to new parameters by principal components, which have eigenvalues of large magnitude. Secondly, the parameters are projected into two-dimensional plane by multi-dimensional scaling (MDS) based on Euclidean distances. Finally, we train an SVM classifier using 20% models which show the most similar or dissimilar well oil production rates (WOPR) with the true values (10% for each). Then, the other 80% models are classified by trained SVM. We select models on side of low WOPR errors. One hundred channel reservoir models are initially generated by single normal equation simulation. By repeating the classification process, we can select models which have similar geological trend with the true reservoir model. The average field of the selected models is utilized as a probability map for regeneration. Newly generated models can preserve correct channel features and exclude wrong geological properties maintaining suitable uncertainty ranges. History matching with the initial models cannot provide trustworthy results. It fails to find out correct geological features of the true model. However, history matching with the regenerated ensemble offers reliable characterization results by figuring out proper channel trend. Furthermore, it gives dependable prediction of future performances with reduced uncertainties. We propose a novel classification scheme which integrates PCA, MDS, and SVM for regenerating reservoir models. The scheme can easily sort out reliable models which have similar channel trend with the reference in lowered dimension space.

Keywords: history matching, principal component analysis, reservoir modelling, support vector machine

Procedia PDF Downloads 139
1909 Characteristics Features and Action Mechanism of Some Country Made Pistols

Authors: Ajitesh Pal, Arpan Datta Roy, H. K. Pratihari

Abstract:

The different illegal firearms crudely made by skilled gunsmith from scrap materials are popularly known as country made firearms. Such firearms along with improvised ammunition are clandestinely marketed at the cheaper price without any license to the extremist group, criminal, poachers and firearm lovers. As per National Crime Records Bureau (NCRB), MHA, Govt of India about 80% firearm cases are committed by country made/improvised firearms. The ballistic division of the laboratory has examined a good number of cases. The analysis of firearm cases received for forensic examination revealed that 7.65mm calibre pistols mostly improvised firearm are commonly used in firearm related crime cases. In the present communication, physical parameters and other characteristics features of some 7.65mm calibre pistols have been discussed in detail. The detailed study on country made (CM) firearm will help to prepare a database related to type of material used, origin of the raw material and tools used for inscription. The study also includes to establish the chemistry of propellants & head stamp pattern. The database will be helpful to the firearm examiners, researchers, students pursuing study on forensic science as reference material.

Keywords: improvised pistol, stringent gun law, working mechanism, parameters, database

Procedia PDF Downloads 55
1908 The Modification of Convolutional Neural Network in Fin Whale Identification

Authors: Jiahao Cui

Abstract:

In the past centuries, due to climate change and intense whaling, the global whale population has dramatically declined. Among the various whale species, the fin whale experienced the most drastic drop in number due to its popularity in whaling. Under this background, identifying fin whale calls could be immensely beneficial to the preservation of the species. This paper uses feature extraction to process the input audio signal, then a network based on AlexNet and three networks based on the ResNet model was constructed to classify fin whale calls. A mixture of the DOSITS database and the Watkins database was used during training. The results demonstrate that a modified ResNet network has the best performance considering precision and network complexity.

Keywords: convolutional neural network, ResNet, AlexNet, fin whale preservation, feature extraction

Procedia PDF Downloads 96
1907 A Novel Framework for User-Friendly Ontology-Mediated Access to Relational Databases

Authors: Efthymios Chondrogiannis, Vassiliki Andronikou, Efstathios Karanastasis, Theodora Varvarigou

Abstract:

A large amount of data is typically stored in relational databases (DB). The latter can efficiently handle user queries which intend to elicit the appropriate information from data sources. However, direct access and use of this data requires the end users to have an adequate technical background, while they should also cope with the internal data structure and values presented. Consequently the information retrieval is a quite difficult process even for IT or DB experts, taking into account the limited contributions of relational databases from the conceptual point of view. Ontologies enable users to formally describe a domain of knowledge in terms of concepts and relations among them and hence they can be used for unambiguously specifying the information captured by the relational database. However, accessing information residing in a database using ontologies is feasible, provided that the users are keen on using semantic web technologies. For enabling users form different disciplines to retrieve the appropriate data, the design of a Graphical User Interface is necessary. In this work, we will present an interactive, ontology-based, semantically enable web tool that can be used for information retrieval purposes. The tool is totally based on the ontological representation of underlying database schema while it provides a user friendly environment through which the users can graphically form and execute their queries.

Keywords: ontologies, relational databases, SPARQL, web interface

Procedia PDF Downloads 255
1906 Rendering Cognition Based Learning in Coherence with Development within the Context of PostgreSQL

Authors: Manuela Nayantara Jeyaraj, Senuri Sucharitharathna, Chathurika Senarath, Yasanthy Kanagaraj, Indraka Udayakumara

Abstract:

PostgreSQL is an Object Relational Database Management System (ORDBMS) that has been in existence for a while. Despite the superior features that it wraps and packages to manage database and data, the database community has not fully realized the importance and advantages of PostgreSQL. Hence, this research tends to focus on provisioning a better environment of development for PostgreSQL in order to induce the utilization and elucidate the importance of PostgreSQL. PostgreSQL is also known to be the world’s most elementary SQL-compliant open source ORDBMS. But, users have not yet resolved to PostgreSQL due to the facts that it is still under the layers and the complexity of its persistent textual environment for an introductory user. Simply stating this, there is a dire need to explicate an easy way of making the users comprehend the procedure and standards with which databases are created, tables and the relationships among them, manipulating queries and their flow based on conditions in PostgreSQL to help the community resolve to PostgreSQL at an augmented rate. Hence, this research under development within the context tends to initially identify the dominant features provided by PostgreSQL over its competitors. Following the identified merits, an analysis on why the database community holds a hesitance in migrating to PostgreSQL’s environment will be carried out. These will be modulated and tailored based on the scope and the constraints discovered. The resultant of the research proposes a system that will serve as a designing platform as well as a learning tool that will provide an interactive method of learning via a visual editor mode and incorporate a textual editor for well-versed users. The study is based on conjuring viable solutions that analyze a user’s cognitive perception in comprehending human computer interfaces and the behavioural processing of design elements. By providing a visually draggable and manipulative environment to work with Postgresql databases and table queries, it is expected to highlight the elementary features displayed by Postgresql over any other existent systems in order to grasp and disseminate the importance and simplicity offered by this to a hesitant user.

Keywords: cognition, database, PostgreSQL, text-editor, visual-editor

Procedia PDF Downloads 255
1905 Analysis and Prediction of COVID-19 by Using Recurrent LSTM Neural Network Model in Machine Learning

Authors: Grienggrai Rajchakit

Abstract:

As we all know that coronavirus is announced as a pandemic in the world by WHO. It is speeded all over the world with few days of time. To control this spreading, every citizen maintains social distance and self-preventive measures are the best strategies. As of now, many researchers and scientists are continuing their research in finding out the exact vaccine. The machine learning model finds that the coronavirus disease behaves in an exponential manner. To abolish the consequence of this pandemic, an efficient step should be taken to analyze this disease. In this paper, a recurrent neural network model is chosen to predict the number of active cases in a particular state. To make this prediction of active cases, we need a database. The database of COVID-19 is downloaded from the KAGGLE website and is analyzed by applying a recurrent LSTM neural network with univariant features to predict the number of active cases of patients suffering from the corona virus. The downloaded database is divided into training and testing the chosen neural network model. The model is trained with the training data set and tested with a testing dataset to predict the number of active cases in a particular state; here, we have concentrated on Andhra Pradesh state.

Keywords: COVID-19, coronavirus, KAGGLE, LSTM neural network, machine learning

Procedia PDF Downloads 138
1904 Block Mining: Block Chain Enabled Process Mining Database

Authors: James Newman

Abstract:

Process mining is an emerging technology that looks to serialize enterprise data in time series data. It has been used by many companies and has been the subject of a variety of research papers. However, the majority of current efforts have looked at how to best create process mining from standard relational databases. This paper is the first pass at outlining a database custom-built for the minimal viable product of process mining. We present Block Miner, a blockchain protocol to store process mining data across a distributed network. We demonstrate the feasibility of storing process mining data on the blockchain. We present a proof of concept and show how the intersection of these two technologies helps to solve a variety of issues, including but not limited to ransomware attacks, tax documentation, and conflict resolution.

Keywords: blockchain, process mining, memory optimization, protocol

Procedia PDF Downloads 71
1903 Modified Active (MA) Algorithm to Generate Semantic Web Related Clustered Hierarchy for Keyword Search

Authors: G. Leena Giri, Archana Mathur, S. H. Manjula, K. R. Venugopal, L. M. Patnaik

Abstract:

Keyword search in XML documents is based on the notion of lowest common ancestors in the labelled trees model of XML documents and has recently gained a lot of research interest in the database community. In this paper, we propose the Modified Active (MA) algorithm which is an improvement over the active clustering algorithm by taking into consideration the entity aspect of the nodes to find the level of the node pertaining to a particular keyword input by the user. A portion of the bibliography database is used to experimentally evaluate the modified active algorithm and results show that it performs better than the active algorithm. Our modification improves the response time of the system and thereby increases the efficiency of the system.

Keywords: keyword matching patterns, MA algorithm, semantic search, knowledge management

Procedia PDF Downloads 384
1902 Investigation of Lead and Zinc Oxide Deposits Using Geological and Geophysical Techniques at Oshiri Province in Onicha Local Government Area of Ebonyi State Located Within Southeastern Part of Nigeria, West Africa

Authors: Amaechi O. Azi, Uche D. Aluge, Lim H. San, Godwin A. Agbo

Abstract:

This paper is centered on the investigation of mineral deposits in selected locations in Oshiri province in Ebonyi State. Mineral deposits contribute immensely to the economic growth of a society. In researching lead and zinc oxide-bearing sites at Oshiri, geological and geophysical research technique was employed. Petrozenith, Earth Resistivity Meter, and Schlumberger setup were selected to examine the electrical characteristics of the subsurface. To determine the apparent resistivity of the subsurface, five soundings were taken, and the field data were processed using WinResist software. The mudstone, lead-shale, shale-granite, and lateritic topsoil were the four geoelectric strata that were found. The third layer, which corresponds to the shale-lead lithology, has a resistivity value between 211.9 m to 807.7 m at a depth of 25 m. Due to its resistivity levels and geological trend, this layer makes an excellent signature for lead-zinc occurrence. This zone is expected to house deposits of lead and zinc oxide in commercial quantity.

Keywords: Schlumberger, current, resistivity, lithology

Procedia PDF Downloads 56
1901 Geo Spatial Database for Railway Assets Management

Authors: Muhammad Umar

Abstract:

Safety and Assets management is considering a backbone of every department. GIS in the Railway become very important to Manage Assets and Security through Digital Maps and Web based GIS Maps. It provides a complete frame of work to the organization for the management of assets. Pakistan Railway is the most common and safest mode of traveling in Pakistan. Due to ever-increasing demand of transporting huge amount of information generated from various sources and this information must be accurate. This creates problems for Passengers and Administration that causes finical and time loss. GIS Solve this problem by Digital Maps & Database. It provides you a real time Spatial and Statistical analysis that helps you to communicate and exchange the information in a sophisticated way to the users. GIS Based Web system provides a facility to different end user to make query at a time as per requirements. This GIS System provides an advancement in an organization for a complete Monitoring, Safety and Decision System for tracks, Stations and Junctions that further use for the Analysis of different areas i.e. analysis of tracks, junctions and Stations in case of reconstruction, Rescue for rail accidents and Natural disasters .This Research work helps to reduce the financial loss and reduce human mistakes helps you provide a complete security and Management system of assets.

Keywords: Geographical Information System (GIS) for assets management, geo spatial database, railway assets management, Pakistan

Procedia PDF Downloads 470
1900 TomoTherapy® System Repositioning Accuracy According to Treatment Localization

Authors: Veronica Sorgato, Jeremy Belhassen, Philippe Chartier, Roddy Sihanath, Nicolas Docquiere, Jean-Yves Giraud

Abstract:

We analyzed the image-guided radiotherapy method used by the TomoTherapy® System (Accuray Corp.) for patient repositioning in clinical routine. The TomoTherapy® System computes X, Y, Z and roll displacements to match the reference CT, on which the dosimetry has been performed, with the pre-treatment MV CT. The accuracy of the repositioning method has been studied according to the treatment localization. For this, a database of 18774 treatment sessions, performed during 2 consecutive years (2016-2017 period) has been used. The database includes the X, Y, Z and roll displacements proposed by TomoTherapy® System as well as the manual correction of these proposals applied by the radiation therapist. This manual correction aims to further improve the repositioning based on the clinical situation and depends on the structures surrounding the target tumor tissue. The statistical analysis performed on the database aims to define repositioning limits to be used as security and guiding tool for the manual adjustment implemented by the radiation therapist. This tool will participate not only to notify potential repositioning errors but also to further improve patient positioning for optimal treatment.

Keywords: accuracy, IGRT MVCT, image-guided radiotherapy megavoltage computed tomography, statistical analysis, tomotherapy, localization

Procedia PDF Downloads 207
1899 Bundle Block Detection Using Spectral Coherence and Levenberg Marquardt Neural Network

Authors: K. Padmavathi, K. Sri Ramakrishna

Abstract:

This study describes a procedure for the detection of Left and Right Bundle Branch Block (LBBB and RBBB) ECG patterns using spectral Coherence(SC) technique and LM Neural Network. The Coherence function finds common frequencies between two signals and evaluate the similarity of the two signals. The QT variations of Bundle Blocks are observed in lead V1 of ECG. Spectral Coherence technique uses Welch method for calculating PSD. For the detection of normal and Bundle block beats, SC output values are given as the input features for the LMNN classifier. Overall accuracy of LMNN classifier is 99.5 percent. The data was collected from MIT-BIH Arrhythmia database.

Keywords: bundle block, SC, LMNN classifier, welch method, PSD, MIT-BIH, arrhythmia database

Procedia PDF Downloads 256
1898 Practices in Planning, Design and Construction of Head Race Tunnel of a Hydroelectric Project

Authors: M. S. Thakur, Mohit Shukla

Abstract:

A channel/tunnel, which carries the water to the penstock/pressure shaft is called headrace tunnel (HRT). It is necessary to know the general topography, geology of the area, state of stress and other mechanical properties of the strata. For this certain topographical and geological investigations, in-situ and laboratory tests, and observations are required to be done. These investigations play an important role in a tunnel design as these help in deciding the optimum layout, shape and size and support requirements of the tunnel. The paper includes inputs from Nathpa Jhakri Hydeoelectric project which is India’s highest capacity (1500 MW) operating hydroelectric project. The paper would help the design engineers with various new concepts and preparedness against geological surprises.

Keywords: tunnelling, geology, HRT, rockmass

Procedia PDF Downloads 234
1897 Intelligent Earthquake Prediction System Based On Neural Network

Authors: Emad Amar, Tawfik Khattab, Fatma Zada

Abstract:

Predicting earthquakes is an important issue in the study of geography. Accurate prediction of earthquakes can help people to take effective measures to minimize the loss of personal and economic damage, such as large casualties, destruction of buildings and broken of traffic, occurred within a few seconds. United States Geological Survey (USGS) science organization provides reliable scientific information of Earthquake Existed throughout history & Preliminary database from the National Center Earthquake Information (NEIC) show some useful factors to predict an earthquake in a seismic area like Aleutian Arc in the U.S. state of Alaska. The main advantage of this prediction method that it does not require any assumption, it makes prediction according to the future evolution of object's time series. The article compares between simulation data result from trained BP and RBF neural network versus actual output result from the system calculations. Therefore, this article focuses on analysis of data relating to real earthquakes. Evaluation results show better accuracy and higher speed by using radial basis functions (RBF) neural network.

Keywords: BP neural network, prediction, RBF neural network, earthquake

Procedia PDF Downloads 473
1896 Knowledge-Driven Decision Support System Based on Knowledge Warehouse and Data Mining by Improving Apriori Algorithm with Fuzzy Logic

Authors: Pejman Hosseinioun, Hasan Shakeri, Ghasem Ghorbanirostam

Abstract:

In recent years, we have seen an increasing importance of research and study on knowledge source, decision support systems, data mining and procedure of knowledge discovery in data bases and it is considered that each of these aspects affects the others. In this article, we have merged information source and knowledge source to suggest a knowledge based system within limits of management based on storing and restoring of knowledge to manage information and improve decision making and resources. In this article, we have used method of data mining and Apriori algorithm in procedure of knowledge discovery one of the problems of Apriori algorithm is that, a user should specify the minimum threshold for supporting the regularity. Imagine that a user wants to apply Apriori algorithm for a database with millions of transactions. Definitely, the user does not have necessary knowledge of all existing transactions in that database, and therefore cannot specify a suitable threshold. Our purpose in this article is to improve Apriori algorithm. To achieve our goal, we tried using fuzzy logic to put data in different clusters before applying the Apriori algorithm for existing data in the database and we also try to suggest the most suitable threshold to the user automatically.

Keywords: decision support system, data mining, knowledge discovery, data discovery, fuzzy logic

Procedia PDF Downloads 311
1895 Web Map Service for Fragmentary Rockfall Inventory

Authors: M. Amparo Nunez-Andres, Nieves Lantada

Abstract:

One of the most harmful geological risks is rockfalls. They cause both economic lost, damaged in buildings and infrastructures, and personal ones. Therefore, in order to estimate the risk of the exposed elements, it is necessary to know the mechanism of this kind of events, since the characteristics of the rock walls, to the propagation of fragments generated by the initial detached rock mass. In the framework of the research RockModels project, several inventories of rockfalls were carried out along the northeast of the Spanish peninsula and the Mallorca island. These inventories have general information about the events, although the important fact is that they contained detailed information about fragmentation. Specifically, the IBSD (Insitu Block Size Distribution) is obtained by photogrammetry from drone or TLS (Terrestrial Laser Scanner) and the RBSD (Rock Block Size Distribution) from the volume of the fragment in the deposit measured by hand. In order to share all this information with other scientists, engineers, members of civil protection, and stakeholders, it is necessary a platform accessible from the internet and following interoperable standards. In all the process, open-software have been used: PostGIS 2.1., Geoserver, and OpenLayers library. In the first step, a spatial database was implemented to manage all the information. We have used the data specifications of INSPIRE for natural risks adding specific and detailed data about fragmentation distribution. The next step was to develop a WMS with Geoserver. A previous phase was the creation of several views in PostGIS to show the information at different scales of visualization and with different degrees of detail. In the first view, the sites are identified with a point, and basic information about the rockfall event is facilitated. In the next level of zoom, at medium scale, the convex hull of the rockfall appears with its real shape and the source of the event and fragments are represented by symbols. The queries at this level offer a major detail about the movement. Eventually, the third level shows all elements: deposit, source, and blocks, in their real size, if it is possible, and in their real localization. The last task was the publication of all information in a web mapping site (www.rockdb.upc.edu) with data classified by levels using libraries in JavaScript as OpenLayers.

Keywords: geological risk, web mapping, WMS, rockfalls

Procedia PDF Downloads 140
1894 Research on Sensitivity of Geological Disasters in Road Area Based on Analytic Hierarchy Process

Authors: Li Yongyi

Abstract:

In order to explore the distribution of geological disasters within the expressway area of Shaanxi Province, the Analytic Hierarchy Process theory is applied based on the geographic information system technology platform, and the ground elevation, rainfall, vegetation coverage and other indicators are selected for analysis, and the expressway area is sensitive Sexual evaluation. The results show that the highway area disasters in Shaanxi Province are mainly distributed in the southern mountainous areas and are dominated by landslides; the disaster area ratio basically increases with the increase in ground elevation, surface slope, surface undulation, rainfall, and vegetation coverage. The increase in the distance from the river shows a decreasing trend; after grading the disaster sensitivity within 5km of the expressway, the extremely sensitive area, the highly sensitive area, the medium sensitive area, the low sensitive area, and the extremely low sensitive area respectively account for 8.17%、15.80%、22.99%、26.22%、26.82%. Highly sensitive road areas are mainly distributed in southern Shaanxi.

Keywords: highway engineering, sensitivity, analytic hierarchy process, geological hazard, road area

Procedia PDF Downloads 79
1893 Geophysical Approach in the Geological Characterization of a Dam Site: Case of the Chebabta-Dam, Meskiana, Oum El-Bouaghi

Authors: Benhammadi Hocine, Djamel Boubaya, Chaffai Hicham

Abstract:

Meskiana Area is characterized by a semi-arid climate where the water supply for irrigation and industry is not sufficient as the priority goes for domestic use. To meet the increasing population growth and development, the authorities have considered building a new water retaining structure on some major temporary water streams. For this purpose Chebabta site on Oued Meskiana was chosen as the future dam site. It is large enough to store the desired volume of water. This study comes to investigate the conditions of the site and the adequacy of the ground as a foundation for the projected dam. The conditions of the site include the geological structure and mainly the presence of discontinuities in the formation on which the dam will be built, the nature of the lithologies under the foundation and the future lake, and the presence of any hazard. This site characterization is usually carried out using different methods in order to highlight any underground buried problematic structure. In this context, the different geophysical technics remain the most used ones. Three geophysical methods were used in the case of the Chebabta dam site, namely, electric survey, seismic refraction, and tomography. The choice of the technics and the location of the scan line was made on the basis of the available geological data. In this sense, profiles have been established on both banks of Oued Meskiana. The obtained results have allowed a better characterization of the geological structure, defining the limit between the surface cover and the bedrock, which is, in other words, the limit between the weathered zone and the bedrock. Their respective thicknesses were also determined by seismic refraction and electrical resistivity sounding. However, the tomography imaging technic has succeeded in positioning a fault structure passing through the right bank of the wadi.

Keywords: dam site, fault, geophysic, investigation, Meskiana

Procedia PDF Downloads 65
1892 Descriptive Analysis of the Database of Poliomyelitis Surveillance System in Mauritania from 2012-2019

Authors: B. Baba Ahmed, P. Yanogo, B. Djibryl. N. Medas

Abstract:

Introduction: Polio is a highly contagious viral infection, with children under 5 years of age being the most affected. It is a public health emergency of international concern. Polio surveillance in Mauritania has been ongoing since 1998 and has achieved "polio free" status in 2007. our objective is to analyse a pidemiological surveillance database of poliomyélitis in Mauritania from 2012-2019. Method: A transversal descriptive analysis of poliomyélitis database was carried out in Mauritania from 2012-2019.An exhaustive sampling was done on all suspected polio cases recorded in the database from 2012 -2019. This study used EPI-INFO 7.4 for frequency calculation for qualitative variables, mean and standard deviation for quantitative variables. Results: We found 459 suspected cases of polio over the study period with an average rate of acute non-polio flaccid paralysis of 25.4 cases/100,000 children under 15 years of age. The age group 0-6 years represented 75.2%. Males constituted 50.2%. Females represented 49.78% with a ratio of M/F=1.Among the 422 observations, the average age is 4 years +/- 3.38. The four regions, TIRIS-ZEMMOUR, INCHIRI, TAGANT, NOUACHCHOTT OUEST recorded the lowest percentages of notifications, respectively (3.28%; 3.93%; 4.37%; 4.8%). 99.34% [98.09-99.78] of cases presented acute flaccid paralysis. And 56.77% [52.19-61.23], had limb asymmetry. We showed that 82.93% [79.21-86.10], had fever. we found that 89.5% of suspected polio cases were investigated before 48 hours. And 88.39% of suspected cases had two adequate samples taken 48 hours apart and within 14 days after the onset of symptoms. Only 30.95% of samples arrived at the referral laboratory before 72 hours. Conclusion: This study has shown that Mauritania has achieved the objectives in most of the quantitative performance indicators of polio surveillance. This study has shown a low notification of cases in the northern and central regions of the country. There is a problem with the transport of samples to the laboratory.

Keywords: analysis, data base, Epi-Info, polio

Procedia PDF Downloads 155
1891 On the Design of a Secure Two-Party Authentication Scheme for Internet of Things Using Cancelable Biometrics and Physically Unclonable Functions

Authors: Behnam Zahednejad, Saeed Kosari

Abstract:

Widespread deployment of Internet of Things (IoT) has raised security and privacy issues in this environment. Designing a secure two-factor authentication scheme between the user and server is still a challenging task. In this paper, we focus on Cancelable Biometric (CB) as an authentication factor in IoT. We show that previous CB-based scheme fail to provide real two-factor security, Perfect Forward Secrecy (PFS) and suffer database attacks and traceability of the user. Then we propose our improved scheme based on CB and Physically Unclonable Functions (PUF), which can provide real two-factor security, PFS, user’s unlinkability, and resistance to database attack. In addition, Key Compromise Impersonation (KCI) resilience is achieved in our scheme. We also prove the security of our proposed scheme formally using both Real-Or-Random (RoR) model and the ProVerif analysis tool. For the usability of our scheme, we conducted a performance analysis and showed that our scheme has the least communication cost compared to the previous CB-based scheme. The computational cost of our scheme is also acceptable for the IoT environment.

Keywords: IoT, two-factor security, cancelable biometric, key compromise impersonation resilience, perfect forward secrecy, database attack, real-or-random model, ProVerif

Procedia PDF Downloads 78