Search results for: end-user trained information extraction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13196

Search results for: end-user trained information extraction

12836 Text Data Preprocessing Library: Bilingual Approach

Authors: Kabil Boukhari

Abstract:

In the context of information retrieval, the selection of the most relevant words is a very important step. In fact, the text cleaning allows keeping only the most representative words for a better use. In this paper, we propose a library for the purpose text preprocessing within an implemented application to facilitate this task. This study has two purposes. The first, is to present the related work of the various steps involved in text preprocessing, presenting the segmentation, stemming and lemmatization algorithms that could be efficient in the rest of study. The second, is to implement a developed tool for text preprocessing in French and English. This library accepts unstructured text as input and provides the preprocessed text as output, based on a set of rules and on a base of stop words for both languages. The proposed library has been made on different corpora and gave an interesting result.

Keywords: text preprocessing, segmentation, knowledge extraction, normalization, text generation, information retrieval

Procedia PDF Downloads 94
12835 Study of Chemical Compounds of Garlic

Authors: A. B. Bazaralieva, A. A. Turgumbayeva

Abstract:

The phytosubstance from garlic was obtained by extraction with liquid carbon dioxide under critical conditions. Methods of processing raw materials are proposed, and the chemical composition of garlic is studied by gas chromatography and mass spectrometry. The garlic extract's composition was determined using gas chromatography (GC) and gas chromatography-mass spectrophotometry (GC-MS). The phytosubstance had 54 constituents. The extract included the following main compounds: Manool (39.56%), Viridifrolol (7%), Podocarpa-1,8,11,13-tetraen-3-one, 14-isopropyl-1,13-dimethoxy- 5,15 percent, (+)-2-Bornanone (4.29%), Thujone (3.49%), Linolic acid ethyl ester (3.41%), and 12-O-Methylcarn.

Keywords: Allium sativum, bioactive compounds of garlic, carbon dioxide extraction of garlic, GS-MS method

Procedia PDF Downloads 112
12834 On Musical Information Geometry with Applications to Sonified Image Analysis

Authors: Shannon Steinmetz, Ellen Gethner

Abstract:

In this paper, a theoretical foundation is developed for patterned segmentation of audio using the geometry of music and statistical manifold. We demonstrate image content clustering using conic space sonification. The algorithm takes a geodesic curve as a model estimator of the three-parameter Gamma distribution. The random variable is parameterized by musical centricity and centric velocity. Model parameters predict audio segmentation in the form of duration and frame count based on the likelihood of musical geometry transition. We provide an example using a database of randomly selected images, resulting in statistically significant clusters of similar image content.

Keywords: sonification, musical information geometry, image, content extraction, automated quantification, audio segmentation, pattern recognition

Procedia PDF Downloads 237
12833 Waters Colloidal Phase Extraction and Preconcentration: Method Comparison

Authors: Emmanuelle Maria, Pierre Crançon, Gaëtane Lespes

Abstract:

Colloids are ubiquitous in the environment and are known to play a major role in enhancing the transport of trace elements, thus being an important vector for contaminants dispersion. Colloids study and characterization are necessary to improve our understanding of the fate of pollutants in the environment. However, in stream water and groundwater, colloids are often very poorly concentrated. It is therefore necessary to pre-concentrate colloids in order to get enough material for analysis, while preserving their initial structure. Many techniques are used to extract and/or pre-concentrate the colloidal phase from bulk aqueous phase, but yet there is neither reference method nor estimation of the impact of these different techniques on the colloids structure, as well as the bias introduced by the separation method. In the present work, we have tested and compared several methods of colloidal phase extraction/pre-concentration, and their impact on colloids properties, particularly their size distribution and their elementary composition. Ultrafiltration methods (frontal, tangential and centrifugal) have been considered since they are widely used for the extraction of colloids in natural waters. To compare these methods, a ‘synthetic groundwater’ was used as a reference. The size distribution (obtained by Field-Flow Fractionation (FFF)) and the chemical composition of the colloidal phase (obtained by Inductively Coupled Plasma Mass Spectrometry (ICPMS) and Total Organic Carbon analysis (TOC)) were chosen as comparison factors. In this way, it is possible to estimate the pre-concentration impact on the colloidal phase preservation. It appears that some of these methods preserve in a more efficient manner the colloidal phase composition while others are easier/faster to use. The choice of the extraction/pre-concentration method is therefore a compromise between efficiency (including speed and ease of use) and impact on the structural and chemical composition of the colloidal phase. In perspective, the use of these methods should enhance the consideration of colloidal phase in the transport of pollutants in environmental assessment studies and forensics.

Keywords: chemical composition, colloids, extraction, preconcentration methods, size distribution

Procedia PDF Downloads 215
12832 Precious and Rare Metals in Overburden Carbonaceous Rocks: Methods of Extraction

Authors: Tatyana Alexandrova, Alexandr Alexandrov, Nadezhda Nikolaeva

Abstract:

A problem of complex mineral resources development is urgent and priority, it is aimed at realization of the processes of their ecologically safe development, one of its components is revealing the influence of the forms of element compounds in raw materials and in the processing products. In view of depletion of the precious metal reserves at the traditional deposits in the XXI century the large-size open cast deposits, localized in black shale strata begin to play the leading role. Carbonaceous (black) shales carry a heightened metallogenic potential. Black shales with high content of carbon are widely distributed within the scope of Bureinsky massif. According to academician Hanchuk`s data black shales of Sutirskaya series contain generally PGEs native form. The presence of high absorptive towards carbonaceous matter gold and PGEs compounds in crude ore results in decrease of valuable components extraction because of their sorption into dissipated carbonaceous matter.

Keywords: сarbonaceous rocks, bitumens, precious metals, concentration, extraction

Procedia PDF Downloads 246
12831 Information Literacy: Concept and Importance

Authors: Gaurav Kumar

Abstract:

An information literate person is one who uses information effectively in all its forms. When presented with questions or problems, an information literate person would know what information to look for, how to search efficiently and be able to access relevant sources. In addition, an information literate person would have the ability to evaluate and select appropriate information sources and to use the information effectively and ethically to answer questions or solve problems. Information literacy has become an important element in higher education. The information literacy movement has internationally recognized standards and learning outcomes. The step-by-step process of achieving information literacy is particularly crucial in an era where knowledge could be disseminated through a variety of media. What is the relationship between information literacy as we define it in higher education and information literacy among non-academic populations? What forces will change how we think about the definition of information literacy in the future and how we will apply the definition in all environments?

Keywords: information literacy, human beings, visual media and computer network etc, information literacy

Procedia PDF Downloads 339
12830 Hydrometallurgical Processing of a Nigerian Chalcopyrite Ore

Authors: Alafara A. Baba, Kuranga I. Ayinla, Folahan A. Adekola, Rafiu B. Bale

Abstract:

Due to increasing demands and diverse applications of copper oxide as pigment in ceramics, cuprammonium hydroxide solution for rayon, p-type semi-conductor, dry cell batteries production and as safety disposal of hazardous materials, a study on the hydrometallurgical operations involving leaching, solvent extraction and precipitation for the recovery of copper for producing high grade copper oxide from a Nigerian chalcopyrite ore in chloride media has been examined. At a particular set of experimental parameter with respect to acid concentration, reaction temperature and particle size, the leaching investigation showed that the ore dissolution increases with increasing acid concentration, temperature and decreasing particle diameter at a moderate stirring. The kinetics data has been analyzed and was found to follow diffusion control mechanism. At optimal conditions, the extent of ore dissolution reached 94.3%. The recovery of the total copper from the hydrochloric acid-leached chalcopyrite ore was undertaken by solvent extraction and precipitation techniques, prior to the beneficiation of the purified solution as copper oxide. The purification of the leach liquor was firstly done by precipitation of total iron and manganese using Ca(OH)2 and H2O2 as oxidizer at pH 3.5 and 4.25, respectively. An extraction efficiency of 97.3% total copper was obtained by 0.2 mol/L Dithizone in kerosene at 25±2ºC within 40 minutes, from which ≈98% Cu from loaded organic phase was successfully stripped by 0.1 mol/L HCl solution. The beneficiation of the recovered pure copper solution was carried out by crystallization through alkali addition followed by calcination at 600ºC to obtain high grade copper oxide (Tenorite, CuO: 05-0661). Finally, a simple hydrometallurgical scheme for the operational extraction procedure amenable for industrial utilization and economic sustainability was provided.

Keywords: chalcopyrite ore, Nigeria, copper, copper oxide, solvent extraction

Procedia PDF Downloads 393
12829 Extraction of Cellulose Nanocrystals from Soy Pods

Authors: Maycon dos Santos, Marivane Turim Koschevic, Karina Sayuri Ueda, Marcello Lima Bertuci, Farayde Matta Fackhouri, Silvia Maria Martelli

Abstract:

The use of cellulose nanocrystals as reinforcing agents in polymer nanocomposites is promising. In this study, we tested four different methods of mercerization were divided into two stages. The sample was treated in 5% NaOH solution for 30 minutes at 50 ° C in the first stage and 30vol H2O2 for 2 hours at 50 ° C in the second step, which showed better results. For the extraction of the sample obtained nanocrystals positive result was that the solution was treated with H2SO4 60% (w / w) for 1 hour at 50 ° C. The results were positive and showed that it is possible to extract CNC at low temperatures.

Keywords: soy pods, cellulose nanocrystals, temperature, acid concentration

Procedia PDF Downloads 297
12828 Characteristic Sentence Stems in Academic English Texts: Definition, Identification, and Extraction

Authors: Jingjie Li, Wenjie Hu

Abstract:

Phraseological units in academic English texts have been a central focus in recent corpus linguistic research. A wide variety of phraseological units have been explored, including collocations, chunks, lexical bundles, patterns, semantic sequences, etc. This paper describes a special category of clause-level phraseological units, namely, Characteristic Sentence Stems (CSSs), with a view to describing their defining criteria and extraction method. CSSs are contiguous lexico-grammatical sequences which contain a subject-predicate structure and which are frame expressions characteristic of academic writing. The extraction of CSSs consists of six steps: Part-of-speech tagging, n-gram segmentation, structure identification, significance of occurrence calculation, text range calculation, and overlapping sequence reduction. Significance of occurrence calculation is the crux of this study. It includes the computing of both the internal association and the boundary independence of a CSS and tests the occurring significance of the CSS from both inside and outside perspectives. A new normalization algorithm is also introduced into the calculation of LocalMaxs for reducing overlapping sequences. It is argued that many sentence stems are so recurrent in academic texts that the most typical of them have become the habitual ways of making meaning in academic writing. Therefore, studies of CSSs could have potential implications and reference value for academic discourse analysis, English for Academic Purposes (EAP) teaching and writing.

Keywords: characteristic sentence stem, extraction method, phraseological unit, the statistical measure

Procedia PDF Downloads 166
12827 Generating Insights from Data Using a Hybrid Approach

Authors: Allmin Susaiyah, Aki Härmä, Milan Petković

Abstract:

Automatic generation of insights from data using insight mining systems (IMS) is useful in many applications, such as personal health tracking, patient monitoring, and business process management. Existing IMS face challenges in controlling insight extraction, scaling to large databases, and generalising to unseen domains. In this work, we propose a hybrid approach consisting of rule-based and neural components for generating insights from data while overcoming the aforementioned challenges. Firstly, a rule-based data 2CNL component is used to extract statistically significant insights from data and represent them in a controlled natural language (CNL). Secondly, a BERTSum-based CNL2NL component is used to convert these CNLs into natural language texts. We improve the model using task-specific and domain-specific fine-tuning. Our approach has been evaluated using statistical techniques and standard evaluation metrics. We overcame the aforementioned challenges and observed significant improvement with domain-specific fine-tuning.

Keywords: data mining, insight mining, natural language generation, pre-trained language models

Procedia PDF Downloads 119
12826 Skull Extraction for Quantification of Brain Volume in Magnetic Resonance Imaging of Multiple Sclerosis Patients

Authors: Marcela De Oliveira, Marina P. Da Silva, Fernando C. G. Da Rocha, Jorge M. Santos, Jaime S. Cardoso, Paulo N. Lisboa-Filho

Abstract:

Multiple Sclerosis (MS) is an immune-mediated disease of the central nervous system characterized by neurodegeneration, inflammation, demyelination, and axonal loss. Magnetic resonance imaging (MRI), due to the richness in the information details provided, is the gold standard exam for diagnosis and follow-up of neurodegenerative diseases, such as MS. Brain atrophy, the gradual loss of brain volume, is quite extensive in multiple sclerosis, nearly 0.5-1.35% per year, far off the limits of normal aging. Thus, the brain volume quantification becomes an essential task for future analysis of the occurrence atrophy. The analysis of MRI has become a tedious and complex task for clinicians, who have to manually extract important information. This manual analysis is prone to errors and is time consuming due to various intra- and inter-operator variability. Nowadays, computerized methods for MRI segmentation have been extensively used to assist doctors in quantitative analyzes for disease diagnosis and monitoring. Thus, the purpose of this work was to evaluate the brain volume in MRI of MS patients. We used MRI scans with 30 slices of the five patients diagnosed with multiple sclerosis according to the McDonald criteria. The computational methods for the analysis of images were carried out in two steps: segmentation of the brain and brain volume quantification. The first image processing step was to perform brain extraction by skull stripping from the original image. In the skull stripper for MRI images of the brain, the algorithm registers a grayscale atlas image to the grayscale patient image. The associated brain mask is propagated using the registration transformation. Then this mask is eroded and used for a refined brain extraction based on level-sets (edge of the brain-skull border with dedicated expansion, curvature, and advection terms). In the second step, the brain volume quantification was performed by counting the voxels belonging to the segmentation mask and converted in cc. We observed an average brain volume of 1469.5 cc. We concluded that the automatic method applied in this work can be used for the brain extraction process and brain volume quantification in MRI. The development and use of computer programs can contribute to assist health professionals in the diagnosis and monitoring of patients with neurodegenerative diseases. In future works, we expect to implement more automated methods for the assessment of cerebral atrophy and brain lesions quantification, including machine-learning approaches. Acknowledgements: This work was supported by a grant from Brazilian agency Fundação de Amparo à Pesquisa do Estado de São Paulo (number 2019/16362-5).

Keywords: brain volume, magnetic resonance imaging, multiple sclerosis, skull stripper

Procedia PDF Downloads 146
12825 A Summary-Based Text Classification Model for Graph Attention Networks

Authors: Shuo Liu

Abstract:

In Chinese text classification tasks, redundant words and phrases can interfere with the formation of extracted and analyzed text information, leading to a decrease in the accuracy of the classification model. To reduce irrelevant elements, extract and utilize text content information more efficiently and improve the accuracy of text classification models. In this paper, the text in the corpus is first extracted using the TextRank algorithm for abstraction, the words in the abstract are used as nodes to construct a text graph, and then the graph attention network (GAT) is used to complete the task of classifying the text. Testing on a Chinese dataset from the network, the classification accuracy was improved over the direct method of generating graph structures using text.

Keywords: Chinese natural language processing, text classification, abstract extraction, graph attention network

Procedia PDF Downloads 100
12824 Advantages of Matrix Solid Phase Dispersive (MSPD) Extraction Associated to MIPS versus MAE Liquid Extraction for the Simultaneous Analysis of PAHs, PCBs and Some Hydroxylated PAHs in Sediments

Authors: F. Portet-Koltalo, Y. Tian, I. Berger, C. Boulanger-Lecomte, A. Benamar, N. Machour

Abstract:

Sediments are complex environments which can accumulate a great variety of persistent toxic contaminants such as polychlorobiphenyles (PCBs), polycyclic aromatic hydrocarbons (PAHs) and some of their more toxic degradation metabolites such as hydroxylated PAHs (OH-PAHs). Owing to their composition, fine clayey sediments can be more difficult to extract than soils using conventional solvent extraction processes. So this study aimed to compare the potential of MSPD (matrix solid phase dispersive extraction) to extract PCBs, PAHs and OH-PAHs, in comparison with microwave assisted extraction (MAE). Methodologies: MAE extraction with various solvent mixtures was used to extract PCBs, PAHs and OH-PAHs from sediments in two runs, followed by two GC-MS analyses. MSPD consisted in crushing the dried sediment with dispersive agents, introducing the mixture in cartridges and eluting the target compounds with an appropriate volume of selected solvents. So MSPD combined with cartridges containing MIPs (molecularly imprinted polymers) designed for OH-PAHs was used to extract the three families of target compounds in only one run, followed by parallel analyses in GC-MS for PAHs/PCBs and HPLC-FLD for OH-PAHs. Results: MAE extraction was optimized to extract from clayey sediments, in two runs, PAHs/PCBs in one hand and OH-PAHs in the other hand. Indeed, the best conditions of extractions (mixtures of extracting solvents, temperature) were different if we consider the polarity and the thermodegradability of the different families of target contaminants: PAHs/PCBs were better extracted using an acetone/toluene 50/50 mixture at 130°C whereas OH-PAHs were better extracted using an acetonitrile/toluene 90/10 mixture at 100°C. Moreover, the two consecutive GC-MS analyses contributed to double the total analysis time. A matrix solid phase dispersive (MSPD) extraction procedure was also optimized, with the first objective of increasing the extraction recovery yields of PAHs and PCBs from fine-grained sediment. The crushing time (2-10 min), the nature of the dispersing agents added for purifying and increasing the extraction yields (Florisil, octadecylsilane, 3-chloropropyle, 4-benzylchloride), the nature and the volume of eluting solvents (methylene chloride, hexane, hexane/acetone…) were studied. It appeared that in the best conditions, MSPD was a better extraction method than MAE for PAHs and PCBs, with respectively, mean increases of 8.2% and 71%. This method was also faster, easier and less expensive. But the other advantage of MSPD was that it allowed to introduce easily, just after the first elution process of PAHs/PCBs, a step permitting the selective recovery of OH-PAHs. A cartridge containing MIPs designed for phenols was coupled to the cartridge containing the dispersed sediment, and various eluting solvents, different from those used for PAHs and PCBs, were tested to selectively concentrate and extract OH-PAHs. Thereafter OH-PAHs could be analyzed at the same time than PAHs and PCBs: the OH-PAH extract could be analyzed with HPLC-FLD, whereas the PAHs/PCBs extract was analyzed with GC-MS, adding only few minutes more to the total duration of the analytical process. Conclusion: MSPD associated to MIPs appeared to be an easy, fast and low expensive method, able to extract in one run a complex mixture of toxic apolar and more polar contaminants present in clayey fine-grained sediments, an environmental matrix which is generally difficult to analyze.

Keywords: contaminated fine-grained sediments, matrix solid phase dispersive extraction, microwave assisted extraction, molecularly imprinted polymers, multi-pollutant analysis

Procedia PDF Downloads 353
12823 Determining the Effective Substance of Cottonseed Extract on the Treatment of Leishmaniasis

Authors: Mehrosadat Mirmohammadi, Sara Taghdisi, Ali Padash, Mohammad Hossein Pazandeh

Abstract:

Gossypol, a yellowish anti-nutritional compound found in cotton plants, exists in various plant parts, including seeds, husks, leaves, and stems. Chemically, gossypol is a potent polyphenolic aldehyde with antioxidant and therapeutic properties. However, its free form can be toxic, posing risks to both humans and animals. Initially, we extracted gossypol from cotton seeds using n-hexane as a solvent (yield: 84.0 ± 4.0%). We also obtained cotton seed and cotton boll extracts via Soxhlet extraction (25:75 hydroalcoholic ratio). These extracts, combined with cornstarch, formed four herbal medicinal formulations. Ethical approval allowed us to investigate their effects on Leishmania-caused skin wounds, comparing them to glucantime (local ampoule). Herbal formulas outperformed the control group (ethanol only) in wound treatment (p-value 0.05). The average wound diameter after two months did not significantly differ between plant extract ointments and topical glucantime. Notably, cotton boll extract with 1% extra gossypol crystal showed the best therapeutic effect. We extracted gossypol from cotton seeds using n-hexane via Soxhlet extraction. Saponification, acidification, and recrystallization steps followed. FTIR, UV-Vis, and HPLC analyses confirmed the product’s identity. Herbal medicines from cotton seeds effectively treated chronic wounds compared to the ethanol-only control group. Wound diameter differed significantly between extract ointments and glucantime injections. It seems that due to the presence of large amounts of fat in the oil, the extraction of gossypol from it faces many obstacles. The extraction of this compound with our technique showed that extraction from oil has a higher efficiency, perhaps because of the preparation of oil by cold pressing method, the possibility of losing this compound is much less than when extraction is done with Soxhlet. On the other hand, the gossypol in the oil is mostly bound to the protein, which somehow protects the gossypol until the last stage of the extraction process. Since this compound is very sensitive to light and heat, it was extracted as a derivative with acetic acid. Also, in the treatment section, it was found that the ointment prepared with the extract is more effective and Gossypol is one of the effective ingredients in the treatment. Therefore, gossypol can be extracted from the oil and added to the extract from which gossypol has been extracted to make an effective medicine with a certain dose.

Keywords: cottonseed, glucantime, gossypol, leishmaniasis

Procedia PDF Downloads 61
12822 Ultrasound/Microwave Assisted Extraction Recovery and Identification of Bioactive Compounds (Polyphenols) from Tarbush (Fluorensia cernua)

Authors: Marisol Rodriguez-Duarte, Aide Saenz-Galindo, Carolina Flores-Gallegos, Raul Rodriguez-Herrera, Juan Ascacio-Valdes

Abstract:

The plant known as tarbush (Fluorensia cernua) is a plant originating in northern Mexico, mainly in the states of Coahuila, Durango, San Luis Potosí, Zacatecas and Chihuahua. It is a branched shrub that belongs to the family Asteraceae, has oval leaves of 6 to 11 cm in length and also has small yellow flowers. In Mexico, the tarbush is a very appreciated plant because it has been used as a traditional medicinal agent, for the treatment of gastrointestinal diseases, skin infections and as a healing agent. This plant has been used mainly as an infusion. Due to its traditional use, the content and type of phytochemicals present in the plant are currently unknown and are responsible for its biological properties, so its recovery and identification is very important because the compounds that it contains have relevant applications in the field of food, pharmaceuticals and medicine. The objective of this work was to determine the best extraction condition of phytochemical compounds (mainly polyphenolic compounds) from the leaf using ultrasound/microwave assisted extraction (U/M-AE). To reach the objective, U/M-AE extractions were performed evaluating three mass/volume ratios (1:8, 1:12, 1:16), three ethanol/water solvent concentrations (0%, 30% and 70%), ultrasound extraction time of 20 min and 5 min at 70°C of microwave treatment. All experiments were performed using a fractional factorial experimental design. Once the best extraction condition was defined, the compounds were recovered by liquid column chromatography using Amberlite XAD-16, the polyphenolic fraction was recovered with ethanol and then evaporated. The recovered polyphenolic compounds were quantified by spectrophotometric techniques and identified by HPLC/ESI/MS. The results obtained showed that the best extraction condition of the compounds was using a mass/volume ratio of 1:8 and solvent ethanol/water concentration of 70%. The concentration obtained from polyphenolic compounds using this condition was 22.74 mg/g and finally, 16 compounds of polyphenolic origin were identified. The results obtained in this work allow us to postulate the Mexican plant known as tarbush as a relevant source of bioactive polyphenolic compounds of food, pharmaceutical and medicinal interest.

Keywords: U/M-AE, tarbush, polyphenols, identification

Procedia PDF Downloads 163
12821 Automated Feature Extraction and Object-Based Detection from High-Resolution Aerial Photos Based on Machine Learning and Artificial Intelligence

Authors: Mohammed Al Sulaimani, Hamad Al Manhi

Abstract:

With the development of Remote Sensing technology, the resolution of optical Remote Sensing images has greatly improved, and images have become largely available. Numerous detectors have been developed for detecting different types of objects. In the past few years, Remote Sensing has benefited a lot from deep learning, particularly Deep Convolution Neural Networks (CNNs). Deep learning holds great promise to fulfill the challenging needs of Remote Sensing and solving various problems within different fields and applications. The use of Unmanned Aerial Systems in acquiring Aerial Photos has become highly used and preferred by most organizations to support their activities because of their high resolution and accuracy, which make the identification and detection of very small features much easier than Satellite Images. And this has opened an extreme era of Deep Learning in different applications not only in feature extraction and prediction but also in analysis. This work addresses the capacity of Machine Learning and Deep Learning in detecting and extracting Oil Leaks from Flowlines (Onshore) using High-Resolution Aerial Photos which have been acquired by UAS fixed with RGB Sensor to support early detection of these leaks and prevent the company from the leak’s losses and the most important thing environmental damage. Here, there are two different approaches and different methods of DL have been demonstrated. The first approach focuses on detecting the Oil Leaks from the RAW Aerial Photos (not processed) using a Deep Learning called Single Shoot Detector (SSD). The model draws bounding boxes around the leaks, and the results were extremely good. The second approach focuses on detecting the Oil Leaks from the Ortho-mosaiced Images (Georeferenced Images) by developing three Deep Learning Models using (MaskRCNN, U-Net and PSP-Net Classifier). Then, post-processing is performed to combine the results of these three Deep Learning Models to achieve a better detection result and improved accuracy. Although there is a relatively small amount of datasets available for training purposes, the Trained DL Models have shown good results in extracting the extent of the Oil Leaks and obtaining excellent and accurate detection.

Keywords: GIS, remote sensing, oil leak detection, machine learning, aerial photos, unmanned aerial systems

Procedia PDF Downloads 33
12820 Comparative Analysis of Feature Extraction and Classification Techniques

Authors: R. L. Ujjwal, Abhishek Jain

Abstract:

In the field of computer vision, most facial variations such as identity, expression, emotions and gender have been extensively studied. Automatic age estimation has been rarely explored. With age progression of a human, the features of the face changes. This paper is providing a new comparable study of different type of algorithm to feature extraction [Hybrid features using HAAR cascade & HOG features] & classification [KNN & SVM] training dataset. By using these algorithms we are trying to find out one of the best classification algorithms. Same thing we have done on the feature selection part, we extract the feature by using HAAR cascade and HOG. This work will be done in context of age group classification model.

Keywords: computer vision, age group, face detection

Procedia PDF Downloads 368
12819 Local Directional Encoded Derivative Binary Pattern Based Coral Image Classification Using Weighted Distance Gray Wolf Optimization Algorithm

Authors: Annalakshmi G., Sakthivel Murugan S.

Abstract:

This paper presents a local directional encoded derivative binary pattern (LDEDBP) feature extraction method that can be applied for the classification of submarine coral reef images. The classification of coral reef images using texture features is difficult due to the dissimilarities in class samples. In coral reef image classification, texture features are extracted using the proposed method called local directional encoded derivative binary pattern (LDEDBP). The proposed approach extracts the complete structural arrangement of the local region using local binary batten (LBP) and also extracts the edge information using local directional pattern (LDP) from the edge response available in a particular region, thereby achieving extra discriminative feature value. Typically the LDP extracts the edge details in all eight directions. The process of integrating edge responses along with the local binary pattern achieves a more robust texture descriptor than the other descriptors used in texture feature extraction methods. Finally, the proposed technique is applied to an extreme learning machine (ELM) method with a meta-heuristic algorithm known as weighted distance grey wolf optimizer (GWO) to optimize the input weight and biases of single-hidden-layer feed-forward neural networks (SLFN). In the empirical results, ELM-WDGWO demonstrated their better performance in terms of accuracy on all coral datasets, namely RSMAS, EILAT, EILAT2, and MLC, compared with other state-of-the-art algorithms. The proposed method achieves the highest overall classification accuracy of 94% compared to the other state of art methods.

Keywords: feature extraction, local directional pattern, ELM classifier, GWO optimization

Procedia PDF Downloads 163
12818 Towards Logical Inference for the Arabic Question-Answering

Authors: Wided Bakari, Patrice Bellot, Omar Trigui, Mahmoud Neji

Abstract:

This article constitutes an opening to think of the modeling and analysis of Arabic texts in the context of a question-answer system. It is a question of exceeding the traditional approaches focused on morphosyntactic approaches. Furthermore, we present a new approach that analyze a text in order to extract correct answers then transform it to logical predicates. In addition, we would like to represent different levels of information within a text to answer a question and choose an answer among several proposed. To do so, we transform both the question and the text into logical forms. Then, we try to recognize all entailment between them. The results of recognizing the entailment are a set of text sentences that can implicate the user’s question. Our work is now concentrated on an implementation step in order to develop a system of question-answering in Arabic using techniques to recognize textual implications. In this context, the extraction of text features (keywords, named entities, and relationships that link them) is actually considered the first step in our process of text modeling. The second one is the use of techniques of textual implication that relies on the notion of inference and logic representation to extract candidate answers. The last step is the extraction and selection of the desired answer.

Keywords: NLP, Arabic language, question-answering, recognition text entailment, logic forms

Procedia PDF Downloads 342
12817 Feature Evaluation Based on Random Subspace and Multiple-K Ensemble

Authors: Jaehong Yu, Seoung Bum Kim

Abstract:

Clustering analysis can facilitate the extraction of intrinsic patterns in a dataset and reveal its natural groupings without requiring class information. For effective clustering analysis in high dimensional datasets, unsupervised dimensionality reduction is an important task. Unsupervised dimensionality reduction can generally be achieved by feature extraction or feature selection. In many situations, feature selection methods are more appropriate than feature extraction methods because of their clear interpretation with respect to the original features. The unsupervised feature selection can be categorized as feature subset selection and feature ranking method, and we focused on unsupervised feature ranking methods which evaluate the features based on their importance scores. Recently, several unsupervised feature ranking methods were developed based on ensemble approaches to achieve their higher accuracy and stability. However, most of the ensemble-based feature ranking methods require the true number of clusters. Furthermore, these algorithms evaluate the feature importance depending on the ensemble clustering solution, and they produce undesirable evaluation results if the clustering solutions are inaccurate. To address these limitations, we proposed an ensemble-based feature ranking method with random subspace and multiple-k ensemble (FRRM). The proposed FRRM algorithm evaluates the importance of each feature with the random subspace ensemble, and all evaluation results are combined with the ensemble importance scores. Moreover, FRRM does not require the determination of the true number of clusters in advance through the use of the multiple-k ensemble idea. Experiments on various benchmark datasets were conducted to examine the properties of the proposed FRRM algorithm and to compare its performance with that of existing feature ranking methods. The experimental results demonstrated that the proposed FRRM outperformed the competitors.

Keywords: clustering analysis, multiple-k ensemble, random subspace-based feature evaluation, unsupervised feature ranking

Procedia PDF Downloads 339
12816 Training a Neural Network to Segment, Detect and Recognize Numbers

Authors: Abhisek Dash

Abstract:

This study had three neural networks, one for number segmentation, one for number detection and one for number recognition all of which are coupled to one another. All networks were trained on the MNIST dataset and were convolutional. It was assumed that the images had lighter background and darker foreground. The segmentation network took 28x28 images as input and had sixteen outputs. Segmentation training starts when a dark pixel is encountered. Taking a window(7x7) over that pixel as focus, the eight neighborhood of the focus was checked for further dark pixels. The segmentation network was then trained to move in those directions which had dark pixels. To this end the segmentation network had 16 outputs. They were arranged as “go east”, ”don’t go east ”, “go south east”, “don’t go south east”, “go south”, “don’t go south” and so on w.r.t focus window. The focus window was resized into a 28x28 image and the network was trained to consider those neighborhoods which had dark pixels. The neighborhoods which had dark pixels were pushed into a queue in a particular order. The neighborhoods were then popped one at a time stitched to the existing partial image of the number one at a time and trained on which neighborhoods to consider when the new partial image was presented. The above process was repeated until the image was fully covered by the 7x7 neighborhoods and there were no more uncovered black pixels. During testing the network scans and looks for the first dark pixel. From here on the network predicts which neighborhoods to consider and segments the image. After this step the group of neighborhoods are passed into the detection network. The detection network took 28x28 images as input and had two outputs denoting whether a number was detected or not. Since the ground truth of the bounds of a number was known during training the detection network outputted in favor of number not found until the bounds were not met and vice versa. The recognition network was a standard CNN that also took 28x28 images and had 10 outputs for recognition of numbers from 0 to 9. This network was activated only when the detection network votes in favor of number detected. The above methodology could segment connected and overlapping numbers. Additionally the recognition unit was only invoked when a number was detected which minimized false positives. It also eliminated the need for rules of thumb as segmentation is learned. The strategy can also be extended to other characters as well.

Keywords: convolutional neural networks, OCR, text detection, text segmentation

Procedia PDF Downloads 161
12815 Knowledge of Quality Assurance and Quality Control in Mammography; A Study among Radiographers of Mammography Settings in Sri Lanka

Authors: H. S. Niroshani, W. M. Ediri Arachchi, R. Tudugala, U. J. M. A. L. Jayasinghe, U. M. U. J. Jayasekara, P. B. Hewavithana

Abstract:

Mammography is used as a screening tool for early diagnosis of breast cancer. It is also useful in refining the diagnosis of breast cancer either by assessment or work up after a suspicious area in the breast has been detected. In order to detect breast cancer accurately and at the earliest possible stage, the image must have an optimum contrast to reveal mass densities and spiculated fibrous structures radiating from them. In addition, the spatial resolution must be adequate to reveal the suffusion of micro calcifications and their shape. The above factors can be optimized by implementing an effective QA programme to enhance the accurate diagnosis of mammographic imaging. Therefore, the radiographer’s knowledge on QA is greatly instrumental in routine mammographic practice. The aim of this study was to assess the radiographer’s knowledge on Quality Assurance and Quality Control programmes in relation to mammographic procedures. A cross-sectional study was carried out among all radiographers working in each mammography setting in Sri Lanka. Pre-tested, anonymous self-administered questionnaires were circulated among the study population and duly filled questionnaires returned within a period of three months were taken into the account. The data on demographical information, knowledge on QA programme and associated QC tests, overall knowledge on QA and QC programmes were obtained. Data analysis was performed using IBM SPSS statistical software (version 20.0). The total response rate was 59.6% and the average knowledge score was 54.15±11.29 SD out of 100. Knowledge was compared on the basis of education level, special training of mammography, and the years of working experience in a mammographic setting of the individuals. Out of 31 subjects, 64.5% (n=20) were graduate radiographers and 35.5% (n=11) were diploma holders while 83.9% (n=26) of radiographers have been specially trained for mammography and 16.1% (n=5) have not been attended for any special training for mammography. It is also noted that 58.1% (n=18) of individuals possessed their experience of less than one year and rest 41.9% (n=13) of them were greater than that. Further, the results found that there is a significant difference (P < 0.05) in the knowledge of QA and overall knowledge on QA and QC programme in the categories of education level and working experience. Also, results imply that there was a significant difference (P < 0.05) in the knowledge of QC test among the groups of trained and non-trained radiographers. This study reveals that education level, working experience and the training obtained particularly in the field of mammography have a significant impact on their knowledge on QA and QC in mammography.

Keywords: knowledge, mammography, quality assurance, quality control

Procedia PDF Downloads 330
12814 A Quality Index Optimization Method for Non-Invasive Fetal ECG Extraction

Authors: Lucia Billeci, Gennaro Tartarisco, Maurizio Varanini

Abstract:

Fetal cardiac monitoring by fetal electrocardiogram (fECG) can provide significant clinical information about the healthy condition of the fetus. Despite this potentiality till now the use of fECG in clinical practice has been quite limited due to the difficulties in its measuring. The recovery of fECG from the signals acquired non-invasively by using electrodes placed on the maternal abdomen is a challenging task because abdominal signals are a mixture of several components and the fetal one is very weak. This paper presents an approach for fECG extraction from abdominal maternal recordings, which exploits the characteristics of pseudo-periodicity of fetal ECG. It consists of devising a quality index (fQI) for fECG and of finding the linear combinations of preprocessed abdominal signals, which maximize these fQI (quality index optimization - QIO). It aims at improving the performances of the most commonly adopted methods for fECG extraction, usually based on maternal ECG (mECG) estimating and canceling. The procedure for the fECG extraction and fetal QRS (fQRS) detection is completely unsupervised and based on the following steps: signal pre-processing; maternal ECG (mECG) extraction and maternal QRS detection; mECG component approximation and canceling by weighted principal component analysis; fECG extraction by fQI maximization and fetal QRS detection. The proposed method was compared with our previously developed procedure, which obtained the highest at the Physionet/Computing in Cardiology Challenge 2013. That procedure was based on removing the mECG from abdominal signals estimated by a principal component analysis (PCA) and applying the Independent component Analysis (ICA) on the residual signals. Both methods were developed and tuned using 69, 1 min long, abdominal measurements with fetal QRS annotation of the dataset A provided by PhysioNet/Computing in Cardiology Challenge 2013. The QIO-based and the ICA-based methods were compared in analyzing two databases of abdominal maternal ECG available on the Physionet site. The first is the Abdominal and Direct Fetal Electrocardiogram Database (ADdb) which contains the fetal QRS annotations thus allowing a quantitative performance comparison, the second is the Non-Invasive Fetal Electrocardiogram Database (NIdb), which does not contain the fetal QRS annotations so that the comparison between the two methods can be only qualitative. In particular, the comparison on NIdb was performed defining an index of quality for the fetal RR series. On the annotated database ADdb the QIO method, provided the performance indexes Sens=0.9988, PPA=0.9991, F1=0.9989 overcoming the ICA-based one, which provided Sens=0.9966, PPA=0.9972, F1=0.9969. The comparison on NIdb was performed defining an index of quality for the fetal RR series. The index of quality resulted higher for the QIO-based method compared to the ICA-based one in 35 records out 55 cases of the NIdb. The QIO-based method gave very high performances with both the databases. The results of this study foresees the application of the algorithm in a fully unsupervised way for the implementation in wearable devices for self-monitoring of fetal health.

Keywords: fetal electrocardiography, fetal QRS detection, independent component analysis (ICA), optimization, wearable

Procedia PDF Downloads 280
12813 Density Determination by Dilution for Extra Heavy Oil Residues Obtained Using Molecular Distillation and Supercritical Fluid Extraction as Upgrading and Refining Process

Authors: Oscar Corredor, Alexander Guzman, Adan Leon

Abstract:

Density is a bulk physical property that indicates the quality of a petroleum fraction. It is also a useful property to estimate various physicochemical properties of fraction and petroleum fluids; however, the determination of density of extra heavy residual (EHR) fractions by standard methodologies, (ASTM D70) shows limitations for samples with higher densities than 1.0879 g/cm3. For this reason, a dilution methodology was developed in order to determinate density for those particular fractions, 87 (EHR) fractions were obtained as products of the fractionation of Colombian typical Vacuum Distillation Residual Fractions using molecular distillation (MD) and extraction with Solvent N-hexane in Supercritical Conditions (SFEF) pilot plants. The proposed methodology showed reliable results that can be demonstrated with the standard deviation of repeatability and reproducibility values of 0.0031 and 0.0061 g/ml respectively. In the same way, it was possible to determine densities in fractions EHR up to 1.1647g/cm3 and °API values obtained were ten times less than the water reference value.

Keywords: API, density, vacuum residual, molecular distillation, supercritical fluid extraction

Procedia PDF Downloads 267
12812 Breakthrough Highly-Effective Extraction of Perfluoroctanoic Acid Using Natural Deep Eutectic Solvents

Authors: Sana Eid, Ahmad S. Darwish, Tarek Lemaoui, Maguy Abi Jaoude, Fawzi Banat, Shadi W. Hasan, Inas M. AlNashef

Abstract:

Addressing the growing challenge of per- and polyfluoroalkyl substances (PFAS) pollution in water bodies, this study introduces natural deep eutectic solvents (NADESs) as a pioneering solution for the efficient extraction of perfluorooctanoic acid (PFOA), one of the most persistent and concerning PFAS pollutants. Among the tested NADESs, trioctylphosphine oxide: lauric acid (TOPO:LauA) in a 1:1 molar ratio was distinguished as the most effective, achieving an extraction efficiency of approximately 99.52% at a solvent-to-feed (S:F) ratio of 1:2, room temperature, and neutral pH. This efficiency is achieved within a notably short mixing time of only one min, which is significantly less than the time required by conventional methods, underscoring the potential of TOPO:LauA for rapid and effective PFAS remediation. TOPO:LauA maintained consistent performance across various operational parameters, including a range of initial PFOA concentrations (0.1 ppm to 1000 ppm), temperatures (15 °C to 100 °C), pH values (3 to 9), and S:F ratios (2:3 to 1:7), demonstrating its versatility and robustness. Furthermore, its effectiveness was consistently high over seven consecutive extraction cycles, highlighting TOPO:LauA as a sustainable, environmentally friendly alternative to hazardous organic solvents, with promising applications for reliable, repeatable use in combating persistent water pollutants such as PFOA.

Keywords: deep eutectic solvents, natural deep eutectic solvents, perfluorooctanoic acid, water remediation

Procedia PDF Downloads 61
12811 Improvement of Ground Truth Data for Eye Location on Infrared Driver Recordings

Authors: Sorin Valcan, Mihail Gaianu

Abstract:

Labeling is a very costly and time consuming process which aims to generate datasets for training neural networks in several functionalities and projects. For driver monitoring system projects, the need for labeled images has a significant impact on the budget and distribution of effort. This paper presents the modifications done to an algorithm used for the generation of ground truth data for 2D eyes location on infrared images with drivers in order to improve the quality of the data and performance of the trained neural networks. The algorithm restrictions become tougher, which makes it more accurate but also less constant. The resulting dataset becomes smaller and shall not be altered by any kind of manual label adjustment before being used in the neural networks training process. These changes resulted in a much better performance of the trained neural networks.

Keywords: labeling automation, infrared camera, driver monitoring, eye detection, convolutional neural networks

Procedia PDF Downloads 117
12810 MarginDistillation: Distillation for Face Recognition Neural Networks with Margin-Based Softmax

Authors: Svitov David, Alyamkin Sergey

Abstract:

The usage of convolutional neural networks (CNNs) in conjunction with the margin-based softmax approach demonstrates the state-of-the-art performance for the face recognition problem. Recently, lightweight neural network models trained with the margin-based softmax have been introduced for the face identification task for edge devices. In this paper, we propose a distillation method for lightweight neural network architectures that outperforms other known methods for the face recognition task on LFW, AgeDB-30 and Megaface datasets. The idea of the proposed method is to use class centers from the teacher network for the student network. Then the student network is trained to get the same angles between the class centers and face embeddings predicted by the teacher network.

Keywords: ArcFace, distillation, face recognition, margin-based softmax

Procedia PDF Downloads 146
12809 Study of Chemical Compounds of Garlic

Authors: Bazaraliyeva Aigerim Bakytzhanovna, Turgumbayeva Aknur Amanbekovna

Abstract:

The phytosubstance from garlic was obtained by extraction with liquid carbon dioxide under critical conditions. Methods of processing raw materials are proposed, and the chemical composition of garlic is studied by gas chromatography and mass spectrometry. The garlic extract's composition was determined using gas chromatography (GC) and gas chromatography-mass spectrophotometry (GC-MS). The phytosubstance had 54 constituents. The extract included the following main compounds: Manool (39.56%), Viridifrolol (7%), Podocarpa-1,8,11,13-tetraen-3-one, 14-isopropyl-1,13-dimethoxy- 5,15 percent, (+)-2-Bornanone (4.29%), Thujone (3.49%), Linolic acid ethyl ester (3.41%), and 12-O-Methylcarn.

Keywords: allium sativum, bioactive compounds of garlic, carbon dioxide extraction of garlic, GS-MS method

Procedia PDF Downloads 80
12808 Intercultural Trainings for Future Global Managers: Evaluating the Effect on the Global Mind-Set

Authors: Nina Dziatzko, Christopher Stehr, Franziska Struve

Abstract:

Intercultural competence as an explicit required skill nearly never appears in job advertisements in international or even global contexts. But especially those who have to deal with different nationalities and cultures in their everyday business need to have several intercultural competencies and further a global mind-set. This way the question arises how potential future global managers can be trained to learn these competencies. In this regard, it might be helpful to see if different types of intercultural trainings have different effects on those skills. This paper outlines lessons learned based on the evaluation of two different intercultural trainings for management students. The main differences between the observed intercultural trainings are the amount of theoretical input in relation to hands-on experiences, the number of trainers as well as the used methods to teach implicit cultural rules. Both groups contain management students with the willingness and perspective to work abroad or to work in international context. The research is carried out with a pre-training-survey and a post-training-survey which consists of questions referring the international context of the students and a self-estimation of 19 identified intercultural and global mind-set skills, such as: cosmopolitanism, empathy, differentiation and adaptability. Whereas there is no clear result which training gets overall a significant higher increase of skills, there is a clear difference between the focus of competencies trained by each of the intercultural trainings. This way this research provides a guideline for both academicals institutions as well as companies for the decision between different types of intercultural trainings, if the to be trained required skills are defined. Therefore the efficiency and the accuracy of fit of the education of future global managers get optimized.

Keywords: global mind-set, intercultural competencies, intercultural training, learning experiences

Procedia PDF Downloads 276
12807 Systems Approach to Design and Production of Picture Books for the Pre-Primary Classes to Attain Educational Goals in Southwest Nigeria

Authors: Azeez Ayodele Ayodele

Abstract:

This paper investigated the problem of picture books design and the quality of the pictures in picture books. The research surveyed nursery and primary schools in four major cities in southwest of Nigeria. The instruments including the descriptive survey questionnaire and a structured interview were developed, validated and administered for collection of relevant data. Descriptive statistics was used in analyzing the data. The result of the study revealed that there were poor quality of pictures in picture books and this is due to scarcity of trained graphic designers who understand systems approach to picture books design and production. There is thus a need for more qualified graphic designers, given in-service professional training as well as a refresher course as criteria for upgrading by the stakeholders.

Keywords: pictures, picture books, pre-primary schools, trained graphic designers

Procedia PDF Downloads 247