Search results for: Arabic Information extraction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 12563

Search results for: Arabic Information extraction

12053 Study of Chemical Compounds of Garlic

Authors: A. B. Bazaralieva, A. A. Turgumbayeva

Abstract:

The phytosubstance from garlic was obtained by extraction with liquid carbon dioxide under critical conditions. Methods of processing raw materials are proposed, and the chemical composition of garlic is studied by gas chromatography and mass spectrometry. The garlic extract's composition was determined using gas chromatography (GC) and gas chromatography-mass spectrophotometry (GC-MS). The phytosubstance had 54 constituents. The extract included the following main compounds: Manool (39.56%), Viridifrolol (7%), Podocarpa-1,8,11,13-tetraen-3-one, 14-isopropyl-1,13-dimethoxy- 5,15 percent, (+)-2-Bornanone (4.29%), Thujone (3.49%), Linolic acid ethyl ester (3.41%), and 12-O-Methylcarn.

Keywords: Allium sativum, bioactive compounds of garlic, carbon dioxide extraction of garlic, GS-MS method

Procedia PDF Downloads 96
12052 Waters Colloidal Phase Extraction and Preconcentration: Method Comparison

Authors: Emmanuelle Maria, Pierre Crançon, Gaëtane Lespes

Abstract:

Colloids are ubiquitous in the environment and are known to play a major role in enhancing the transport of trace elements, thus being an important vector for contaminants dispersion. Colloids study and characterization are necessary to improve our understanding of the fate of pollutants in the environment. However, in stream water and groundwater, colloids are often very poorly concentrated. It is therefore necessary to pre-concentrate colloids in order to get enough material for analysis, while preserving their initial structure. Many techniques are used to extract and/or pre-concentrate the colloidal phase from bulk aqueous phase, but yet there is neither reference method nor estimation of the impact of these different techniques on the colloids structure, as well as the bias introduced by the separation method. In the present work, we have tested and compared several methods of colloidal phase extraction/pre-concentration, and their impact on colloids properties, particularly their size distribution and their elementary composition. Ultrafiltration methods (frontal, tangential and centrifugal) have been considered since they are widely used for the extraction of colloids in natural waters. To compare these methods, a ‘synthetic groundwater’ was used as a reference. The size distribution (obtained by Field-Flow Fractionation (FFF)) and the chemical composition of the colloidal phase (obtained by Inductively Coupled Plasma Mass Spectrometry (ICPMS) and Total Organic Carbon analysis (TOC)) were chosen as comparison factors. In this way, it is possible to estimate the pre-concentration impact on the colloidal phase preservation. It appears that some of these methods preserve in a more efficient manner the colloidal phase composition while others are easier/faster to use. The choice of the extraction/pre-concentration method is therefore a compromise between efficiency (including speed and ease of use) and impact on the structural and chemical composition of the colloidal phase. In perspective, the use of these methods should enhance the consideration of colloidal phase in the transport of pollutants in environmental assessment studies and forensics.

Keywords: chemical composition, colloids, extraction, preconcentration methods, size distribution

Procedia PDF Downloads 204
12051 Text Data Preprocessing Library: Bilingual Approach

Authors: Kabil Boukhari

Abstract:

In the context of information retrieval, the selection of the most relevant words is a very important step. In fact, the text cleaning allows keeping only the most representative words for a better use. In this paper, we propose a library for the purpose text preprocessing within an implemented application to facilitate this task. This study has two purposes. The first, is to present the related work of the various steps involved in text preprocessing, presenting the segmentation, stemming and lemmatization algorithms that could be efficient in the rest of study. The second, is to implement a developed tool for text preprocessing in French and English. This library accepts unstructured text as input and provides the preprocessed text as output, based on a set of rules and on a base of stop words for both languages. The proposed library has been made on different corpora and gave an interesting result.

Keywords: text preprocessing, segmentation, knowledge extraction, normalization, text generation, information retrieval

Procedia PDF Downloads 84
12050 Precious and Rare Metals in Overburden Carbonaceous Rocks: Methods of Extraction

Authors: Tatyana Alexandrova, Alexandr Alexandrov, Nadezhda Nikolaeva

Abstract:

A problem of complex mineral resources development is urgent and priority, it is aimed at realization of the processes of their ecologically safe development, one of its components is revealing the influence of the forms of element compounds in raw materials and in the processing products. In view of depletion of the precious metal reserves at the traditional deposits in the XXI century the large-size open cast deposits, localized in black shale strata begin to play the leading role. Carbonaceous (black) shales carry a heightened metallogenic potential. Black shales with high content of carbon are widely distributed within the scope of Bureinsky massif. According to academician Hanchuk`s data black shales of Sutirskaya series contain generally PGEs native form. The presence of high absorptive towards carbonaceous matter gold and PGEs compounds in crude ore results in decrease of valuable components extraction because of their sorption into dissipated carbonaceous matter.

Keywords: сarbonaceous rocks, bitumens, precious metals, concentration, extraction

Procedia PDF Downloads 236
12049 Hydrometallurgical Processing of a Nigerian Chalcopyrite Ore

Authors: Alafara A. Baba, Kuranga I. Ayinla, Folahan A. Adekola, Rafiu B. Bale

Abstract:

Due to increasing demands and diverse applications of copper oxide as pigment in ceramics, cuprammonium hydroxide solution for rayon, p-type semi-conductor, dry cell batteries production and as safety disposal of hazardous materials, a study on the hydrometallurgical operations involving leaching, solvent extraction and precipitation for the recovery of copper for producing high grade copper oxide from a Nigerian chalcopyrite ore in chloride media has been examined. At a particular set of experimental parameter with respect to acid concentration, reaction temperature and particle size, the leaching investigation showed that the ore dissolution increases with increasing acid concentration, temperature and decreasing particle diameter at a moderate stirring. The kinetics data has been analyzed and was found to follow diffusion control mechanism. At optimal conditions, the extent of ore dissolution reached 94.3%. The recovery of the total copper from the hydrochloric acid-leached chalcopyrite ore was undertaken by solvent extraction and precipitation techniques, prior to the beneficiation of the purified solution as copper oxide. The purification of the leach liquor was firstly done by precipitation of total iron and manganese using Ca(OH)2 and H2O2 as oxidizer at pH 3.5 and 4.25, respectively. An extraction efficiency of 97.3% total copper was obtained by 0.2 mol/L Dithizone in kerosene at 25±2ºC within 40 minutes, from which ≈98% Cu from loaded organic phase was successfully stripped by 0.1 mol/L HCl solution. The beneficiation of the recovered pure copper solution was carried out by crystallization through alkali addition followed by calcination at 600ºC to obtain high grade copper oxide (Tenorite, CuO: 05-0661). Finally, a simple hydrometallurgical scheme for the operational extraction procedure amenable for industrial utilization and economic sustainability was provided.

Keywords: chalcopyrite ore, Nigeria, copper, copper oxide, solvent extraction

Procedia PDF Downloads 379
12048 On Musical Information Geometry with Applications to Sonified Image Analysis

Authors: Shannon Steinmetz, Ellen Gethner

Abstract:

In this paper, a theoretical foundation is developed for patterned segmentation of audio using the geometry of music and statistical manifold. We demonstrate image content clustering using conic space sonification. The algorithm takes a geodesic curve as a model estimator of the three-parameter Gamma distribution. The random variable is parameterized by musical centricity and centric velocity. Model parameters predict audio segmentation in the form of duration and frame count based on the likelihood of musical geometry transition. We provide an example using a database of randomly selected images, resulting in statistically significant clusters of similar image content.

Keywords: sonification, musical information geometry, image, content extraction, automated quantification, audio segmentation, pattern recognition

Procedia PDF Downloads 213
12047 An Exploratory Study into the Suggestive Impact of Alaa Al-Aswany's Political Essays

Authors: Valerii Dudin

Abstract:

With the continuous increase in quantity and importance of the information surrounding our daily lives, it has become crucial to understand what makes information stand out and affect our point of view, regardless of the accuracy of the facts involved. Alaa Al-Aswany’s numerous works have been an inspiration for millions of his readers in Egypt and all across the Arab World. While highly factual, the author’s political essays are both lexically and stylistically rich; they also implement descriptive allusions and proverbs to support the presented opinions. We have undertaken an effort to explore the impact on the individual perception through these political works of the author. In this study, we have overviewed previously made research on similar subjects and through contextual, intertextual, linguistic and corpus analyses we have come to realize the presence of suggestive themes in these works, capable of shaping the reader’s perception regarding a certain topic, specifically targeting the reader’s emotional bias. The findings presented in the study will reveal an overview of such examples of suggestive elements used in the author’s works, as well as various new insights on what can be considered suggestive in the context of modern Arabic printed press.

Keywords: Alaa al-Aswany, cognitive linguistics, political essays, suggestion

Procedia PDF Downloads 145
12046 Extraction of Cellulose Nanocrystals from Soy Pods

Authors: Maycon dos Santos, Marivane Turim Koschevic, Karina Sayuri Ueda, Marcello Lima Bertuci, Farayde Matta Fackhouri, Silvia Maria Martelli

Abstract:

The use of cellulose nanocrystals as reinforcing agents in polymer nanocomposites is promising. In this study, we tested four different methods of mercerization were divided into two stages. The sample was treated in 5% NaOH solution for 30 minutes at 50 ° C in the first stage and 30vol H2O2 for 2 hours at 50 ° C in the second step, which showed better results. For the extraction of the sample obtained nanocrystals positive result was that the solution was treated with H2SO4 60% (w / w) for 1 hour at 50 ° C. The results were positive and showed that it is possible to extract CNC at low temperatures.

Keywords: soy pods, cellulose nanocrystals, temperature, acid concentration

Procedia PDF Downloads 284
12045 Characteristic Sentence Stems in Academic English Texts: Definition, Identification, and Extraction

Authors: Jingjie Li, Wenjie Hu

Abstract:

Phraseological units in academic English texts have been a central focus in recent corpus linguistic research. A wide variety of phraseological units have been explored, including collocations, chunks, lexical bundles, patterns, semantic sequences, etc. This paper describes a special category of clause-level phraseological units, namely, Characteristic Sentence Stems (CSSs), with a view to describing their defining criteria and extraction method. CSSs are contiguous lexico-grammatical sequences which contain a subject-predicate structure and which are frame expressions characteristic of academic writing. The extraction of CSSs consists of six steps: Part-of-speech tagging, n-gram segmentation, structure identification, significance of occurrence calculation, text range calculation, and overlapping sequence reduction. Significance of occurrence calculation is the crux of this study. It includes the computing of both the internal association and the boundary independence of a CSS and tests the occurring significance of the CSS from both inside and outside perspectives. A new normalization algorithm is also introduced into the calculation of LocalMaxs for reducing overlapping sequences. It is argued that many sentence stems are so recurrent in academic texts that the most typical of them have become the habitual ways of making meaning in academic writing. Therefore, studies of CSSs could have potential implications and reference value for academic discourse analysis, English for Academic Purposes (EAP) teaching and writing.

Keywords: characteristic sentence stem, extraction method, phraseological unit, the statistical measure

Procedia PDF Downloads 151
12044 Advantages of Matrix Solid Phase Dispersive (MSPD) Extraction Associated to MIPS versus MAE Liquid Extraction for the Simultaneous Analysis of PAHs, PCBs and Some Hydroxylated PAHs in Sediments

Authors: F. Portet-Koltalo, Y. Tian, I. Berger, C. Boulanger-Lecomte, A. Benamar, N. Machour

Abstract:

Sediments are complex environments which can accumulate a great variety of persistent toxic contaminants such as polychlorobiphenyles (PCBs), polycyclic aromatic hydrocarbons (PAHs) and some of their more toxic degradation metabolites such as hydroxylated PAHs (OH-PAHs). Owing to their composition, fine clayey sediments can be more difficult to extract than soils using conventional solvent extraction processes. So this study aimed to compare the potential of MSPD (matrix solid phase dispersive extraction) to extract PCBs, PAHs and OH-PAHs, in comparison with microwave assisted extraction (MAE). Methodologies: MAE extraction with various solvent mixtures was used to extract PCBs, PAHs and OH-PAHs from sediments in two runs, followed by two GC-MS analyses. MSPD consisted in crushing the dried sediment with dispersive agents, introducing the mixture in cartridges and eluting the target compounds with an appropriate volume of selected solvents. So MSPD combined with cartridges containing MIPs (molecularly imprinted polymers) designed for OH-PAHs was used to extract the three families of target compounds in only one run, followed by parallel analyses in GC-MS for PAHs/PCBs and HPLC-FLD for OH-PAHs. Results: MAE extraction was optimized to extract from clayey sediments, in two runs, PAHs/PCBs in one hand and OH-PAHs in the other hand. Indeed, the best conditions of extractions (mixtures of extracting solvents, temperature) were different if we consider the polarity and the thermodegradability of the different families of target contaminants: PAHs/PCBs were better extracted using an acetone/toluene 50/50 mixture at 130°C whereas OH-PAHs were better extracted using an acetonitrile/toluene 90/10 mixture at 100°C. Moreover, the two consecutive GC-MS analyses contributed to double the total analysis time. A matrix solid phase dispersive (MSPD) extraction procedure was also optimized, with the first objective of increasing the extraction recovery yields of PAHs and PCBs from fine-grained sediment. The crushing time (2-10 min), the nature of the dispersing agents added for purifying and increasing the extraction yields (Florisil, octadecylsilane, 3-chloropropyle, 4-benzylchloride), the nature and the volume of eluting solvents (methylene chloride, hexane, hexane/acetone…) were studied. It appeared that in the best conditions, MSPD was a better extraction method than MAE for PAHs and PCBs, with respectively, mean increases of 8.2% and 71%. This method was also faster, easier and less expensive. But the other advantage of MSPD was that it allowed to introduce easily, just after the first elution process of PAHs/PCBs, a step permitting the selective recovery of OH-PAHs. A cartridge containing MIPs designed for phenols was coupled to the cartridge containing the dispersed sediment, and various eluting solvents, different from those used for PAHs and PCBs, were tested to selectively concentrate and extract OH-PAHs. Thereafter OH-PAHs could be analyzed at the same time than PAHs and PCBs: the OH-PAH extract could be analyzed with HPLC-FLD, whereas the PAHs/PCBs extract was analyzed with GC-MS, adding only few minutes more to the total duration of the analytical process. Conclusion: MSPD associated to MIPs appeared to be an easy, fast and low expensive method, able to extract in one run a complex mixture of toxic apolar and more polar contaminants present in clayey fine-grained sediments, an environmental matrix which is generally difficult to analyze.

Keywords: contaminated fine-grained sediments, matrix solid phase dispersive extraction, microwave assisted extraction, molecularly imprinted polymers, multi-pollutant analysis

Procedia PDF Downloads 339
12043 Determining the Effective Substance of Cottonseed Extract on the Treatment of Leishmaniasis

Authors: Mehrosadat Mirmohammadi, Sara Taghdisi, Ali Padash, Mohammad Hossein Pazandeh

Abstract:

Gossypol, a yellowish anti-nutritional compound found in cotton plants, exists in various plant parts, including seeds, husks, leaves, and stems. Chemically, gossypol is a potent polyphenolic aldehyde with antioxidant and therapeutic properties. However, its free form can be toxic, posing risks to both humans and animals. Initially, we extracted gossypol from cotton seeds using n-hexane as a solvent (yield: 84.0 ± 4.0%). We also obtained cotton seed and cotton boll extracts via Soxhlet extraction (25:75 hydroalcoholic ratio). These extracts, combined with cornstarch, formed four herbal medicinal formulations. Ethical approval allowed us to investigate their effects on Leishmania-caused skin wounds, comparing them to glucantime (local ampoule). Herbal formulas outperformed the control group (ethanol only) in wound treatment (p-value 0.05). The average wound diameter after two months did not significantly differ between plant extract ointments and topical glucantime. Notably, cotton boll extract with 1% extra gossypol crystal showed the best therapeutic effect. We extracted gossypol from cotton seeds using n-hexane via Soxhlet extraction. Saponification, acidification, and recrystallization steps followed. FTIR, UV-Vis, and HPLC analyses confirmed the product’s identity. Herbal medicines from cotton seeds effectively treated chronic wounds compared to the ethanol-only control group. Wound diameter differed significantly between extract ointments and glucantime injections. It seems that due to the presence of large amounts of fat in the oil, the extraction of gossypol from it faces many obstacles. The extraction of this compound with our technique showed that extraction from oil has a higher efficiency, perhaps because of the preparation of oil by cold pressing method, the possibility of losing this compound is much less than when extraction is done with Soxhlet. On the other hand, the gossypol in the oil is mostly bound to the protein, which somehow protects the gossypol until the last stage of the extraction process. Since this compound is very sensitive to light and heat, it was extracted as a derivative with acetic acid. Also, in the treatment section, it was found that the ointment prepared with the extract is more effective and Gossypol is one of the effective ingredients in the treatment. Therefore, gossypol can be extracted from the oil and added to the extract from which gossypol has been extracted to make an effective medicine with a certain dose.

Keywords: cottonseed, glucantime, gossypol, leishmaniasis

Procedia PDF Downloads 47
12042 Ultrasound/Microwave Assisted Extraction Recovery and Identification of Bioactive Compounds (Polyphenols) from Tarbush (Fluorensia cernua)

Authors: Marisol Rodriguez-Duarte, Aide Saenz-Galindo, Carolina Flores-Gallegos, Raul Rodriguez-Herrera, Juan Ascacio-Valdes

Abstract:

The plant known as tarbush (Fluorensia cernua) is a plant originating in northern Mexico, mainly in the states of Coahuila, Durango, San Luis Potosí, Zacatecas and Chihuahua. It is a branched shrub that belongs to the family Asteraceae, has oval leaves of 6 to 11 cm in length and also has small yellow flowers. In Mexico, the tarbush is a very appreciated plant because it has been used as a traditional medicinal agent, for the treatment of gastrointestinal diseases, skin infections and as a healing agent. This plant has been used mainly as an infusion. Due to its traditional use, the content and type of phytochemicals present in the plant are currently unknown and are responsible for its biological properties, so its recovery and identification is very important because the compounds that it contains have relevant applications in the field of food, pharmaceuticals and medicine. The objective of this work was to determine the best extraction condition of phytochemical compounds (mainly polyphenolic compounds) from the leaf using ultrasound/microwave assisted extraction (U/M-AE). To reach the objective, U/M-AE extractions were performed evaluating three mass/volume ratios (1:8, 1:12, 1:16), three ethanol/water solvent concentrations (0%, 30% and 70%), ultrasound extraction time of 20 min and 5 min at 70°C of microwave treatment. All experiments were performed using a fractional factorial experimental design. Once the best extraction condition was defined, the compounds were recovered by liquid column chromatography using Amberlite XAD-16, the polyphenolic fraction was recovered with ethanol and then evaporated. The recovered polyphenolic compounds were quantified by spectrophotometric techniques and identified by HPLC/ESI/MS. The results obtained showed that the best extraction condition of the compounds was using a mass/volume ratio of 1:8 and solvent ethanol/water concentration of 70%. The concentration obtained from polyphenolic compounds using this condition was 22.74 mg/g and finally, 16 compounds of polyphenolic origin were identified. The results obtained in this work allow us to postulate the Mexican plant known as tarbush as a relevant source of bioactive polyphenolic compounds of food, pharmaceutical and medicinal interest.

Keywords: U/M-AE, tarbush, polyphenols, identification

Procedia PDF Downloads 148
12041 Skull Extraction for Quantification of Brain Volume in Magnetic Resonance Imaging of Multiple Sclerosis Patients

Authors: Marcela De Oliveira, Marina P. Da Silva, Fernando C. G. Da Rocha, Jorge M. Santos, Jaime S. Cardoso, Paulo N. Lisboa-Filho

Abstract:

Multiple Sclerosis (MS) is an immune-mediated disease of the central nervous system characterized by neurodegeneration, inflammation, demyelination, and axonal loss. Magnetic resonance imaging (MRI), due to the richness in the information details provided, is the gold standard exam for diagnosis and follow-up of neurodegenerative diseases, such as MS. Brain atrophy, the gradual loss of brain volume, is quite extensive in multiple sclerosis, nearly 0.5-1.35% per year, far off the limits of normal aging. Thus, the brain volume quantification becomes an essential task for future analysis of the occurrence atrophy. The analysis of MRI has become a tedious and complex task for clinicians, who have to manually extract important information. This manual analysis is prone to errors and is time consuming due to various intra- and inter-operator variability. Nowadays, computerized methods for MRI segmentation have been extensively used to assist doctors in quantitative analyzes for disease diagnosis and monitoring. Thus, the purpose of this work was to evaluate the brain volume in MRI of MS patients. We used MRI scans with 30 slices of the five patients diagnosed with multiple sclerosis according to the McDonald criteria. The computational methods for the analysis of images were carried out in two steps: segmentation of the brain and brain volume quantification. The first image processing step was to perform brain extraction by skull stripping from the original image. In the skull stripper for MRI images of the brain, the algorithm registers a grayscale atlas image to the grayscale patient image. The associated brain mask is propagated using the registration transformation. Then this mask is eroded and used for a refined brain extraction based on level-sets (edge of the brain-skull border with dedicated expansion, curvature, and advection terms). In the second step, the brain volume quantification was performed by counting the voxels belonging to the segmentation mask and converted in cc. We observed an average brain volume of 1469.5 cc. We concluded that the automatic method applied in this work can be used for the brain extraction process and brain volume quantification in MRI. The development and use of computer programs can contribute to assist health professionals in the diagnosis and monitoring of patients with neurodegenerative diseases. In future works, we expect to implement more automated methods for the assessment of cerebral atrophy and brain lesions quantification, including machine-learning approaches. Acknowledgements: This work was supported by a grant from Brazilian agency Fundação de Amparo à Pesquisa do Estado de São Paulo (number 2019/16362-5).

Keywords: brain volume, magnetic resonance imaging, multiple sclerosis, skull stripper

Procedia PDF Downloads 132
12040 Comparative Analysis of Feature Extraction and Classification Techniques

Authors: R. L. Ujjwal, Abhishek Jain

Abstract:

In the field of computer vision, most facial variations such as identity, expression, emotions and gender have been extensively studied. Automatic age estimation has been rarely explored. With age progression of a human, the features of the face changes. This paper is providing a new comparable study of different type of algorithm to feature extraction [Hybrid features using HAAR cascade & HOG features] & classification [KNN & SVM] training dataset. By using these algorithms we are trying to find out one of the best classification algorithms. Same thing we have done on the feature selection part, we extract the feature by using HAAR cascade and HOG. This work will be done in context of age group classification model.

Keywords: computer vision, age group, face detection

Procedia PDF Downloads 355
12039 Analysis of Feminist Translation in Subtitling from Arabic into English: A Case Study

Authors: Ghada Ahmed

Abstract:

Feminist translation is one of the strategies adopted in the field of translation studies when a gendered content is being rendered from one language to another, and this strategy has been examined in previous studies on written texts. This research, however, addresses the practice of feminist translation in audiovisual texts that are concerned with the screen, dialogue, image and visual aspects. In this thesis, the objectives are studying feminist translation and its adaptation in subtitling from Arabic into English. It addresses the connections between gender and translation as one domain and feminist translation practices with particular consideration of feminist translation strategies in English subtitles. It examines the visibility of the translator throughout the process, assuming that feminist translation is a product directed by the translator’s feminist position, culture, and ideology as a means of helping unshadow women. It also discusses how subtitling constraints impact feminist translation and how the image that has a narrative value can be integrated into the content of the English subtitles. The reasons for conducting this research project are to study language sexism in English and look into Arabic into English gendered content, taking into consideration the Arabic cultural concepts that may lose their connotations when they are translated into English. This research is also analysing the image in an audiovisual text and its contribution to the written dialogue in subtitling. Thus, this research attempts to answer the following questions: To what extent is there a form of affinity between a gendered content and translation? Is feminist translation an act of merely working on a feminist text or feminising the language of any text, by incorporating the translator’s ideology? How can feminist translation practices be applied in an audiovisual text? How likely is it to adapt feminist translation looking into visual components as well as subtitling constraints? Moreover, the paper searches into the fields of gender and translation; feminist translation, language sexism, media studies, and the gap in the literature related to feminist translation practice in visual texts. For my case study, the "Speed Sisters" film has been chosen so as to analyze its English subtitles for my research. The film is a documentary that was produced in 2015 and directed by Amber Fares. It is about five Palestinian women who try to break the stereotypes about women, and have taken their passion about car-racing forward to be the first all-women car-racing driving team in the Middle East. It tackles the issue of gender in both content and language and this is reflected in the translation. As the research topic is semiotic-channelled, the choice for the theoretical approaches varies and combines between translation studies, audiovisual translation, gender studies, and media studies. Each of which will contribute to understanding a specific field of the research and the results will eventually be integrated to achieve the intended objectives in a way that demonstrates rendering a gendered content in one of the audiovisual translation modes from a language into another.

Keywords: audiovisual translation, feminist translation, films gendered content, subtitling conventions and constraints

Procedia PDF Downloads 283
12038 Information Literacy: Concept and Importance

Authors: Gaurav Kumar

Abstract:

An information literate person is one who uses information effectively in all its forms. When presented with questions or problems, an information literate person would know what information to look for, how to search efficiently and be able to access relevant sources. In addition, an information literate person would have the ability to evaluate and select appropriate information sources and to use the information effectively and ethically to answer questions or solve problems. Information literacy has become an important element in higher education. The information literacy movement has internationally recognized standards and learning outcomes. The step-by-step process of achieving information literacy is particularly crucial in an era where knowledge could be disseminated through a variety of media. What is the relationship between information literacy as we define it in higher education and information literacy among non-academic populations? What forces will change how we think about the definition of information literacy in the future and how we will apply the definition in all environments?

Keywords: information literacy, human beings, visual media and computer network etc, information literacy

Procedia PDF Downloads 323
12037 A Summary-Based Text Classification Model for Graph Attention Networks

Authors: Shuo Liu

Abstract:

In Chinese text classification tasks, redundant words and phrases can interfere with the formation of extracted and analyzed text information, leading to a decrease in the accuracy of the classification model. To reduce irrelevant elements, extract and utilize text content information more efficiently and improve the accuracy of text classification models. In this paper, the text in the corpus is first extracted using the TextRank algorithm for abstraction, the words in the abstract are used as nodes to construct a text graph, and then the graph attention network (GAT) is used to complete the task of classifying the text. Testing on a Chinese dataset from the network, the classification accuracy was improved over the direct method of generating graph structures using text.

Keywords: Chinese natural language processing, text classification, abstract extraction, graph attention network

Procedia PDF Downloads 85
12036 Local Directional Encoded Derivative Binary Pattern Based Coral Image Classification Using Weighted Distance Gray Wolf Optimization Algorithm

Authors: Annalakshmi G., Sakthivel Murugan S.

Abstract:

This paper presents a local directional encoded derivative binary pattern (LDEDBP) feature extraction method that can be applied for the classification of submarine coral reef images. The classification of coral reef images using texture features is difficult due to the dissimilarities in class samples. In coral reef image classification, texture features are extracted using the proposed method called local directional encoded derivative binary pattern (LDEDBP). The proposed approach extracts the complete structural arrangement of the local region using local binary batten (LBP) and also extracts the edge information using local directional pattern (LDP) from the edge response available in a particular region, thereby achieving extra discriminative feature value. Typically the LDP extracts the edge details in all eight directions. The process of integrating edge responses along with the local binary pattern achieves a more robust texture descriptor than the other descriptors used in texture feature extraction methods. Finally, the proposed technique is applied to an extreme learning machine (ELM) method with a meta-heuristic algorithm known as weighted distance grey wolf optimizer (GWO) to optimize the input weight and biases of single-hidden-layer feed-forward neural networks (SLFN). In the empirical results, ELM-WDGWO demonstrated their better performance in terms of accuracy on all coral datasets, namely RSMAS, EILAT, EILAT2, and MLC, compared with other state-of-the-art algorithms. The proposed method achieves the highest overall classification accuracy of 94% compared to the other state of art methods.

Keywords: feature extraction, local directional pattern, ELM classifier, GWO optimization

Procedia PDF Downloads 149
12035 An Analysis of the Five Most Used Numerals and a Proposal for the Adoption of a Universally Acceptable Numeral (UAN)

Authors: Mufutau Ayinla Abdul-Yakeen

Abstract:

An analysis of the five most used numerals and a proposal for the adoption of a Universally Acceptable Numerals (UAN), came up as a result of the researchers inquisitiveses of the need for a set of numerals that is universally accepted. The researcher sought for the meaning of the first letter, “Nun”, “ن”, of the first verse of Suratul-Kalam (Chapter of the Pen), the Sixty-Eighth Chapter of the Holy Qur'an. It was observed that there was no universally accepted, economical, explainable, linkable and consistent set of numerals used by all scientists up till the moment of making this enquiry. As a theoretical paper, explanatory method is used to review five of the most used numerals (Tally Marks, Roman Figure, Hindu-Arabic, Arabic, and Chinese) and the urgent need for a universally accepted, economical, explainable, linkable and consistent set of numerals arises. The study discovers: ., I, \, _, L, U, =, C, O, 9, and 1.; to be used as numeral 0, 1, 2, 3, 4, 5, 6, 7, 8, 9 and 10 respectively; as a set of universally acceptable, economical, explainable, linkable, sustainable, convertible and consistent set of numerals that originates from Islam. They can be called Islameconumerals or UAN. With UAN, everything dropped, written, drawn and/or scribbled has meaning(s) as postulated by the first verse of Qur'an 68 and everyone can easily document all figures within the shortest period. It is suggested that there should be a discipline called Numeralnomics (Study of optimum utilization of Numerals) and everybody should start using the UAN, now, in order in know their strengths and weaknesses so as to suggest a better and acceptable set of numerals for the interested readers. Similarly study can be conducted for the alphabets.

Keywords: acceptable, economical, explainable, Islameconumerals, numeralnomics

Procedia PDF Downloads 308
12034 Feature Evaluation Based on Random Subspace and Multiple-K Ensemble

Authors: Jaehong Yu, Seoung Bum Kim

Abstract:

Clustering analysis can facilitate the extraction of intrinsic patterns in a dataset and reveal its natural groupings without requiring class information. For effective clustering analysis in high dimensional datasets, unsupervised dimensionality reduction is an important task. Unsupervised dimensionality reduction can generally be achieved by feature extraction or feature selection. In many situations, feature selection methods are more appropriate than feature extraction methods because of their clear interpretation with respect to the original features. The unsupervised feature selection can be categorized as feature subset selection and feature ranking method, and we focused on unsupervised feature ranking methods which evaluate the features based on their importance scores. Recently, several unsupervised feature ranking methods were developed based on ensemble approaches to achieve their higher accuracy and stability. However, most of the ensemble-based feature ranking methods require the true number of clusters. Furthermore, these algorithms evaluate the feature importance depending on the ensemble clustering solution, and they produce undesirable evaluation results if the clustering solutions are inaccurate. To address these limitations, we proposed an ensemble-based feature ranking method with random subspace and multiple-k ensemble (FRRM). The proposed FRRM algorithm evaluates the importance of each feature with the random subspace ensemble, and all evaluation results are combined with the ensemble importance scores. Moreover, FRRM does not require the determination of the true number of clusters in advance through the use of the multiple-k ensemble idea. Experiments on various benchmark datasets were conducted to examine the properties of the proposed FRRM algorithm and to compare its performance with that of existing feature ranking methods. The experimental results demonstrated that the proposed FRRM outperformed the competitors.

Keywords: clustering analysis, multiple-k ensemble, random subspace-based feature evaluation, unsupervised feature ranking

Procedia PDF Downloads 324
12033 Density Determination by Dilution for Extra Heavy Oil Residues Obtained Using Molecular Distillation and Supercritical Fluid Extraction as Upgrading and Refining Process

Authors: Oscar Corredor, Alexander Guzman, Adan Leon

Abstract:

Density is a bulk physical property that indicates the quality of a petroleum fraction. It is also a useful property to estimate various physicochemical properties of fraction and petroleum fluids; however, the determination of density of extra heavy residual (EHR) fractions by standard methodologies, (ASTM D70) shows limitations for samples with higher densities than 1.0879 g/cm3. For this reason, a dilution methodology was developed in order to determinate density for those particular fractions, 87 (EHR) fractions were obtained as products of the fractionation of Colombian typical Vacuum Distillation Residual Fractions using molecular distillation (MD) and extraction with Solvent N-hexane in Supercritical Conditions (SFEF) pilot plants. The proposed methodology showed reliable results that can be demonstrated with the standard deviation of repeatability and reproducibility values of 0.0031 and 0.0061 g/ml respectively. In the same way, it was possible to determine densities in fractions EHR up to 1.1647g/cm3 and °API values obtained were ten times less than the water reference value.

Keywords: API, density, vacuum residual, molecular distillation, supercritical fluid extraction

Procedia PDF Downloads 256
12032 Breakthrough Highly-Effective Extraction of Perfluoroctanoic Acid Using Natural Deep Eutectic Solvents

Authors: Sana Eid, Ahmad S. Darwish, Tarek Lemaoui, Maguy Abi Jaoude, Fawzi Banat, Shadi W. Hasan, Inas M. AlNashef

Abstract:

Addressing the growing challenge of per- and polyfluoroalkyl substances (PFAS) pollution in water bodies, this study introduces natural deep eutectic solvents (NADESs) as a pioneering solution for the efficient extraction of perfluorooctanoic acid (PFOA), one of the most persistent and concerning PFAS pollutants. Among the tested NADESs, trioctylphosphine oxide: lauric acid (TOPO:LauA) in a 1:1 molar ratio was distinguished as the most effective, achieving an extraction efficiency of approximately 99.52% at a solvent-to-feed (S:F) ratio of 1:2, room temperature, and neutral pH. This efficiency is achieved within a notably short mixing time of only one min, which is significantly less than the time required by conventional methods, underscoring the potential of TOPO:LauA for rapid and effective PFAS remediation. TOPO:LauA maintained consistent performance across various operational parameters, including a range of initial PFOA concentrations (0.1 ppm to 1000 ppm), temperatures (15 °C to 100 °C), pH values (3 to 9), and S:F ratios (2:3 to 1:7), demonstrating its versatility and robustness. Furthermore, its effectiveness was consistently high over seven consecutive extraction cycles, highlighting TOPO:LauA as a sustainable, environmentally friendly alternative to hazardous organic solvents, with promising applications for reliable, repeatable use in combating persistent water pollutants such as PFOA.

Keywords: deep eutectic solvents, natural deep eutectic solvents, perfluorooctanoic acid, water remediation

Procedia PDF Downloads 44
12031 The Effect of the Pronunciation of Emphatic Sounds on Perceived Masculinity/Femininity

Authors: M. Sayyour, M. Abdulkareem, O. Osman, S. Salmeh

Abstract:

Emphatic sounds in Arabic are /tˤ/, /sˤ/, /dˤ/, and /ðˤ/. They involve a secondary articulation in the pharynx area as opposed to their counterparts: /t/,/s/,/d/and /ð/. Although they are present in most Arabic dialects, some dialects have lost this class as a historical development, such as Maltese Arabic. It has been found that there is a difference in the pronunciation of these emphatic sounds between the two genders, arguing that males tend to produce more evident emphasis than females. This study builds on these studies by trying to investigate whether listeners perceive fully emphatic sounds as more masculine and less emphatic sounds as more feminine. Furthermore, the study aims to find out which is more important in this perception process: the emphatic consonant itself or the vowel following it. To test this, natural and manipulated tokens of two male and two female speakers were used. The natural tokens include words that have emphatic consonant and emphatic vowel and tokens that have plain consonant and plain vowel. The manipulated tokens include words that have emphatic consonant but central vowel and plain consonant followed by the same central vowel. These manipulated tokens allow us to see whether the consonant will still affect the perception even if the vowel is controlled. Another group of words that contained no emphatic sounds was used as a control group. The total number of tokens (natural, manipulated, and control) are 160 tokens. After that, 60 university students (30 males and 30 females) listened to these tokens and responded by choosing a specific character that they think is likely to produce each token. The characters’ descriptions are carefully written with two degrees of femininity and two degrees of masculinity. The preliminary results for the femininity level showed that the highest degree of femininity was for tokens that contain a plain consonant and a plain vowel. The lowest level of femininity was given for tokens that have fully emphatic consonant and vowel. For the manipulated tokens that contained plain consonant and central vowel, the femininity degree was high which indicates that the consonant is more important than the vowel, while for the manipulated tokens that contain emphatic consonant and a central vowel, the femininity level was higher than that for the tokens that have emphatic consonant and emphatic vowel, which indicates that the vowel is more important for the perception of emphatic consonants. These results are interpreted in light of feminist linguistic theories, linguistic expectations, performed gender and linguistic change theories.

Keywords: Emphatic sounds, gender studies, perception, sociophonetics

Procedia PDF Downloads 366
12030 A Quality Index Optimization Method for Non-Invasive Fetal ECG Extraction

Authors: Lucia Billeci, Gennaro Tartarisco, Maurizio Varanini

Abstract:

Fetal cardiac monitoring by fetal electrocardiogram (fECG) can provide significant clinical information about the healthy condition of the fetus. Despite this potentiality till now the use of fECG in clinical practice has been quite limited due to the difficulties in its measuring. The recovery of fECG from the signals acquired non-invasively by using electrodes placed on the maternal abdomen is a challenging task because abdominal signals are a mixture of several components and the fetal one is very weak. This paper presents an approach for fECG extraction from abdominal maternal recordings, which exploits the characteristics of pseudo-periodicity of fetal ECG. It consists of devising a quality index (fQI) for fECG and of finding the linear combinations of preprocessed abdominal signals, which maximize these fQI (quality index optimization - QIO). It aims at improving the performances of the most commonly adopted methods for fECG extraction, usually based on maternal ECG (mECG) estimating and canceling. The procedure for the fECG extraction and fetal QRS (fQRS) detection is completely unsupervised and based on the following steps: signal pre-processing; maternal ECG (mECG) extraction and maternal QRS detection; mECG component approximation and canceling by weighted principal component analysis; fECG extraction by fQI maximization and fetal QRS detection. The proposed method was compared with our previously developed procedure, which obtained the highest at the Physionet/Computing in Cardiology Challenge 2013. That procedure was based on removing the mECG from abdominal signals estimated by a principal component analysis (PCA) and applying the Independent component Analysis (ICA) on the residual signals. Both methods were developed and tuned using 69, 1 min long, abdominal measurements with fetal QRS annotation of the dataset A provided by PhysioNet/Computing in Cardiology Challenge 2013. The QIO-based and the ICA-based methods were compared in analyzing two databases of abdominal maternal ECG available on the Physionet site. The first is the Abdominal and Direct Fetal Electrocardiogram Database (ADdb) which contains the fetal QRS annotations thus allowing a quantitative performance comparison, the second is the Non-Invasive Fetal Electrocardiogram Database (NIdb), which does not contain the fetal QRS annotations so that the comparison between the two methods can be only qualitative. In particular, the comparison on NIdb was performed defining an index of quality for the fetal RR series. On the annotated database ADdb the QIO method, provided the performance indexes Sens=0.9988, PPA=0.9991, F1=0.9989 overcoming the ICA-based one, which provided Sens=0.9966, PPA=0.9972, F1=0.9969. The comparison on NIdb was performed defining an index of quality for the fetal RR series. The index of quality resulted higher for the QIO-based method compared to the ICA-based one in 35 records out 55 cases of the NIdb. The QIO-based method gave very high performances with both the databases. The results of this study foresees the application of the algorithm in a fully unsupervised way for the implementation in wearable devices for self-monitoring of fetal health.

Keywords: fetal electrocardiography, fetal QRS detection, independent component analysis (ICA), optimization, wearable

Procedia PDF Downloads 268
12029 Study of Chemical Compounds of Garlic

Authors: Bazaraliyeva Aigerim Bakytzhanovna, Turgumbayeva Aknur Amanbekovna

Abstract:

The phytosubstance from garlic was obtained by extraction with liquid carbon dioxide under critical conditions. Methods of processing raw materials are proposed, and the chemical composition of garlic is studied by gas chromatography and mass spectrometry. The garlic extract's composition was determined using gas chromatography (GC) and gas chromatography-mass spectrophotometry (GC-MS). The phytosubstance had 54 constituents. The extract included the following main compounds: Manool (39.56%), Viridifrolol (7%), Podocarpa-1,8,11,13-tetraen-3-one, 14-isopropyl-1,13-dimethoxy- 5,15 percent, (+)-2-Bornanone (4.29%), Thujone (3.49%), Linolic acid ethyl ester (3.41%), and 12-O-Methylcarn.

Keywords: allium sativum, bioactive compounds of garlic, carbon dioxide extraction of garlic, GS-MS method

Procedia PDF Downloads 72
12028 Comparative Pre-treatment Analysis of RNA-Extraction Methods and Efficient Detection of SARS-COV-2 and PMMoV in Influents and 1ˢᵗ Sedimentation from a Wastewater Treatment Plan

Authors: Jesmin Akter, Chang Hyuk Ahn, Ilho Kim, Fumitake Nishimura, Jaiyeop Lee

Abstract:

This study aimed to compare two pre-treatment and two RNA extraction methods, namely PEG, and Nano bubble, Viral RNA Soil, and Mini Kit, in terms of their efficiency in detecting SARS-CoV-2 and PMMoV in influent and 1st sedimentation samples from a wastewater treatment plant. The extracted RNA samples were quantified and evaluated for purity, yield, and integrity. The results indicated that the nanobubble PEG method provided the highest yield of RNA, while the QIAamp Viral RNA Mini Kit produced the purest RNA samples. In terms of sensitivity and specificity, all these methods were able to detect SARS-CoV-2 and PMMoV in both influent and 1st sedimentation samples. However, the nanobubble PEG method showed slightly higher sensitivity compared to the other methods. These findings suggest that the choice of RNA extraction method should depend on the downstream application and the quality of the RNA required. The study also highlights the potential of wastewater-based epidemiology as an effective and non-invasive method for monitoring the spread of infectious diseases in a community.

Keywords: influent, PMMoV, SARS-CoV-2, wastewater based epidemiology

Procedia PDF Downloads 85
12027 Status of Bio-Graphene Extraction from Biomass: A Review

Authors: Simon Peter Wafula, Ziporah Nakabazzi Kitooke

Abstract:

Graphene is a carbon allotrope made of a two-dimensional shape. This material has got a number of materials researchers’ interest due to its properties that are special compared to ordinary material. Graphene is thought to enhance a number of material properties in the manufacturing, energy, and construction industries. Many studies consider graphene to be a wonder material, just like plastic in the 21st century. This shows how much should be invested in graphene research. This review highlights the status of graphene extracted from various biomass sources together with their appropriate extraction techniques, including the pretreatment methods for a better product. The functional groups and structure of graphene extracted using several common methods of synthesis are in this paper as well. The review explores methods like chemical vapor deposition (CVD), hydrothermal, chemical exfoliation method, liquid exfoliation, and Hummers. Comparative analysis of the various extraction techniques gives an insight into each of their advantages, challenges, and potential scalability. The review also highlights the pretreatment process for biomass before carbonation for better quality of bio-graphene. The various graphene modes, as well as their applications, are in this study. Recommendations for future research for improving the efficiency and sustainability of bio-graphene are highlighted.

Keywords: exfoliation, nanomaterials, biochar, large-scale, two-dimension

Procedia PDF Downloads 35
12026 Comprehensive Feature Extraction for Optimized Condition Assessment of Fuel Pumps

Authors: Ugochukwu Ejike Akpudo, Jank-Wook Hur

Abstract:

The increasing demand for improved productivity, maintainability, and reliability has prompted rapidly increasing research studies on the emerging condition-based maintenance concept- Prognostics and health management (PHM). Varieties of fuel pumps serve critical functions in several hydraulic systems; hence, their failure can have daunting effects on productivity, safety, etc. The need for condition monitoring and assessment of these pumps cannot be overemphasized, and this has led to the uproar in research studies on standard feature extraction techniques for optimized condition assessment of fuel pumps. By extracting time-based, frequency-based and the more robust time-frequency based features from these vibrational signals, a more comprehensive feature assessment (and selection) can be achieved for a more accurate and reliable condition assessment of these pumps. With the aid of emerging deep classification and regression algorithms like the locally linear embedding (LLE), we propose a method for comprehensive condition assessment of electromagnetic fuel pumps (EMFPs). Results show that the LLE as a comprehensive feature extraction technique yields better feature fusion/dimensionality reduction results for condition assessment of EMFPs against the use of single features. Also, unlike other feature fusion techniques, its capabilities as a fault classification technique were explored, and the results show an acceptable accuracy level using standard performance metrics for evaluation.

Keywords: electromagnetic fuel pumps, comprehensive feature extraction, condition assessment, locally linear embedding, feature fusion

Procedia PDF Downloads 106
12025 An Analysis of Language Borrowing among Algerian University Students Using Online Facebook Conversations

Authors: Messaouda Annab

Abstract:

The rapid development of technology has led to an important context in which different languages and structures are used in the same conversations. This paper investigates the practice of language borrowing within social media platform, namely, Facebook among Algerian Vernacular Arabic (AVA) students. In other words, this study will explore how Algerian students have incorporated lexical English borrowing in their online conversations. This paper will examine the relationships between language, culture and identity among a multilingual group. The main objective is to determine the cultural and linguistic functions that borrowing fulfills in social media and to explain the possible factors underlying English borrowing. The nature of the study entails the use of an online research method that includes ten online Facebook conversations in the form of private messages collected from Bachelor and Masters Algerian students recruited from the English department at the University of Oum El-Bouaghi. The analysis of data revealed that social media platform provided the users with opportunities to shift from one language to another. This practice was noticed in students’ online conversations. English borrowing was the most relevant language performance in accordance with Arabic which is the mother tongue of the chosen sample. The analysis has assumed that participants are skilled in more than one language.

Keywords: borrowing, language performance, linguistic background, social media

Procedia PDF Downloads 145
12024 Feature Extraction and Classification Based on the Bayes Test for Minimum Error

Authors: Nasar Aldian Ambark Shashoa

Abstract:

Classification with a dimension reduction based on Bayesian approach is proposed in this paper . The first step is to generate a sample (parameter) of fault-free mode class and faulty mode class. The second, in order to obtain good classification performance, a selection of important features is done with the discrete karhunen-loeve expansion. Next, the Bayes test for minimum error is used to classify the classes. Finally, the results for simulated data demonstrate the capabilities of the proposed procedure.

Keywords: analytical redundancy, fault detection, feature extraction, Bayesian approach

Procedia PDF Downloads 514