Search results for: text mining
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2218

Search results for: text mining

688 Post-Structural Study of Gender in Shakespearean Othello from Butlerian Perspective

Authors: Muhammad Shakeel Rehman Hissam

Abstract:

This study aims at analyzing gender in Othello by applying Judith Butler’s Post-Structural theory of gender and gender performance. The analysis of the play provides us context by which we can examine what kinds of effects the drama have on understanding of the researchers regarding gender identity. The study sets out to examine that, is there any evidence or ground in Shakespearean selected work which leads to challenge the patriarchal taken for granted prescribed roles of gender? This would be the focal point in study of Othello that actions and performances of characters determine their gender identity rather than their sexuality. It argues that gender of Shakespearean characters has no constant, fixed and structural impression. On the contrary, they undergo consistent variations in their behavior and performance which impart fluidity and volatility to them. The focal point of the present study is Butler’s prominent work; Gender Trouble: Feminism and subversion of Identity and her post structural theory of Gender performativity as the theoretical underpinning of the text. It analyzes the selected play in Post-Structural gender perspective. The gender-centric plot of the play is riddled with fluidity of gender. The most fascinating aspect of the play is the transformations of genders on the basis of performances by different characters and through these transformations; gender identity is revealed and determined. The study reconstructs the accepted gender norms by challenging the traditional concept of gender that is based on sexual differences of characters.

Keywords: post structural, gender, performativity, socio-cultural gender norms, binaries, Othello, Butler, identity

Procedia PDF Downloads 349
687 Wolof Voice Response Recognition System: A Deep Learning Model for Wolof Audio Classification

Authors: Krishna Mohan Bathula, Fatou Bintou Loucoubar, FNU Kaleemunnisa, Christelle Scharff, Mark Anthony De Castro

Abstract:

Voice recognition algorithms such as automatic speech recognition and text-to-speech systems with African languages can play an important role in bridging the digital divide of Artificial Intelligence in Africa, contributing to the establishment of a fully inclusive information society. This paper proposes a Deep Learning model that can classify the user responses as inputs for an interactive voice response system. A dataset with Wolof language words ‘yes’ and ‘no’ is collected as audio recordings. A two stage Data Augmentation approach is adopted for enhancing the dataset size required by the deep neural network. Data preprocessing and feature engineering with Mel-Frequency Cepstral Coefficients are implemented. Convolutional Neural Networks (CNNs) have proven to be very powerful in image classification and are promising for audio processing when sounds are transformed into spectra. For performing voice response classification, the recordings are transformed into sound frequency feature spectra and then applied image classification methodology using a deep CNN model. The inference model of this trained and reusable Wolof voice response recognition system can be integrated with many applications associated with both web and mobile platforms.

Keywords: automatic speech recognition, interactive voice response, voice response recognition, wolof word classification

Procedia PDF Downloads 96
686 Technology in the Calculation of People Health Level: Design of a Computational Tool

Authors: Sara Herrero Jaén, José María Santamaría García, María Lourdes Jiménez Rodríguez, Jorge Luis Gómez González, Adriana Cercas Duque, Alexandra González Aguna

Abstract:

Background: Health concept has evolved throughout history. The health level is determined by the own individual perception. It is a dynamic process over time so that you can see variations from one moment to the next. In this way, knowing the health of the patients you care for, will facilitate decision making in the treatment of care. Objective: To design a technological tool that calculates the people health level in a sequential way over time. Material and Methods: Deductive methodology through text analysis, extraction and logical knowledge formalization and education with expert group. Studying time: September 2015- actually. Results: A computational tool for the use of health personnel has been designed. It has 11 variables. Each variable can be given a value from 1 to 5, with 1 being the minimum value and 5 being the maximum value. By adding the result of the 11 variables we obtain a magnitude in a certain time, the health level of the person. The health calculator allows to represent people health level at a time, establishing temporal cuts being useful to determine the evolution of the individual over time. Conclusion: The Information and Communication Technologies (ICT) allow training and help in various disciplinary areas. It is important to highlight their relevance in the field of health. Based on the health formalization, care acts can be directed towards some of the propositional elements of the concept above. The care acts will modify the people health level. The health calculator allows the prioritization and prediction of different strategies of health care in hospital units.

Keywords: calculator, care, eHealth, health

Procedia PDF Downloads 244
685 Advancement of Computer Science Research in Nigeria: A Bibliometric Analysis of the Past Three Decades

Authors: Temidayo O. Omotehinwa, David O. Oyewola, Friday J. Agbo

Abstract:

This study aims to gather a proper perspective of the development landscape of Computer Science research in Nigeria. Therefore, a bibliometric analysis of 4,333 bibliographic records of Computer Science research in Nigeria in the last 31 years (1991-2021) was carried out. The bibliographic data were extracted from the Scopus database and analyzed using VOSviewer and the bibliometrix R package through the biblioshiny web interface. The findings of this study revealed that Computer Science research in Nigeria has a growth rate of 24.19%. The most developed and well-studied research areas in the Computer Science field in Nigeria are machine learning, data mining, and deep learning. The social structure analysis result revealed that there is a need for improved international collaborations. Sparsely established collaborations are largely influenced by geographic proximity. The funding analysis result showed that Computer Science research in Nigeria is under-funded. The findings of this study will be useful for researchers conducting Computer Science related research. Experts can gain insights into how to develop a strategic framework that will advance the field in a more impactful manner. Government agencies and policymakers can also utilize the outcome of this research to develop strategies for improved funding for Computer Science research.

Keywords: bibliometric analysis, biblioshiny, computer science, Nigeria, science mapping

Procedia PDF Downloads 91
684 DCDNet: Lightweight Document Corner Detection Network Based on Attention Mechanism

Authors: Kun Xu, Yuan Xu, Jia Qiao

Abstract:

The document detection plays an important role in optical character recognition and text analysis. Because the traditional detection methods have weak generalization ability, and deep neural network has complex structure and large number of parameters, which cannot be well applied in mobile devices, this paper proposes a lightweight Document Corner Detection Network (DCDNet). DCDNet is a two-stage architecture. The first stage with Encoder-Decoder structure adopts depthwise separable convolution to greatly reduce the network parameters. After introducing the Feature Attention Union (FAU) module, the second stage enhances the feature information of spatial and channel dim and adaptively adjusts the size of receptive field to enhance the feature expression ability of the model. Aiming at solving the problem of the large difference in the number of pixel distribution between corner and non-corner, Weighted Binary Cross Entropy Loss (WBCE Loss) is proposed to define corner detection problem as a classification problem to make the training process more efficient. In order to make up for the lack of Dataset of document corner detection, a Dataset containing 6620 images named Document Corner Detection Dataset (DCDD) is made. Experimental results show that the proposed method can obtain fast, stable and accurate detection results on DCDD.

Keywords: document detection, corner detection, attention mechanism, lightweight

Procedia PDF Downloads 335
683 Walls, Barriers, and Fences to Informal Political Economy of Land Resource Accesses: A Case of Banyabunagana Along with Uganda–Congo Border, South Western Uganda, Kisoro District

Authors: Niringiye Fred

Abstract:

Banyabunagana has always had access to land resources for grazing animals, sand mining, and farmland across the border in the Democratic Republic of Congo during the pre-colonial and colonial times, usually on an informal arrangement facilitated by kinship ties and rent transactions for these resources. However, in recent periods, the government of the Democratic Republic of the Congo (DRC) has been pursuing a policy of constructing barriers such as walls and fences so that Banyabunagana communities do not access the land on the DRC side of the border. This is happening in the background of increased and intensified demand for land use on the side of the Ugandan community. This paper will attempt to discuss the reasons behind the construction of walls, fences, and other barriers which deny access to land for Banyabunagana communities in Bunagana Parish, Muramba Sub-county- Kisoro district, Uganda. The research will attempt to answer the following main questions, among others, whether there are the factors that explain the construction of walls and fences which could limit or deny access to the informal use of land and other resources and whether policy options to ensure continued access to land and other resources for local communities.

Keywords: border, walls, fences, land resource access

Procedia PDF Downloads 96
682 A Systematic Review Examining the Experimental methodology behind in vivo testing of hiatus hernia and Diaphragmatic Hernia Mesh

Authors: Whitehead-Clarke T., Beynon V., Banks J., Karanjia R., Mudera V., Windsor A., Kureshi A.

Abstract:

Introduction: Mesh implants are regularly used to help repair both hiatus hernias (HH) and diaphragmatic hernias (DH). In vivo studies are used to test not only mesh safety but increasingly comparative efficacy. Our work examines the field of in vivo mesh testing for HH and DH models to establish current practices and standards. Method: This systematic review was registered with PROSPERO. Medline and Embase databases were searched for relevant in vivo studies. 44 articles were identified and underwent abstract review, where 22 were excluded. 4 further studies were excluded after full text review – leaving 18 to undergo data extraction. Results: Of 18 studies identified, 9 used an in vivo HH model and 9 a DH model. 5 studies undertook mechanical testing on tissue samples – all uniaxial in nature. Testing strip widths ranged from 1-20mm (median 3mm). Testing speeds varied from 1.5-60mm/minute. Upon histology, the most commonly assessed structural and cellular factors were neovascularization and macrophages, respectively (n=9 each). Structural analysis was mostly qualitative, where cellular analysis was equally likely to be quantitative. 11 studies assessed adhesion formation, of which 8 used one of four scoring systems. 8 studies measured mesh shrinkage. Discussion: In vivo studies assessing mesh for HH and DH repair are uncommon. Within this relatively young field, we encourage surgical and materials testing institutions to discuss its standardisation.

Keywords: hiatus, diaphragmatic, hernia, mesh, materials testing, in vivo

Procedia PDF Downloads 200
681 The 5S Responses of Obese Teenagers in Verbal Bullying

Authors: Alpha Bolinao, Francine Rose De Castro, Jessie Kate Lumba, Raztine Mae Paeste, Hannah Grace Tosio

Abstract:

The present study aimed to know the role of verbal bullying in the lives of obese teenagers exposed to it. The study employed a qualitative design specifically the phenomenological approach that focuses on the obese teenagers’ verbal bullying experiences. The study also used the social constructivism approach wherein it described the obese teenagers’ verbal bullying experiences as they interact with the social world. Through purposive and referral sampling technique, the researchers were able to choose twelve (12) respondents from different schools around the City of Manila, enrolled in the School Year 2015-2016, ages 16-21 years old, has experienced verbal bullying for the last ten (10) years and with the Body Mass Index (BMI) of equal to or greater than 30. Upon the consent of the respondents, ethical considerations were ensured. In-depth one (1) hour interviews were guided by the researchers’ aide memoir. The recorded interviews were transcribed into a field text and the responses were thoroughly analyzed through Thematic Analysis and Kelly’s Repertory Grid. It was found that the role of verbal bullying in the lives of obese teenagers exposed to it is a process and is best described through a syringe, or the 5S Responses of Obese Teenagers in Bullying, with five conceptual themes which also signify the experiences and the process that obese teenagers have gone through after experiencing verbal bullying. The themes conceptualized were: Suffering, self-doubt, suppression, self-acceptance and sanguineness. This paper may serve as a basis for a counseling program to help the obese teenagers cope with their bullying experiences.

Keywords: obesity, obese teenagers, bullying, experiences

Procedia PDF Downloads 344
680 Multi-Criteria Inventory Classification Process Based on Logical Analysis of Data

Authors: Diana López-Soto, Soumaya Yacout, Francisco Ángel-Bello

Abstract:

Although inventories are considered as stocks of money sitting on shelve, they are needed in order to secure a constant and continuous production. Therefore, companies need to have control over the amount of inventory in order to find the balance between excessive and shortage of inventory. The classification of items according to certain criteria such as the price, the usage rate and the lead time before arrival allows any company to concentrate its investment in inventory according to certain ranking or priority of items. This makes the decision making process for inventory management easier and more justifiable. The purpose of this paper is to present a new approach for the classification of new items based on the already existing criteria. This approach is called the Logical Analysis of Data (LAD). It is used in this paper to assist the process of ABC items classification based on multiple criteria. LAD is a data mining technique based on Boolean theory that is used for pattern recognition. This technique has been tested in medicine, industry, credit risk analysis, and engineering with remarkable results. An application on ABC inventory classification is presented for the first time, and the results are compared with those obtained when using the well-known AHP technique and the ANN technique. The results show that LAD presented very good classification accuracy.

Keywords: ABC multi-criteria inventory classification, inventory management, multi-class LAD model, multi-criteria classification

Procedia PDF Downloads 856
679 Perceiving Interpersonal Conflict and the Big Five Personality Traits

Authors: Emily Rivera, Toni DiDona

Abstract:

The Big Five personality traits is a hierarchical classification of personality traits that applies factor analysis to a personality survey data in order to describe human personality using five broad dimensions: Extraversion, Agreeableness, Conscientiousness, Neuroticism, and Openness (Fetvadjiev & Van de Vijer, 2015). Research shows that personality constructs underline individual differences in processing conflict and interpersonal relations. (Graziano et al., 1996). This research explores the understudied correlation between the Big Five personality traits and perceived interpersonal conflict in the workplace. It revises social psychological literature on Big Five personality traits within a social context and discusses organizational development journal articles on the perceived efficacy of conflict tactics and approach to interpersonal relationships. The study also presents research undertaken on a survey group of 867 subjects over the age of 18 that were recruited by means of convenience sampling through social media, email, and text messaging. The central finding of this study is that only two of the Big Five personality traits had a significant correlation with perceiving interpersonal conflict in the workplace. Individuals who score higher on agreeableness and neuroticism, perceive more interpersonal conflict in the workplace compared to those that score lower on each dimension. The relationship between both constructs is worthy of research due to its everyday frequency and unique individual psycho-social consequences. This multimethod research associated the Big Five personality dimensions to interpersonal conflict. Its findings that can be utilized to further understand social cognition, person perception, complex social behavior and social relationships in the work environment.

Keywords: five-factor model, interpersonal conflict, personality, The Big Five personality traits

Procedia PDF Downloads 139
678 Artificial Intelligence Assisted Sentiment Analysis of Hotel Reviews Using Topic Modeling

Authors: Sushma Ghogale

Abstract:

With a surge in user-generated content or feedback or reviews on the internet, it has become possible and important to know consumers' opinions about products and services. This data is important for both potential customers and businesses providing the services. Data from social media is attracting significant attention and has become the most prominent channel of expressing an unregulated opinion. Prospective customers look for reviews from experienced customers before deciding to buy a product or service. Several websites provide a platform for users to post their feedback for the provider and potential customers. However, the biggest challenge in analyzing such data is in extracting latent features and providing term-level analysis of the data. This paper proposes an approach to use topic modeling to classify the reviews into topics and conduct sentiment analysis to mine the opinions. This approach can analyse and classify latent topics mentioned by reviewers on business sites or review sites, or social media using topic modeling to identify the importance of each topic. It is followed by sentiment analysis to assess the satisfaction level of each topic. This approach provides a classification of hotel reviews using multiple machine learning techniques and comparing different classifiers to mine the opinions of user reviews through sentiment analysis. This experiment concludes that Multinomial Naïve Bayes classifier produces higher accuracy than other classifiers.

Keywords: latent Dirichlet allocation, topic modeling, text classification, sentiment analysis

Procedia PDF Downloads 87
677 Characteristic Sentence Stems in Academic English Texts: Definition, Identification, and Extraction

Authors: Jingjie Li, Wenjie Hu

Abstract:

Phraseological units in academic English texts have been a central focus in recent corpus linguistic research. A wide variety of phraseological units have been explored, including collocations, chunks, lexical bundles, patterns, semantic sequences, etc. This paper describes a special category of clause-level phraseological units, namely, Characteristic Sentence Stems (CSSs), with a view to describing their defining criteria and extraction method. CSSs are contiguous lexico-grammatical sequences which contain a subject-predicate structure and which are frame expressions characteristic of academic writing. The extraction of CSSs consists of six steps: Part-of-speech tagging, n-gram segmentation, structure identification, significance of occurrence calculation, text range calculation, and overlapping sequence reduction. Significance of occurrence calculation is the crux of this study. It includes the computing of both the internal association and the boundary independence of a CSS and tests the occurring significance of the CSS from both inside and outside perspectives. A new normalization algorithm is also introduced into the calculation of LocalMaxs for reducing overlapping sequences. It is argued that many sentence stems are so recurrent in academic texts that the most typical of them have become the habitual ways of making meaning in academic writing. Therefore, studies of CSSs could have potential implications and reference value for academic discourse analysis, English for Academic Purposes (EAP) teaching and writing.

Keywords: characteristic sentence stem, extraction method, phraseological unit, the statistical measure

Procedia PDF Downloads 148
676 A Hybrid Feature Selection and Deep Learning Algorithm for Cancer Disease Classification

Authors: Niousha Bagheri Khulenjani, Mohammad Saniee Abadeh

Abstract:

Learning from very big datasets is a significant problem for most present data mining and machine learning algorithms. MicroRNA (miRNA) is one of the important big genomic and non-coding datasets presenting the genome sequences. In this paper, a hybrid method for the classification of the miRNA data is proposed. Due to the variety of cancers and high number of genes, analyzing the miRNA dataset has been a challenging problem for researchers. The number of features corresponding to the number of samples is high and the data suffer from being imbalanced. The feature selection method has been used to select features having more ability to distinguish classes and eliminating obscures features. Afterward, a Convolutional Neural Network (CNN) classifier for classification of cancer types is utilized, which employs a Genetic Algorithm to highlight optimized hyper-parameters of CNN. In order to make the process of classification by CNN faster, Graphics Processing Unit (GPU) is recommended for calculating the mathematic equation in a parallel way. The proposed method is tested on a real-world dataset with 8,129 patients, 29 different types of tumors, and 1,046 miRNA biomarkers, taken from The Cancer Genome Atlas (TCGA) database.

Keywords: cancer classification, feature selection, deep learning, genetic algorithm

Procedia PDF Downloads 95
675 An Analysis of Miguel Syjuco’s Ilustrado: The Reconstructed Oriental Image

Authors: Christine Ivy A. Nogot

Abstract:

Under the colony of Spain for more than three centuries, the Philippines has a deep-rooted structure of Western ideologies and colonialism. The late 19th century, the period of Enlightenment, created a significant impact on our history when a group of middle-class Filipino men were sent to Europe to study. They were called Ilustrados, a Spanish word for erudite. They were the enlightened; the well-educated, intellectual scholars. Their writings provide intellectual grounds for the awakening of national consciousness that eventually prompted national movements and revolutions. They helped to establish a postcolonial society. In the modern era, Miguel Syjuco, a Filipino expatriate, wrote a novel and titled it Ilustrado. It is a representation of the liberal mind of the diasporic author in contemporary discourse. It provides a critical examination of the ilustrado in transition through the character of Miguel, who is also an expatriate writer. Using Syjuco’s award-winning novel as the primary text and anchored on Said’s concept of Orientalism, this paper examines how the depiction of features of the Eastern world is presented in the literary discourse. This paper looks into Said’s concept of orientalism as a hegemonic discursive structure and shows how Western superiority influences the Eastern culture in literary discourse. It explores Gramsci’s theory of cultural hegemony to explore Said’s argument that Western powers conquer the orient through culture and ideology. This paper presents how dominant ideologies and the social context redefine the ilustrado in the contemporary era.

Keywords: cultural hegemony, ilustrado, orientalism, postcolonial

Procedia PDF Downloads 54
674 Improving Second Language Speaking Skills via Video Exchange

Authors: Nami Takase

Abstract:

Computer-mediated-communication allows people to connect and interact with each other as if they were sharing the same space. The current study examined the effects of using video letters (VLs) on the development of second language speaking skills of Common European Framework of Reference for Languages (CEFR) A1 and CEFR B2 level learners of English as a foreign language. Two groups were formed to measure the impact of VLs. The experimental and control groups were given the same topic, and both groups worked with a native English-speaking university student from the United States of America. Students in the experimental group exchanged VLs, and students in the control group used video conferencing. Pre- and post-tests were conducted to examine the effects of each practice mode. The transcribed speech-text data showed that the VL group had improved speech accuracy scores, while the video conferencing group had increased sentence complexity scores. The use of VLs may be more effective for beginner-level learners because they are able to notice their own errors and replay videos to better understand the native speaker’s speech at their own pace. Both the VL and video conferencing groups provided positive feedback regarding their interactions with native speakers. The results showed how different types of computer-mediated communication impacts different areas of language learning and speaking practice and how each of these types of online communication tool is suited to different teaching objectives.

Keywords: computer-assisted-language-learning, computer-mediated-communication, english as a foreign language, speaking

Procedia PDF Downloads 87
673 CompPSA: A Component-Based Pairwise RNA Secondary Structure Alignment Algorithm

Authors: Ghada Badr, Arwa Alturki

Abstract:

The biological function of an RNA molecule depends on its structure. The objective of the alignment is finding the homology between two or more RNA secondary structures. Knowing the common functionalities between two RNA structures allows a better understanding and a discovery of other relationships between them. Besides, identifying non-coding RNAs -that is not translated into a protein- is a popular application in which RNA structural alignment is the first step A few methods for RNA structure-to-structure alignment have been developed. Most of these methods are partial structure-to-structure, sequence-to-structure, or structure-to-sequence alignment. Less attention is given in the literature to the use of efficient RNA structure representation and the structure-to-structure alignment methods are lacking. In this paper, we introduce an O(N2) Component-based Pairwise RNA Structure Alignment (CompPSA) algorithm, where structures are given as a component-based representation and where N is the maximum number of components in the two structures. The proposed algorithm compares the two RNA secondary structures based on their weighted component features rather than on their base-pair details. Extensive experiments are conducted illustrating the efficiency of the CompPSA algorithm when compared to other approaches and on different real and simulated datasets. The CompPSA algorithm shows an accurate similarity measure between components. The algorithm gives the flexibility for the user to align the two RNA structures based on their weighted features (position, full length, and/or stem length). Moreover, the algorithm proves scalability and efficiency in time and memory performance.

Keywords: alignment, RNA secondary structure, pairwise, component-based, data mining

Procedia PDF Downloads 440
672 Application of Latent Class Analysis and Self-Organizing Maps for the Prediction of Treatment Outcomes for Chronic Fatigue Syndrome

Authors: Ben Clapperton, Daniel Stahl, Kimberley Goldsmith, Trudie Chalder

Abstract:

Chronic fatigue syndrome (CFS) is a condition characterised by chronic disabling fatigue and other symptoms that currently can't be explained by any underlying medical condition. Although clinical trials support the effectiveness of cognitive behaviour therapy (CBT), the success rate for individual patients is modest. Patients vary in their response and little is known which factors predict or moderate treatment outcomes. The aim of the project is to develop a prediction model from baseline characteristics of patients, such as demographics, clinical and psychological variables, which may predict likely treatment outcome and provide guidance for clinical decision making and help clinicians to recommend the best treatment. The project is aimed at identifying subgroups of patients with similar baseline characteristics that are predictive of treatment effects using modern cluster analyses and data mining machine learning algorithms. The characteristics of these groups will then be used to inform the types of individuals who benefit from a specific treatment. In addition, results will provide a better understanding of for whom the treatment works. The suitability of different clustering methods to identify subgroups and their response to different treatments of CFS patients is compared.

Keywords: chronic fatigue syndrome, latent class analysis, prediction modelling, self-organizing maps

Procedia PDF Downloads 205
671 MIMIC: A Multi Input Micro-Influencers Classifier

Authors: Simone Leonardi, Luca Ardito

Abstract:

Micro-influencers are effective elements in the marketing strategies of companies and institutions because of their capability to create an hyper-engaged audience around a specific topic of interest. In recent years, many scientific approaches and commercial tools have handled the task of detecting this type of social media users. These strategies adopt solutions ranging from rule based machine learning models to deep neural networks and graph analysis on text, images, and account information. This work compares the existing solutions and proposes an ensemble method to generalize them with different input data and social media platforms. The deployed solution combines deep learning models on unstructured data with statistical machine learning models on structured data. We retrieve both social media accounts information and multimedia posts on Twitter and Instagram. These data are mapped into feature vectors for an eXtreme Gradient Boosting (XGBoost) classifier. Sixty different topics have been analyzed to build a rule based gold standard dataset and to compare the performances of our approach against baseline classifiers. We prove the effectiveness of our work by comparing the accuracy, precision, recall, and f1 score of our model with different configurations and architectures. We obtained an accuracy of 0.91 with our best performing model.

Keywords: deep learning, gradient boosting, image processing, micro-influencers, NLP, social media

Procedia PDF Downloads 157
670 Unpacking Chilean Preservice Teachers’ Beliefs on Practicum Experiences through Digital Stories

Authors: Claudio Díaz, Mabel Ortiz

Abstract:

An EFL teacher education programme in Chile takes five years to train a future teacher of English. Preservice teachers are prepared to learn an advanced level of English and teach the language from 5th to 12th grade in the Chilean educational system. In the context of their first EFL Methodology course in year four, preservice teachers have to create a five-minute digital story that starts from a critical incident they have experienced as teachers-to-be during their observations or interventions in the schools. A critical incident can be defined as a happening, a specific incident or event either observed by them or involving them. The happening sparks their thinking and may make them subsequently think differently about the particular event. When they create their digital stories, preservice teachers put technology, teaching practice and theory together to narrate a story that is complemented by still images, moving images, text, sound effects and music. The story should be told as a personal narrative, which explains the critical incident. This presentation will focus on the creation process of 50 Chilean preservice teachers’ digital stories highlighting the critical incidents they started their stories. It will also unpack preservice teachers’ beliefs and reflections when approaching their teaching practices in schools. These beliefs will be coded and categorized through content analysis to evidence preservice teachers’ most rooted conceptions about English teaching and learning in Chilean schools. The findings seem to indicate that preservice teachers’ beliefs are strongly mediated by contextual and affective factors.

Keywords: beliefs, digital stories, preservice teachers, practicum

Procedia PDF Downloads 420
669 A Critical Geography of Reforestation Program in Ghana

Authors: John Narh

Abstract:

There is high rate of deforestation in Ghana due to agricultural expansion, illegal mining and illegal logging. While it is attempting to address the illegalities, Ghana has also initiated a reforestation program known as the Modified Taungya System (MTS). Within the MTS framework, farmers are allocated degraded forestland and provided with tree seedlings to practice agroforestry until the trees form canopy. Yet, the political, ecological and economic models that inform the selection of tree species, the motivations of participating farmers as well as the factors that accounts for differential access to the land and performance of farmers engaged in the program lie underexplored. Using a sequential explanatory mixed methods approach in five forest-fringe communities in the Eastern Region of Ghana, the study reveals that economic factors and Ghana’s commitment to international conventions on the environment underpin the selection of tree species for the MTS program. Social network and access to remittances play critical roles in having access to, and enhances poor farmers’ chances in the program respectively. Farmers are more motivated by the access to degraded forestland to cultivate food crops than having a share in the trees that they plant. As such, in communities where participating farmers are not informed about their benefit in the tree that they plant, the program is largely unsuccessful.

Keywords: translocality, deforestation, forest management, social network

Procedia PDF Downloads 75
668 Leaching Properties of Phosphate Rocks in the Nile River

Authors: Abdelkader T. Ahmed

Abstract:

Phosphate Rocks (PR) are natural sediment rocks. These rocks contain several chemical compositions of heavy metals and radioactive elements. Mining and transportation these rocks beside or through the natural water streams may lead to water contamination. When PR is in contact with water in the field, as a consequence of precipitation events, changes in water table or sinking in water streams, elements such as salts and heavy metals, may be released to the water. In this work, the leaching properties of PR in Nile River water was investigated by experimental lab work. The study focused on evaluating potential environmental impacts of some constituents, including phosphors, cadmium, curium and lead of PR on the water quality of Nile by applying tank leaching tests. In these tests the potential impact of changing conditions, such as phosphate content in PR, liquid to solid ratio (L/S) and pH value, was studied on the long-term release of heavy metals and salts. Experimental results showed that cadmium and lead were released in very low concentrations but curium and phosphors were in high concentrations. Results showed also that the release rate from PR for all constituents was low even in long periods.

Keywords: leaching tests, Nile river, phosphate rocks, water quality

Procedia PDF Downloads 305
667 Information Technology Approaches to Literature Text Analysis

Authors: Ayse Tarhan, Mustafa Ilkan, Mohammad Karimzadeh

Abstract:

Science was considered as part of philosophy in ancient Greece. By the nineteenth century, it was understood that philosophy was very inclusive and that social and human sciences such as literature, history, and psychology should be separated and perceived as an autonomous branch of science. The computer was also first seen as a tool of mathematical science. Over time, computer science has grown by encompassing every area in which technology exists, and its growth compelled the division of computer science into different disciplines, just as philosophy had been divided into different branches of science. Now there is almost no branch of science in which computers are not used. One of the newer autonomous disciplines of computer science is digital humanities, and one of the areas of digital humanities is literature. The material of literature is words, and thanks to the software tools created using computer programming languages, data that a literature researcher would need months to complete, can be achieved quickly and objectively. In this article, three different tools that literary researchers can use in their work will be introduced. These studies were created with the computer programming languages Python and R and brought to the world of literature. The purpose of introducing the aforementioned studies is to set an example for the development of special tools or programs on Ottoman language and literature in the future and to support such initiatives. The first example to be introduced is the Stylometry tool developed with the R language. The other is The Metrical Tool, which is used to measure data in poems and was developed with Python. The latest literature analysis tool in this article is Voyant Tools, which is a multifunctional and easy-to-use tool.

Keywords: DH, literature, information technologies, stylometry, the metrical tool, voyant tools

Procedia PDF Downloads 133
666 Ideology Shift in Political Translation

Authors: Jingsong Ma

Abstract:

In political translation, ideology plays an important role in conveying implications accurately. Ideological collisions can occur in political translation when there existdifferences of political environments embedded in the translingual political texts in both source and target languages. To reach an accurate translationrequires the translatorto understand the ideologies implied in (and often transcending) the texts. This paper explores the conditions, procedure, and purpose of processingideological collision and resolution of such issues in political translation. These points will be elucidated by case studies of translating English and Chinese political texts. First, there are specific political terminologies in certain political environments. These terminological peculiarities in one language are often determined by ideological elements rather than by syntactical and semantical understanding. The translation of these ideological-loaded terminologiesis a process and operation consisting of understanding the ideological context, including cultural, historical, and political situations. This will be explained with characteristic Chinese political terminologies and their renderings in English. Second, when the ideology in the source language fails to match with the ideology in the target language, the decisions to highlight or disregard these conflicts are shaped by power relations, political engagement, social context, etc. It thus is necessary to go beyond linguisticanalysis of the context by deciphering ideology in political documents to provide a faithful or equivalent rendering of certain messages. Finally, one of the practical issues is about equivalence in political translation by redefining the notion of faithfulness and retainment of ideological messages in the source language in translations of political texts. To avoid distortion, the translator should be liberated from grip the literal meaning, instead diving into functional meanings of the text.

Keywords: translation, ideology, politics, society

Procedia PDF Downloads 100
665 South Africa’s Post-Apartheid Film Narratives of HIV/AIDS: A Case of ‘Yesterday’

Authors: Moyahabo Molefe

Abstract:

The persistence of HIV/AIDS infection rates in SA has not only been a subject of academic debate but a mediated narrative that has dominated SA’s post-apartheid film space over the last two decades. SA’s colonial geo-spatial architecture still influences migrant labour patterns, which the Oscar-nominated (2003) SA film ‘Yesterday’ has erstwhile reflected upon, yet continues to account for the spread of HIV/AIDS in SA society. Accordingly, men who had left their homes in the rural areas to work in the mines in the cities become infected with HIV/AIDS, only to return home to infect their wives or partners in the rural areas. This paper analyses, through Social Semiotic theory, how SA geo-spatial arrangement had raptured family structures with both men and women taking new residences in the urban areas where they work away from their homes. By using Social semiotic theory, this paper seeks to understand how images and discourses have been deployed in the film ‘Yesterday’ to demonstrate how HIV/AIDS is embedded in the socio-cultural, economic and political architect of SA society. The study uses qualitative approach and content/text/visual semiotic analysis to decipher meanings from array of imagery and discourses/dialogues that are used to mythologise the relationship between the spread of HIV/AIDS and SA migrant labour patterns. The findings of the study are significant to propose a conceptual framework that can be used to mitigate the spread of HIV/AIDS among SA populace, against the backdrop of changing migrant labour patterns and other related factors

Keywords: colonialism, decoloniality, HIV/AIDS, labour migration patterns, social semiotics

Procedia PDF Downloads 52
664 Convergence and Stability in Federated Learning with Adaptive Differential Privacy Preservation

Authors: Rizwan Rizwan

Abstract:

This paper provides an overview of Federated Learning (FL) and its application in enhancing data security, privacy, and efficiency. FL utilizes three distinct architectures to ensure privacy is never compromised. It involves training individual edge devices and aggregating their models on a server without sharing raw data. This approach not only provides secure models without data sharing but also offers a highly efficient privacy--preserving solution with improved security and data access. Also we discusses various frameworks used in FL and its integration with machine learning, deep learning, and data mining. In order to address the challenges of multi--party collaborative modeling scenarios, a brief review FL scheme combined with an adaptive gradient descent strategy and differential privacy mechanism. The adaptive learning rate algorithm adjusts the gradient descent process to avoid issues such as model overfitting and fluctuations, thereby enhancing modeling efficiency and performance in multi-party computation scenarios. Additionally, to cater to ultra-large-scale distributed secure computing, the research introduces a differential privacy mechanism that defends against various background knowledge attacks.

Keywords: federated learning, differential privacy, gradient descent strategy, convergence, stability, threats

Procedia PDF Downloads 8
663 A Qualitative Study of Health-Related Beliefs and Practices among Vegetarians

Authors: Lorena Antonovici, Maria Nicoleta Turliuc

Abstract:

The process of becoming a vegetarian involves changes in several life aspects, including health. Despite its relevance, however, little research has been carried out to analyze vegetarians' self-perceived health, and even less empirical attention has received in the Romanian population. This study aimed to assess health-related beliefs and practices among vegetarian adults in a Romanian sample. We have undertaken 20 semi-structured interviews (10 males, 10 females) based on a snowball sample with a mean age of 31 years. The interview guide was divided into three sections: causes of adopting the diet, general aspects (beliefs, practices, tensions, and conflicts) and consequences of adopting the diet (significant changes, positive aspects, and difficulties, physical and mental health). Additional anamnestic data were reported by means of a questionnaire. Data analyses were performed using Tropes text analysis software (v. 8.2) and SPSS software (v. 24.0.) Findings showed that most of the participants considered a vegetarian diet as a natural and healthy choice as opposed to meat-eating, which is not healthy, and its consumption should be moderated among omnivores. A higher proportion of participants (65%) had an average body mass index (BMI), and several women even assumed having certain affections that no longer occur after following a vegetarian diet. Moreover, participants admitted having better moods and mental health status, given their self-contentment with the dietary choice. Relatives were perceived as more skeptical about their practices than others, and especially women had this view. This study provides a valuable insight into health-related beliefs and practices and how a vegetarian diet might interact.

Keywords: beliefs, health, practices, vegetarians

Procedia PDF Downloads 106
662 Python Implementation for S1000D Applicability Depended Processing Model - SALERNO

Authors: Theresia El Khoury, Georges Badr, Amir Hajjam El Hassani, Stéphane N’Guyen Van Ky

Abstract:

The widespread adoption of machine learning and artificial intelligence across different domains can be attributed to the digitization of data over several decades, resulting in vast amounts of data, types, and structures. Thus, data processing and preparation turn out to be a crucial stage. However, applying these techniques to S1000D standard-based data poses a challenge due to its complexity and the need to preserve logical information. This paper describes SALERNO, an S1000d AppLicability dEpended pRocessiNg mOdel. This python-based model analyzes and converts the XML S1000D-based files into an easier data format that can be used in machine learning techniques while preserving the different logic and relationships in files. The model parses the files in the given folder, filters them, and extracts the required information to be saved in appropriate data frames and Excel sheets. Its main idea is to group the extracted information by applicability. In addition, it extracts the full text by replacing internal and external references while maintaining the relationships between files, as well as the necessary requirements. The resulting files can then be saved in databases and used in different models. Documents in both English and French languages were tested, and special characters were decoded. Updates on the technical manuals were taken into consideration as well. The model was tested on different versions of the S1000D, and the results demonstrated its ability to effectively handle the applicability, requirements, references, and relationships across all files and on different levels.

Keywords: aeronautics, big data, data processing, machine learning, S1000D

Procedia PDF Downloads 110
661 Balance Transfer of Heavy Metals in Marine Environments Subject to Natural and Anthropogenic Inputs: A Case Study on the Mejerda River Delta

Authors: Mohamed Amine Helali, Walid Oueslati, Ayed Added

Abstract:

Sedimentation rates and total fluxes of heavy metals (Fe, Mn, Pb, Zn and Cu) was measured in three different depths (10m, 20m and 40m) during March and August 2012, offshore of the Mejerda River outlet (Gulf of Tunis, Tunisia). The sedimentation rates are estimated from the fluxes of the suspended particulate matter at 7.32, 5.45 and 4.39 mm y⁻¹ respectively at 10m, 20m and 40m depth. Heavy metals sequestration in sediments was determined by chemical speciation and the total metal contents in each core collected from 10, 20 and 40m depth. Heavy metals intake to the sediment was measured also from the suspended particulate matter, while the fluxes from the sediment to the water column was determined using the benthic chambers technique and from the diffusive fluxes in the pore water. Results shown that iron is the only metal for which the balance transfer between intake/uptake (45 to 117 / 1.8 to 5.8 g m² y⁻¹) and sequestration (277 to 378 g m² y⁻¹) was negative, at the opposite of the Lead which intake fluxes (360 to 480 mg m² y⁻¹) are more than sequestration fluxes (50 to 92 mg m² y⁻¹). The balance transfer is neutral for Mn, Zn, and Cu. These clearly indicate that the contributions of Mejerda have consistently varied over time, probably due to the migration of the River mouth and to the changes in the mining activity in the Mejerda catchment and the recent human activities which affect the delta area.

Keywords: delta, fluxes, heavy metals, sediments, sedimentation rates

Procedia PDF Downloads 192
660 A Corpus-Based Approach to Understanding Market Access in Fisheries and Aquaculture: A Systematic Literature Review

Authors: Cheryl Marie Cordeiro

Abstract:

Although fisheries and aquaculture studies might seem marginal to international business (IB) studies in general, fisheries and aquaculture IB (FAIB) management is currently facing increasing pressure to meet global demand and consumption for fish in the next coming decades. In part address to this challenge, the purpose of this systematic review of literature (SLR) study is to investigate the use of the term ‘market access’ in its context of use in the generic literature and business sector discourse, in comparison to the more specific literature and discourse in fisheries, aquaculture and seafood. This SLR aims to uncover the knowledge/interest gaps between the academic subject discourses and business sector practices. Corpus driven in methodology and using a triangulation method of three different text analysis software including AntConc, VOSviewer and Web of Science (WoS) analytics, the SLR results indicate a gap in conceptual knowledge and business practices in how ‘market access’ is conceived and used in the context of the pharmaceutical healthcare industry and FAIB research and practice. While it is acknowledged that the product orientation of different business sectors might differ, this SLR study works with the assumption that both business sectors are global in orientation. These business sectors are complex in their operations from product to market. This SLR suggests a conceptual model in understanding the challenges, the potential barriers as well as avenues for solutions to developing market access for FAIB.

Keywords: market access, fisheries and aquaculture, international business, systematic literature review

Procedia PDF Downloads 130
659 Comparing the Sequence and Effectiveness of Teaching the Four Basic Operations and Mathematics in Primary Schools

Authors: Abubakar Sadiq Mensah, Hassan Usman

Abstract:

The study compared the effectiveness of Audition, Multiplication, subtraction and Division (AMSD) and Addition, subtraction, Multiplication and Division (ASMD), sequence of teaching these four basic operations in mathematics to primary one pupil’s in Katsina Local Government, Katsina State. The study determined the sequence that was more effective and mostly adopted by teachers of the operations. One hundred (100) teachers and sixty pupils (60) from primary one were used for the study. The pupils were divided into two equal groups. The researcher taught these operations to each group separately for four weeks (4 weeks). Group one was taught using the ASMD sequence, while group two was taught using ASMD sequence. In order to generate the needed data for the study, questionnaires and tests were administered on the samples. Data collected were analyzed and major findings were arrived at: (i) Two primary mathematics text books were used in all the primary schools in the area; (ii) Each of the textbooks contained the ASMD sequence; (iii) 73% of the teachers sampled adopted the ASMD sequence of teaching these operations; and (iv) Group one of the pupils (taught using AMSD sequence) performed significantly better than their counter parts in group two (taught using AMSD sequence). On the basis of this, the researcher concluded that the AMSD sequence was more effective in teaching the operations than the ASMD sequence. Consequently, the researcher concluded that primary schools teachers, authors of primary mathematics textbooks, and curriculum planner should adopt the AMSD sequence of teaching these operations.

Keywords: matematic, high school, four basic operations, effectiveness of teaching

Procedia PDF Downloads 237