Search results for: text comprehension EIAH
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1628

Search results for: text comprehension EIAH

1538 Activation of Google Classroom Features to Engage Introvert Students in Comprehensible Output

Authors: Raghad Dwaik

Abstract:

It is well known in language acquisition literature that a mere understanding of a reading text is not enough to help students build proficiency in comprehension. Students should rather follow understanding by attempting to express what has been understood by pushing their competence to the limit. Learners' attempt to push their competence was given the term "comprehensible output" by Swain (1985). Teachers in large classes, however, find it sometimes difficult to give all students a chance to communicate their views or to share their ideas during the short class time. In most cases, students who are outgoing dominate class discussion and get more opportunities for practice which leads to ignoring the shy students totally while helping the good ones become better. This paper presents the idea of using Google Classroom features of posting and commenting to allow students who hesitate to participate in class discussions about a reading text to write their views on the wall of a Google Classroom and share them later after they have received feedback and comments from classmates. Such attempts lead to developing their proficiency through additional practice in comprehensible output and to enhancing their confidence in themselves and their views. It was found that virtual classroom interaction would help students maintain vocabulary, use more complex structures and focus on meaning besides form.

Keywords: learning groups, reading TESOL, Google Classroom, comprehensible output

Procedia PDF Downloads 43
1537 Improving Vocabulary and Listening Comprehension via Watching French Films without Subtitles: Positive Results

Authors: Yelena Mazour-Matusevich, Jean-Robert Ancheta

Abstract:

This study is based on more than fifteen years of experience of teaching a foreign language, in my case French, to the English-speaking students. It represents a qualitative research on foreign language learners’ reaction and their gains in terms of vocabulary and listening comprehension through repeatedly viewing foreign feature films with the original sountrack but without English subtitles. The initial idea emerged upon realization that the first challenge faced by my students when they find themselves in a francophone environment has been their lack of listening comprehension. Their inability to understand colloquial speech affects not only their academic performance, but their psychological health as well. To remedy this problem, I have designed and applied for many years my own teaching method based on one particular French film, exceptionally suited, for the reasons described in detail in the paper, for the intermediate-advanced level foreign language learners. This project, conducted together with my undergraduate assistant and mentoree J-R Ancheta, aims at showing how the paralinguistic features, such as characters’ facial expressions, settings, music, historical background, images provided before the actual viewing, etc., offer crucial support and enhance students’ listening comprehension. The study, based on students’ interviews, also offers special pedagogical techniques, such as ‘anticipatory’ vocabulary lists and exercises, drills, quizzes and composition topics that have proven to boost students’ performance. For this study, only the listening proficiency and vocabulary gains of the interviewed participants were assessed.

Keywords: comprehension, film, listening, subtitles, vocabulary

Procedia PDF Downloads 597
1536 Experimental Study of Hyperparameter Tuning a Deep Learning Convolutional Recurrent Network for Text Classification

Authors: Bharatendra Rai

Abstract:

The sequence of words in text data has long-term dependencies and is known to suffer from vanishing gradient problems when developing deep learning models. Although recurrent networks such as long short-term memory networks help to overcome this problem, achieving high text classification performance is a challenging problem. Convolutional recurrent networks that combine the advantages of long short-term memory networks and convolutional neural networks can be useful for text classification performance improvements. However, arriving at suitable hyperparameter values for convolutional recurrent networks is still a challenging task where fitting a model requires significant computing resources. This paper illustrates the advantages of using convolutional recurrent networks for text classification with the help of statistically planned computer experiments for hyperparameter tuning.

Keywords: long short-term memory networks, convolutional recurrent networks, text classification, hyperparameter tuning, Tukey honest significant differences

Procedia PDF Downloads 90
1535 Off-Topic Text Detection System Using a Hybrid Model

Authors: Usama Shahid

Abstract:

Be it written documents, news columns, or students' essays, verifying the content can be a time-consuming task. Apart from the spelling and grammar mistakes, the proofreader is also supposed to verify whether the content included in the essay or document is relevant or not. The irrelevant content in any document or essay is referred to as off-topic text and in this paper, we will address the problem of off-topic text detection from a document using machine learning techniques. Our study aims to identify the off-topic content from a document using Echo state network model and we will also compare data with other models. The previous study uses Convolutional Neural Networks and TFIDF to detect off-topic text. We will rearrange the existing datasets and take new classifiers along with new word embeddings and implement them on existing and new datasets in order to compare the results with the previously existing CNN model.

Keywords: off topic, text detection, eco state network, machine learning

Procedia PDF Downloads 57
1534 Towards Logical Inference for the Arabic Question-Answering

Authors: Wided Bakari, Patrice Bellot, Omar Trigui, Mahmoud Neji

Abstract:

This article constitutes an opening to think of the modeling and analysis of Arabic texts in the context of a question-answer system. It is a question of exceeding the traditional approaches focused on morphosyntactic approaches. Furthermore, we present a new approach that analyze a text in order to extract correct answers then transform it to logical predicates. In addition, we would like to represent different levels of information within a text to answer a question and choose an answer among several proposed. To do so, we transform both the question and the text into logical forms. Then, we try to recognize all entailment between them. The results of recognizing the entailment are a set of text sentences that can implicate the user’s question. Our work is now concentrated on an implementation step in order to develop a system of question-answering in Arabic using techniques to recognize textual implications. In this context, the extraction of text features (keywords, named entities, and relationships that link them) is actually considered the first step in our process of text modeling. The second one is the use of techniques of textual implication that relies on the notion of inference and logic representation to extract candidate answers. The last step is the extraction and selection of the desired answer.

Keywords: NLP, Arabic language, question-answering, recognition text entailment, logic forms

Procedia PDF Downloads 311
1533 The Process of Irony Comprehension in Young Children: Evidence from Monolingual and Bilingual Preschoolers

Authors: Natalia Banasik

Abstract:

Comprehension of verbal irony is an example of pragmatic competence in understanding figurative language. The knowledge of how it develops may shed new light on the understanding of social and communicative competence that is crucial for one's effective functioning in the society. Researchers agree it is a competence that develops late in a child’s development. One of the abilities that seems crucial for irony comprehension is theory of mind (ToM), that is the ability to understand that others may have beliefs, desires and intentions different from one’s own. Although both theory of mind and irony comprehension require the ability to understand the figurative use of the false description of the reality, the exact relationship between them is still unknown. Also, even though irony comprehension in children has been studied for over thirty years, the results of the studies are inconsistent as to the age when this competence are acquired. The presented study aimed to answer questions about the developmental trajectories of irony comprehension and ascribing function to ironic utterances by preschool children. Specifically, we were interested in how it is related to the development of ToM and how comprehension of the function of irony changes with age. Data was collected from over 150 monolingual, Polish-speaking children and (so far) thirty bilingual children speaking Polish and English who live in the US. Four-, five- and six-year-olds were presented with a story comprehension task in the form of audio and visual stimuli programmed in the E-prime software (pre-recorded narrated stories, some of which included ironic utterances, and pictures accompanying the stories displayed on a touch screen). Following the presentation, the children were then asked to answer a series of questions. The questions checked the children’s understanding of the intended utterance meaning, evaluation of the degree to which it was funny and evaluation of how nice the speaker was. The children responded by touching the screen, which made it possible to measure reaction times. Additionally, the children were asked to explain why the speaker had uttered the ironic statement. Both quantitive and qualitative analyses were applied. The results of our study indicate that for irony recognition there is a significant difference among the three age groups, but what is new is that children as young as four do understand the real meaning behind the ironic statement as long as the utterance is not grammtically or lexically complex also, there is a clear correlation of ToM and irony comprehension. Although four-year olds and six-year olds understand the real meaning of the ironic utterance, it is not earlier than at the age of six when children start to explain the reason of using this marked form of expression. They talk about the speaker's intention to tell a joke, be funny, or to protect the listener's emotions. There are also some metalinguistic references, such as "mommy sometimes says things that don't make sense and this is called a metaphor".

Keywords: child's pragmatics, figurative speech, irony comprehension in children, theory of mind and irony

Procedia PDF Downloads 289
1532 The Composer’s Hand: An Analysis of Arvo Pärt’s String Orchestral Work, Psalom

Authors: Mark K. Johnson

Abstract:

Arvo Pärt has composed over 80 text-based compositions based on nine different languages. But prior to 2015, it was not publicly known what texts the composer used in composing a number of his non-vocal works, nor the language of those texts. Because of this lack of information, few if any musical scholars have illustrated in any detail how textual structure applies to any of Pärt’s instrumental compositions. However, in early 2015, the Arvo Pärt Centre in Estonia published In Principio, a compendium of the texts Pärt has used to derive many of the parameters of his text-based compositions. This paper provides the first detailed analysis of the relationship between structural aspects of the Church Slavonic Eastern Orthodox text of Psalm 112 and the musical parameters that Pärt used when composing the string orchestral work Psalom. It demonstrates that Pärt’s text-based compositions are carefully crafted works, and that evidence of the presence of the ‘invisible’ hand of the composer can be found within every aspect of the underpinning structures, at the more elaborate middle ground level, and even within surface aspects of these works. Based on the analysis of Psalom, it is evident that the text Pärt selected for Psalom informed many of his decisions regarding the musical structures, parameters and processes that he deployed in composing this non-vocal text-based work. Many of these composerly decisions in relation to these various aspects cannot be fathomed without access to, and an understanding of, the text associated with the work.

Keywords: Arvo Pärt, minimalism, psalom, text-based process music

Procedia PDF Downloads 200
1531 N400 Investigation of Semantic Priming Effect to Symbolic Pictures in Text

Authors: Thomas Ousterhout

Abstract:

The purpose of this study was to investigate if incorporating meaningful pictures of gestures and facial expressions in short sentences of text could supplement the text with enough semantic information to produce and N400 effect when probe words incongruent to the picture were subsequently presented. Event-related potentials (ERPs) were recorded from a 14-channel commercial grade EEG headset while subjects performed congruent/incongruent reaction time discrimination tasks. Since pictures of meaningful gestures have been shown to be semantically processed in the brain in a similar manner as words are, it is believed that pictures will add supplementary information to text just as the inclusion of their equivalent synonymous word would. The hypothesis is that when subjects read the text/picture mixed sentences, they will process the images and words just like in face-to-face communication and therefore probe words incongruent to the image will produce an N400.

Keywords: EEG, ERP, N400, semantics, congruency, facilitation, Emotiv

Procedia PDF Downloads 238
1530 The Use of Videos: Effects on Children's Language and Literacy Skills

Authors: Rahimah Saimin

Abstract:

Previous research has shown that young children can learn from educational television programmes, videos or other technological media. However, the blending of any of these with traditional printed-based text appears to be omitted. Repeated viewing is an important factor in children's ability to comprehend the content or plot. The present study combined videos with traditional printed-based text and required repeated viewing and is original and distinctive. The first study was a pilot study to explore whether the intervention is implementable in ordinary classrooms. The second study explored whether the curricular embedding is important or whether the video with curricular embedding is effective. The third study explored the effect of “dosage”, i.e. whether a longer/ more intense intervention has a proportionately greater effect on outcomes. Both measured outcomes (comprehension, word sounds, and early word recognition) and unmeasured outcomes (engagement to reading traditional printed-based texts or/and multimodal texts) were obtained from this study. Observation indicated degree of engagement in reading. The theoretical framework was multimodality theory combined with Piaget’s and Vygotsky’s learning theories. An experimental design was used with 4-5-year-old children in nursery schools and primary schools. Six links to video clips exploring non-fiction science content were provided to teachers. The first session is whole-class and subsequent sessions small-group. The teacher then engaged the children in dialogue using supplementary materials. About half of each class was selected randomly for pre-post assessments. Two assessments were used the British Picture Vocabulary Scale (BPVSIII) and the York Assessment of Reading for Comprehension (YARC): Early Reading. Different programme fidelity means were deployed- observations, teacher self-reports attendance logs and post-delivery interviews. Data collection is in progress and results will be available shortly. If this multiphase study show effectiveness in one or other application, then teachers will have other tools which they can use to enhance vocabulary, letter knowledge and word reading. This would be a valuable addition to their repertoire.

Keywords: language skills, literacy skills, multimodality, video

Procedia PDF Downloads 309
1529 Electroencephalogram during Natural Reading: Theta and Alpha Rhythms as Analytical Tools for Assessing a Reader’s Cognitive State

Authors: D. Zhigulskaya, V. Anisimov, A. Pikunov, K. Babanova, S. Zuev, A. Latyshkova, K. Сhernozatonskiy, A. Revazov

Abstract:

Electrophysiology of information processing in reading is certainly a popular research topic. Natural reading, however, has been relatively poorly studied, despite having broad potential applications for learning and education. In the current study, we explore the relationship between text categories and spontaneous electroencephalogram (EEG) while reading. Thirty healthy volunteers (mean age 26,68 ± 1,84) participated in this study. 15 Russian-language texts were used as stimuli. The first text was used for practice and was excluded from the final analysis. The remaining 14 were opposite pairs of texts in one of 7 categories, the most important of which were: interesting/boring, fiction/non-fiction, free reading/reading with an instruction, reading a text/reading a pseudo text (consisting of strings of letters that formed meaningless words). Participants had to read the texts sequentially on an Apple iPad Pro. EEG was recorded from 12 electrodes simultaneously with eye movement data via ARKit Technology by Apple. EEG spectral amplitude was analyzed in Fz for theta-band (4-8 Hz) and in C3, C4, P3, and P4 for alpha-band (8-14 Hz) using the Friedman test. We found that reading an interesting text was accompanied by an increase in theta spectral amplitude in Fz compared to reading a boring text (3,87 µV ± 0,12 and 3,67 µV ± 0,11, respectively). When instructions are given for reading, we see less alpha activity than during free reading of the same text (3,34 µV ± 0,20 and 3,73 µV ± 0,28, respectively, for C4 as the most representative channel). The non-fiction text elicited less activity in the alpha band (C4: 3,60 µV ± 0,25) than the fiction text (C4: 3,66 µV ± 0,26). A significant difference in alpha spectral amplitude was also observed between the regular text (C4: 3,64 µV ± 0,29) and the pseudo text (C4: 3,38 µV ± 0,22). These results suggest that some brain activity we see on EEG is sensitive to particular features of the text. We propose that changes in theta and alpha bands during reading may serve as electrophysiological tools for assessing the reader’s cognitive state as well as his or her attitude to the text and the perceived information. These physiological markers have prospective practical value for developing technological solutions and biofeedback systems for reading in particular and for education in general.

Keywords: EEG, natural reading, reader's cognitive state, theta-rhythm, alpha-rhythm

Procedia PDF Downloads 58
1528 Literature Review on Text Comparison Techniques: Analysis of Text Extraction, Main Comparison and Visual Representation Tools

Authors: Andriana Mkrtchyan, Vahe Khlghatyan

Abstract:

The choice of a profession is one of the most important decisions people make throughout their life. With the development of modern science, technologies, and all the spheres existing in the modern world, more and more professions are being arisen that complicate even more the process of choosing. Hence, there is a need for a guiding platform to help people to choose a profession and the right career path based on their interests, skills, and personality. This review aims at analyzing existing methods of comparing PDF format documents and suggests that a 3-stage approach is implemented for the comparison, that is – 1. text extraction from PDF format documents, 2. comparison of the extracted text via NLP algorithms, 3. comparison representation using special shape and color psychology methodology.

Keywords: color psychology, data acquisition/extraction, data augmentation, disambiguation, natural language processing, outlier detection, semantic similarity, text-mining, user evaluation, visual search

Procedia PDF Downloads 44
1527 Interactive, Topic-Oriented Search Support by a Centroid-Based Text Categorisation

Authors: Mario Kubek, Herwig Unger

Abstract:

Centroid terms are single words that semantically and topically characterise text documents and so may serve as their very compact representation in automatic text processing. In the present paper, centroids are used to measure the relevance of text documents with respect to a given search query. Thus, a new graphbased paradigm for searching texts in large corpora is proposed and evaluated against keyword-based methods. The first, promising experimental results demonstrate the usefulness of the centroid-based search procedure. It is shown that especially the routing of search queries in interactive and decentralised search systems can be greatly improved by applying this approach. A detailed discussion on further fields of its application completes this contribution.

Keywords: search algorithm, centroid, query, keyword, co-occurrence, categorisation

Procedia PDF Downloads 256
1526 Binarization and Recognition of Characters from Historical Degraded Documents

Authors: Bency Jacob, S.B. Waykar

Abstract:

Degradations in historical document images appear due to aging of the documents. It is very difficult to understand and retrieve text from badly degraded documents as there is variation between the document foreground and background. Thresholding of such document images either result in broken characters or detection of false texts. Numerous algorithms exist that can separate text and background efficiently in the textual regions of the document; but portions of background are mistaken as text in areas that hardly contain any text. This paper presents a way to overcome these problems by a robust binarization technique that recovers the text from a severely degraded document images and thereby increases the accuracy of optical character recognition systems. The proposed document recovery algorithm efficiently removes degradations from document images. Here we are using the ostus method ,local thresholding and global thresholding and after the binarization training and recognizing the characters in the degraded documents.

Keywords: binarization, denoising, global thresholding, local thresholding, thresholding

Procedia PDF Downloads 314
1525 Effects of Audiovisual Contextualization of L2 Idioms on Enhancing Students’ Comprehension and Retention

Authors: Monica Karlsson

Abstract:

The positive effect of a supportive written context on comprehension and retention when faced with a previously unknown idiomatic expression is today an indisputable fact, especially if relevant clues are given in close proximity of the item in question. Also, giving learners a chance of visualizing the meaning of an idiom by offering them its source domain and/or by elaborating etymologically, i.e. providing a mental picture in addition to the spoken/written form (referred to as dual coding), seems to enhance comprehension and retention even further, especially if the idiom is of a more transparent kind. For example, by explaining that walk the plank has a maritime origin and a canary in a coal mine comes from the time when canaries were kept in cages to warn miners if gas was leaking out at which point the canaries succumbed immediately, learners’ comprehension and retention have been shown to increase. The present study aims to investigate whether contextualization of an audiovisual kind could help increase comprehension and retention of L2 idioms. 40 Swedish first-term university students studying English as part of their education to become middle-school teachers participated in the investigation, which tested 24 idioms, all of which were ascertained to be previously unknown to the informants. While half of the learners were subjected to a test in which they were asked to watch scenes from various TV programmes, each scene including one idiomatic expression in a supportive context, the remaining 20 students, as a point of reference, were only offered written contexts, though equally supportive. Immediately after these sessions, both groups were given the same idioms in a decontextualized form and asked to give their meaning. After five weeks, finally, the students were subjected to yet another decontextualized comprehension test. Furthermore, since mastery of idioms in one’s L1 appears to correlate to a great extent with a person’s ability to comprehend idioms in an L2, all the informants were also asked to take a test focusing on idioms in their L1. The result on this test is thus seen to indicate each student’s potential for understanding and memorizing various idiomatic expressions from a more general perspective. Preliminary results clearly show that audiovisual contextualization indeed has a positive effect on learners’ retention. In addition, preliminary results also show that those learners’ who were able to recall most meanings were those who had a propensity for idiom comprehension in their L1.

Keywords: English, L2, idioms, audiovisual context

Procedia PDF Downloads 317
1524 Adaptation of Projection Profile Algorithm for Skewed Handwritten Text Line Detection

Authors: Kayode A. Olaniyi, Tola. M. Osifeko, Adeola A. Ogunleye

Abstract:

Text line segmentation is an important step in document image processing. It represents a labeling process that assigns the same label using distance metric probability to spatially aligned units. Text line detection techniques have successfully been implemented mainly in printed documents. However, processing of the handwritten texts especially unconstrained documents has remained a key problem. This is because the unconstrained hand-written text lines are often not uniformly skewed. The spaces between text lines may not be obvious, complicated by the nature of handwriting and, overlapping ascenders and/or descenders of some characters. Hence, text lines detection and segmentation represents a leading challenge in handwritten document image processing. Text line detection methods that rely on the traditional global projection profile of the text document cannot efficiently confront with the problem of variable skew angles between different text lines. Hence, the formulation of a horizontal line as a separator is often not efficient. This paper presents a technique to segment a handwritten document into distinct lines of text. The proposed algorithm starts, by partitioning the initial text image into columns, across its width into chunks of about 5% each. At each vertical strip of 5%, the histogram of horizontal runs is projected. We have worked with the assumption that text appearing in a single strip is almost parallel to each other. The algorithm developed provides a sliding window through the first vertical strip on the left side of the page. It runs through to identify the new minimum corresponding to a valley in the projection profile. Each valley would represent the starting point of the orientation line and the ending point is the minimum point on the projection profile of the next vertical strip. The derived text-lines traverse around any obstructing handwritten vertical strips of connected component by associating it to either the line above or below. A decision of associating such connected component is made by the probability obtained from a distance metric decision. The technique outperforms the global projection profile for text line segmentation and it is robust to handle skewed documents and those with lines running into each other.

Keywords: connected-component, projection-profile, segmentation, text-line

Procedia PDF Downloads 95
1523 An Experimental Study of Scalar Implicature Processing in Chinese

Authors: Liu Si, Wang Chunmei, Liu Huangmei

Abstract:

A prominent component of the semantic versus pragmatic debate, scalar implicature (SI) has been gaining great attention ever since it was proposed by Horn. The constant debate is between the structural and pragmatic approach. The former claims that generation of SI is costless, automatic, and dependent mostly on the structural properties of sentences, whereas the latter advocates both that such generation is largely dependent upon context, and that the process is costly. Many experiments, among which Katsos’s text comprehension experiments are influential, have been designed and conducted in order to verify their views, but the results are not conclusive. Besides, most of the experiments were conducted in English language materials. Katsos conducted one off-line and three on-line text comprehension experiments, in which the previous shortcomings were addressed on a certain extent and the conclusion was in favor of the pragmatic approach. We intend to test the results of Katsos’s experiment in Chinese scalar implicature. Four experiments in both off-line and on-line conditions to examine the generation and response time of SI in Chinese "yixie" (some) and "quanbu (dou)" (all) will be conducted in order to find out whether the structural or the pragmatic approach could be sustained. The study mainly aims to answer the following questions: (1) Can SI be generated in the upper- and lower-bound contexts as Katsos confirmed when Chinese language materials are used in the experiment? (2) Can SI be first generated, then cancelled as default view claimed or can it not be generated in a neutral context when Chinese language materials are used in the experiment? (3) Is SI generation costless or costly in terms of processing resources? (4) In line with the SI generation process, what conclusion can be made about the cognitive processing model of language meaning? Is it a parallel model or a linear model? Or is it a dynamic and hierarchical model? According to previous theoretical debates and experimental conflicts, presumptions could be made that SI, in Chinese language, might be generated in the upper-bound contexts. Besides, the response time might be faster in upper-bound than that found in lower-bound context. SI generation in neutral context might be the slowest. At last, a conclusion would be made that the processing model of SI could not be verified by either absolute structural or pragmatic approaches. It is, rather, a dynamic and complex processing mechanism, in which the interaction of language forms, ad hoc context, mental context, background knowledge, speakers’ interaction, etc. are involved.

Keywords: cognitive linguistics, pragmatics, scalar implicture, experimental study, Chinese language

Procedia PDF Downloads 338
1522 Glossematics and Textual Structure

Authors: Abdelhadi Nadjer

Abstract:

The structure of the text to the systemic school -(glossématique-Helmslev). At the beginning of the note we have a cursory look around the concepts of general linguistics The science that studies scientific study of human language based on the description and preview the facts away from the trend of education than we gave a detailed overview the founder of systemic school and most important customers and more methods and curriculum theory and analysis they extend to all humanities, practical action each offset by a theoretical and the procedure can be analyzed through the elements that pose as another method we talked to its links with other language schools where they are based on the sharp criticism of the language before and deflected into consideration for the field of language and its erection has outside or language network and its participation in the actions (non-linguistic) and after that we started our Valglosamatik analytical structure of the text is ejected text terminal or all of the words to was put for expression. This text Negotiable divided into types in turn are divided into classes and class should not be carrying a contradiction and be inclusive. It is on the same materials as described relationships that combine language and seeks to describe their relations and identified.

Keywords: text, language schools, linguistics, human language

Procedia PDF Downloads 430
1521 We Wonder If They Mind: An Empirical Inquiry into the Narratological Function of Mind Wandering in Readers of Literary Texts

Authors: Tina Ternes, Florian Kleinau

Abstract:

The study investigates the content and triggers of mind wandering (MW) in readers of fictional texts. It asks whether readers’ MW is productive (text-related) or unproductive (text-unrelated). Methodologically, it bridges the gap between narratological and data-driven approaches by utilizing a sentence-by-sentence self-paced reading paradigm combined with thought probes in the reading of an excerpt of A. L. Kennedy’s “Baby Blue”. Results show that the contents of MW can be linked to text properties. We validated the role of self-reference in MW and found prediction errors to be triggers of MW. Results also indicate that the content of MW often travels along the lines of the text at hand and can thus be viewed as productive and integral to interpretation.

Keywords: narratology, mind wandering, reading fiction, meta cognition

Procedia PDF Downloads 55
1520 Story Readers’ Self-Reflection on Their past Study Experiences: In Comparison of the Languages Used in a Self-Regulated Learning -Themed Story

Authors: Mayuko Matsuoka

Abstract:

This presentation reports the relationships among EFL(English as a Foreign Language) students’ story comprehension in reading a story written in English and Japanese and empathic reactions. The main focus is put on their self-reflection on past study experiences, one of the empathic reactions after reading a story. One hundred fifty-five first-year university students in Japan read three SRL-themed stories written in English (their foreign language) and those written in Japanese (their mother tongue). The levels of the stories are equivalent, at CEFR(Common European Framework of Reference for Languages) B2 level. The result of categorical correlation analysis shows significant moderate correlations among three empathic reactions in a group reading English versions: having similar emotions as a protagonist, reflecting on their past study experiences, and getting lessons from a story. In addition, the result of logistic regression analysis for the data in a group reading English versions shows the chance of getting lessons from a story significantly approximately doubles if participants’ scores of a comprehension test increases by one, while it approximately triples if participants’ self-reflection occurs. These results do not appear in a group reading Japanese versions. The findings imply that self-reflection may support their comprehension of the English texts and leads to the participants’ getting lessons about SRL.

Keywords: comprehension, lesson, self-reflection, SRL

Procedia PDF Downloads 154
1519 Incorporating Information Gain in Regular Expressions Based Classifiers

Authors: Rosa L. Figueroa, Christopher A. Flores, Qing Zeng-Treitler

Abstract:

A regular expression consists of sequence characters which allow describing a text path. Usually, in clinical research, regular expressions are manually created by programmers together with domain experts. Lately, there have been several efforts to investigate how to generate them automatically. This article presents a text classification algorithm based on regexes. The algorithm named REX was designed, and then, implemented as a simplified method to create regexes to classify Spanish text automatically. In order to classify ambiguous cases, such as, when multiple labels are assigned to a testing example, REX includes an information gain method Two sets of data were used to evaluate the algorithm’s effectiveness in clinical text classification tasks. The results indicate that the regular expression based classifier proposed in this work performs statically better regarding accuracy and F-measure than Support Vector Machine and Naïve Bayes for both datasets.

Keywords: information gain, regular expressions, smith-waterman algorithm, text classification

Procedia PDF Downloads 294
1518 Memory Retrieval and Implicit Prosody during Reading: Anaphora Resolution by L1 and L2 Speakers of English

Authors: Duong Thuy Nguyen, Giulia Bencini

Abstract:

The present study examined structural and prosodic factors on the computation of antecedent-reflexive relationships and sentence comprehension in native English (L1) and Vietnamese-English bilinguals (L2). Participants read sentences presented on the computer screen in one of three presentation formats aimed at manipulating prosodic parsing: word-by-word (RSVP), phrase-segment (self-paced), or whole-sentence (self-paced), then completed a grammaticality rating and a comprehension task (following Pratt & Fernandez, 2016). The design crossed three factors: syntactic structure (simple; complex), grammaticality (target-match; target-mismatch) and presentation format. An example item is provided in (1): (1) The actress that (Mary/John) interviewed at the awards ceremony (about two years ago/organized outside the theater) described (herself/himself) as an extreme workaholic). Results showed that overall, both L1 and L2 speakers made use of a good-enough processing strategy at the expense of more detailed syntactic analyses. L1 and L2 speakers’ comprehension and grammaticality judgements were negatively affected by the most prosodically disrupting condition (word-by-word). However, the two groups demonstrated differences in their performance in the other two reading conditions. For L1 speakers, the whole-sentence and the phrase-segment formats were both facilitative in the grammaticality rating and comprehension tasks; for L2, compared with the whole-sentence condition, the phrase-segment paradigm did not significantly improve accuracy or comprehension. These findings are consistent with the findings of Pratt & Fernandez (2016), who found a similar pattern of results in the processing of subject-verb agreement relations using the same experimental paradigm and prosodic manipulation with English L1 and L2 English-Spanish speakers. The results provide further support for a Good-Enough cue model of sentence processing that integrates cue-based retrieval and implicit prosodic parsing (Pratt & Fernandez, 2016) and highlights similarities and differences between L1 and L2 sentence processing and comprehension.

Keywords: anaphora resolution, bilingualism, implicit prosody, sentence processing

Procedia PDF Downloads 121
1517 The English as a Foreign Language Teachers’ Perceptions and Practices of Infusing Critical Thinking Skills to Improve Students’ Reading Comprehension

Authors: Michael Amale Kirko, Abebe Gebretsadik

Abstract:

In the 21st century, developing students’ critical thinking skills has become a prime concern in higher education institutions. Cognizant of this fact, the Ethiopian higher education policy document used critical thinking as one of the guiding principles. The study aims to explore how English as a foreign language (EFL) teachers perceive and practice critical thinking skills (CTS) in teaching reading to improve reading comprehension at Wolaita Sodo University, Ethiopia. A descriptive survey study used an exploratory mixed-methods approach. The study involved 20 EFL instructors and 40 2nd-year English majoring students. The numerical data were collected using teacher and student surveys and classroom observations; the qualitative data were obtained through content analysis and interviews. Teacher survey results indicated that teachers' perceptions are above average (mean = 3.41). And the result of classroom observations showed the practice CTS in class was below average (mean=2.61). The content analysis result revealed instructors utilized fewer higher-order thinking questions during class activities, quizzes, midterm, and final exams. The teachers perceived that teacher, student, and material-related challenges were hindering the practice of CT to improve students’ reading comprehension. Finally, spearman’s rho output showed r=0.97 and p<0.05. Therefore, the results showed that the EFL teachers’ practices of CTS to improve students’ reading comprehension were less frequent; there was a strong, positive, and statistically significant relationship between the teachers’ perceptions and practices of CTS in reading class.

Keywords: perceptions, critical thinking skills, practices, infusing thinking skills, reading comprehension

Procedia PDF Downloads 50
1516 Examining the Links between Established Principles, Iranian Teachers' Perceptions of Reading Comprehension, and Their Actual Practice in English for Specific Purposes Courses

Authors: Zahra Alimorad

Abstract:

There is a strong belief that language teachers' actual practices in the classroom context are largely determined by the underlying perceptions they hold about the nature of language and language learning. That being so, it can be envisaged that teaching procedures of ESP (English for Specific Purposes) teachers teaching reading comprehension will mainly be driven by their perceptions about the nature of reading. To examine this issue, four Iranian university professors holding Ph.D. in either TEFL (Teaching English as a Foreign Language) or English Literature who were teaching English to Engineering and Sciences students were recruited to participate in this study. To collect the necessary data, classroom observations and follow-up semi-structured interviews were used. Furthermore, the materials utilized by the teachers such as textbooks, syllabuses, and tests were also examined. Although it can be argued that their perceptions were partially compatible with the established principles, results of the study pointed to a lack of congruence between these teachers' perceptions and their practices, on the one hand, and between the established principles and the practices, on the other. While the literature mostly supports a metacognitive-strategy approach to reading comprehension, the teachers were mainly adopting a skills-based approach to the teaching of reading. That is, they primarily focused on translation as the core activity in the classroom followed by reading aloud, defining words, and explaining grammatical structures. This divergence was partly attributed to the contextual constraints and partly to students' lack of motivation by the teachers.

Keywords: English teachers, perceptions, practice, principles, reading comprehension

Procedia PDF Downloads 218
1515 Multi-Class Text Classification Using Ensembles of Classifiers

Authors: Syed Basit Ali Shah Bukhari, Yan Qiang, Saad Abdul Rauf, Syed Saqlaina Bukhari

Abstract:

Text Classification is the methodology to classify any given text into the respective category from a given set of categories. It is highly important and vital to use proper set of pre-processing , feature selection and classification techniques to achieve this purpose. In this paper we have used different ensemble techniques along with variance in feature selection parameters to see the change in overall accuracy of the result and also on some other individual class based features which include precision value of each individual category of the text. After subjecting our data through pre-processing and feature selection techniques , different individual classifiers were tested first and after that classifiers were combined to form ensembles to increase their accuracy. Later we also studied the impact of decreasing the classification categories on over all accuracy of data. Text classification is highly used in sentiment analysis on social media sites such as twitter for realizing people’s opinions about any cause or it is also used to analyze customer’s reviews about certain products or services. Opinion mining is a vital task in data mining and text categorization is a back-bone to opinion mining.

Keywords: Natural Language Processing, Ensemble Classifier, Bagging Classifier, AdaBoost

Procedia PDF Downloads 207
1514 Resource Creation Using Natural Language Processing Techniques for Malay Translated Qur'an

Authors: Nor Diana Ahmad, Eric Atwell, Brandon Bennett

Abstract:

Text processing techniques for English have been developed for several decades. But for the Malay language, text processing methods are still far behind. Moreover, there are limited resources, tools for computational linguistic analysis available for the Malay language. Therefore, this research presents the use of natural language processing (NLP) in processing Malay translated Qur’an text. As the result, a new language resource for Malay translated Qur’an was created. This resource will help other researchers to build the necessary processing tools for the Malay language. This research also develops a simple question-answer prototype to demonstrate the use of the Malay Qur’an resource for text processing. This prototype has been developed using Python. The prototype pre-processes the Malay Qur’an and an input query using a stemming algorithm and then searches for occurrences of the query word stem. The result produced shows improved matching likelihood between user query and its answer. A POS-tagging algorithm has also been produced. The stemming and tagging algorithms can be used as tools for research related to other Malay texts and can be used to support applications such as information retrieval, question answering systems, ontology-based search and other text analysis tasks.

Keywords: language resource, Malay translated Qur'an, natural language processing (NLP), text processing

Procedia PDF Downloads 287
1513 Text Based Shuffling Algorithm on Graphics Processing Unit for Digital Watermarking

Authors: Zayar Phyo, Ei Chaw Htoon

Abstract:

In a New-LSB based Steganography method, the Fisher-Yates algorithm is used to permute an existing array randomly. However, that algorithm performance became slower and occurred memory overflow problem while processing the large dimension of images. Therefore, the Text-Based Shuffling algorithm aimed to select only necessary pixels as hiding characters at the specific position of an image according to the length of the input text. In this paper, the enhanced text-based shuffling algorithm is presented with the powered of GPU to improve more excellent performance. The proposed algorithm employs the OpenCL Aparapi framework, along with XORShift Kernel including the Pseudo-Random Number Generator (PRNG) Kernel. PRNG is applied to produce random numbers inside the kernel of OpenCL. The experiment of the proposed algorithm is carried out by practicing GPU that it can perform faster-processing speed and better efficiency without getting the disruption of unnecessary operating system tasks.

Keywords: LSB based steganography, Fisher-Yates algorithm, text-based shuffling algorithm, OpenCL, XORShiftKernel

Procedia PDF Downloads 122
1512 Text Mining of Twitter Data Using a Latent Dirichlet Allocation Topic Model and Sentiment Analysis

Authors: Sidi Yang, Haiyi Zhang

Abstract:

Twitter is a microblogging platform, where millions of users daily share their attitudes, views, and opinions. Using a probabilistic Latent Dirichlet Allocation (LDA) topic model to discern the most popular topics in the Twitter data is an effective way to analyze a large set of tweets to find a set of topics in a computationally efficient manner. Sentiment analysis provides an effective method to show the emotions and sentiments found in each tweet and an efficient way to summarize the results in a manner that is clearly understood. The primary goal of this paper is to explore text mining, extract and analyze useful information from unstructured text using two approaches: LDA topic modelling and sentiment analysis by examining Twitter plain text data in English. These two methods allow people to dig data more effectively and efficiently. LDA topic model and sentiment analysis can also be applied to provide insight views in business and scientific fields.

Keywords: text mining, Twitter, topic model, sentiment analysis

Procedia PDF Downloads 153
1511 Text Localization in Fixed-Layout Documents Using Convolutional Networks in a Coarse-to-Fine Manner

Authors: Beier Zhu, Rui Zhang, Qi Song

Abstract:

Text contained within fixed-layout documents can be of great semantic value and so requires a high localization accuracy, such as ID cards, invoices, cheques, and passports. Recently, algorithms based on deep convolutional networks achieve high performance on text detection tasks. However, for text localization in fixed-layout documents, such algorithms detect word bounding boxes individually, which ignores the layout information. This paper presents a novel architecture built on convolutional neural networks (CNNs). A global text localization network and a regional bounding-box regression network are introduced to tackle the problem in a coarse-to-fine manner. The text localization network simultaneously locates word bounding points, which takes the layout information into account. The bounding-box regression network inputs the features pooled from arbitrarily sized RoIs and refine the localizations. These two networks share their convolutional features and are trained jointly. A typical type of fixed-layout documents: ID cards, is selected to evaluate the effectiveness of the proposed system. These networks are trained on data cropped from nature scene images, and synthetic data produced by a synthetic text generation engine. Experiments show that our approach locates high accuracy word bounding boxes and achieves state-of-the-art performance.

Keywords: bounding box regression, convolutional networks, fixed-layout documents, text localization

Procedia PDF Downloads 167
1510 Recognition of Cursive Arabic Handwritten Text Using Embedded Training Based on Hidden Markov Models (HMMs)

Authors: Rabi Mouhcine, Amrouch Mustapha, Mahani Zouhir, Mammass Driss

Abstract:

In this paper, we present a system for offline recognition cursive Arabic handwritten text based on Hidden Markov Models (HMMs). The system is analytical without explicit segmentation used embedded training to perform and enhance the character models. Extraction features preceded by baseline estimation are statistical and geometric to integrate both the peculiarities of the text and the pixel distribution characteristics in the word image. These features are modelled using hidden Markov models and trained by embedded training. The experiments on images of the benchmark IFN/ENIT database show that the proposed system improves recognition.

Keywords: recognition, handwriting, Arabic text, HMMs, embedded training

Procedia PDF Downloads 325
1509 Poetics of the Connecting ha’: A Textual Study in the Poetry of Al-Husari Al-Qayrawani

Authors: Mahmoud al-Ashiriy

Abstract:

This paper begins from the idea that the real history of literature is the history of its style. And since the rhyme –as known- is not merely the last letter, that have received a lot of analysis and investigation, but it is a collection of other values in addition to its different markings. This paper will explore the work of the connecting ha’ and its effectiveness in shaping the text of poetry, since it establishes vocal rhythms in addition to its role in indicating references through the pronoun, vertically through the poem through the sequence of its verses, also horizontally through what environs the one verse of sentences. If the scientific formation of prosody stopped at the possibilities and prohibitions; literary criticism and poetry studies should explore what is above the rule of aesthetic horizon of poetic effectiveness that varies from a text to another, a poet to another, a literary period to another, or from a poetic taste to another. Then the paper will explore this poetic essence in the texts of the famous Andalusian Poet Al-Husari Al-Qayrawani through his well-known Daliyya (a poem that its verses end with the letter D), and the role of the connecting ha’ in fulfilling its text and the accomplishment of its poetics, departing from this to the diwan (the big collection of poems) also as a higher text that surpasses the text/poem, and through what it represents of effectiveness the work of the phenomenon in accomplishing the poetics of the poem of Al-Husari Al-Qayrawani who is one of the pillars of Arabic poetics in Andalusia.

Keywords: Al-Husari Al-Qayrawni, poetics, rhyme, stylistics, science of the text

Procedia PDF Downloads 538