Search results for: working memory model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19894

Search results for: working memory model

19654 Motivation Needs in Working of the Employees in Rayong Province: A Case Study of Panakom Co., Ltd.

Authors: Ganratchakan Ninlawan, Witthaya Mekhum

Abstract:

The objective of this research was to investigate motivation needs in working of the employees in Rayong Province at Panakom Co., Ltd. The sample group included 59 operational employees of Panakom Co., Ltd divided into 2 main parts to complete the questionnaires. Part 1 dealt with personal information of the staff in the form of checklist questions. Part 2 was about the motivation needs in working in the form of 5 rating scales. The data were analyzed to find frequency, percentage, arithmetic mean, and SD with the results as follows. They reported their working conditions in the moderate level while the first aspect was on the regulation of the Working Support Department. The second rated aspect was on the modern facilities and office stationeries followed by the working environment such as lighting, temperature, sound system, and atmosphere. The last aspect was on the provision of enough working equipment. On the part of work stability and work progress, they rated this aspect at the moderate level with the details below. The first aspect was their satisfaction in work rotation followed by the encouragement in joining training and seminar to increase working knowledge. The third aspect was their perception on the progress of their careers. The last aspect was on the chance to get promoted in special cases. On the payment and fringe benefit, they rated this part in the moderate level with the highest aspect being on the provided fringe benefit such as health care. The second aspect was on the suitable salary compared to their knowledge and proficiency. The third aspect was on the satisfaction on the activities and seminars provided by the company. Finally, the last aspect was on the sufficient salary when compared with the current cost of living.

Keywords: motivation needs, working, employees, Rayong Province

Procedia PDF Downloads 244
19653 A Corpus-Based Study on the Lexical, Syntactic and Sequential Features across Interpreting Types

Authors: Qianxi Lv, Junying Liang

Abstract:

Among the various modes of interpreting, simultaneous interpreting (SI) is regarded as a ‘complex’ and ‘extreme condition’ of cognitive tasks while consecutive interpreters (CI) do not have to share processing capacity between tasks. Given that SI exerts great cognitive demand, it makes sense to posit that the output of SI may be more compromised than that of CI in the linguistic features. The bulk of the research has stressed the varying cognitive demand and processes involved in different modes of interpreting; however, related empirical research is sparse. In keeping with our interest in investigating the quantitative linguistic factors discriminating between SI and CI, the current study seeks to examine the potential lexical simplification, syntactic complexity and sequential organization mechanism with a self-made inter-model corpus of transcribed simultaneous and consecutive interpretation, translated speech and original speech texts with a total running word of 321960. The lexical features are extracted in terms of the lexical density, list head coverage, hapax legomena, and type-token ratio, as well as core vocabulary percentage. Dependency distance, an index for syntactic complexity and reflective of processing demand is employed. Frequency motif is a non-grammatically-bound sequential unit and is also used to visualize the local function distribution of interpreting the output. While SI is generally regarded as multitasking with high cognitive load, our findings evidently show that CI may impose heavier or taxing cognitive resource differently and hence yields more lexically and syntactically simplified output. In addition, the sequential features manifest that SI and CI organize the sequences from the source text in different ways into the output, to minimize the cognitive load respectively. We reasoned the results in the framework that cognitive demand is exerted both on maintaining and coordinating component of Working Memory. On the one hand, the information maintained in CI is inherently larger in volume compared to SI. On the other hand, time constraints directly influence the sentence reformulation process. The temporal pressure from the input in SI makes the interpreters only keep a small chunk of information in the focus of attention. Thus, SI interpreters usually produce the output by largely retaining the source structure so as to relieve the information from the working memory immediately after formulated in the target language. Conversely, CI interpreters receive at least a few sentences before reformulation, when they are more self-paced. CI interpreters may thus tend to retain and generate the information in a way to lessen the demand. In other words, interpreters cope with the high demand in the reformulation phase of CI by generating output with densely distributed function words, more content words of higher frequency values and fewer variations, simpler structures and more frequently used language sequences. We consequently propose a revised effort model based on the result for a better illustration of cognitive demand during both interpreting types.

Keywords: cognitive demand, corpus-based, dependency distance, frequency motif, interpreting types, lexical simplification, sequential units distribution, syntactic complexity

Procedia PDF Downloads 168
19652 Music Therapy Intervention as a Means of Stimulating Communicative Abilities of Seniors with Neurocognitive Disorders – Theory versus Practice

Authors: Pavel Svoboda, Oldřich Müller

Abstract:

The paper contains a screening of the opinions of helping professional workers working in a home for seniors with individuals with neurocognitive disorders and compares them with the opinions of a younger generation of students who are just preparing for this work. The authors carried out a comparative questionnaire survey with both target groups, focusing on the analysis and comparison of possible differences in their knowledge in the field of care for elderly people with neurocognitive disorders. Specifically, they focused on knowledge and experience with approaches, methods and tools applicable within the framework of music therapy interventions, as they are understood in practice in comparison with the theoretical knowledge of secondary school students focused on social work. The questionnaire was mainly aimed at assessing the knowledge of the possibilities of effective memory stimulation of the elderly and their communication skills using the means of music. The conducted investigation was based on the research of studies dealing with so-called non-pharmacological approaches to the given clientele; for professional caregivers, it followed music therapy lessons, which the authors regularly implemented from the beginning of 2022. Its results will, among other things, serve as the basis for an upcoming study with a scoping design review.

Keywords: neurocognitive disorders, seniors, music therapy intervention, melody, rhythm, text, memory stimulation, communication skills

Procedia PDF Downloads 66
19651 Security Design of Root of Trust Based on RISC-V

Authors: Kang Huang, Wanting Zhou, Shiwei Yuan, Lei Li

Abstract:

Since information technology develops rapidly, the security issue has become an increasingly critical for computer system. In particular, as cloud computing and the Internet of Things (IoT) continue to gain widespread adoption, computer systems need to new security threats and attacks. The Root of Trust (RoT) is the foundation for providing basic trusted computing, which is used to verify the security and trustworthiness of other components. Design a reliable Root of Trust and guarantee its own security are essential for improving the overall security and credibility of computer systems. In this paper, we discuss the implementation of self-security technology based on the RISC-V Root of Trust at the hardware level. To effectively safeguard the security of the Root of Trust, researches on security safeguard technology on the Root of Trust have been studied. At first, a lightweight and secure boot framework is proposed as a secure mechanism. Secondly, two kinds of memory protection mechanism are built to against memory attacks. Moreover, hardware implementation of proposed method has been also investigated. A series of experiments and tests have been carried on to verify to effectiveness of the proposed method. The experimental results demonstrated that the proposed approach is effective in verifying the integrity of the Root of Trust’s own boot rom, user instructions, and data, ensuring authenticity and enabling the secure boot of the Root of Trust’s own system. Additionally, our approach provides memory protection against certain types of memory attacks, such as cache leaks and tampering, and ensures the security of root-of-trust sensitive information, including keys.

Keywords: root of trust, secure boot, memory protection, hardware security

Procedia PDF Downloads 196
19650 Reducing Falls in Memory Care through Implementation of the Stopping Elderly Accidents, Deaths, and Injuries Program

Authors: Cory B. Lord

Abstract:

Falls among the elderly population has become an area of concern in healthcare today. The negative impacts of falls lead to increased morbidity, mortality, and financial burdens for both patients and healthcare systems. Falls in the United States is reported at an annual rate of 36 million in those aged 65 and older. Each year, one out of four people in this age group will suffer a fall, with 20% of these falls causing injury. The setting for this Doctor of Nursing Practice (DNP) project was a memory care unit in an assisted living community, as these facilities house cognitively impaired older adults. These communities lack fall prevention programs; therefore, the need exists to add to the body of knowledge to positively impact this population. The objective of this project was to reduce fall rates through the implementation of the Center for Disease Control and Prevention (CDC) STEADI (stopping elderly accidents, deaths, and injuries) program. The DNP project performed was a quality improvement pilot study with a pre and post-test design. This program was implemented in the memory care setting over 12 weeks. The project included an educational session for staff and a fall risk assessment with appropriate resident referrals. The three aims of the DNP project were to reduce fall rates among the elderly aged 65 and older who reside in the memory care unit, increase staff knowledge of STEADI fall prevention measures after an educational session, and assess the willingness of memory care unit staff to adopt an evidence-based a fall prevention program. The Donabedian model was used as a guiding conceptual framework for this quality improvement pilot study. The fall rate data for 12 months before the intervention was evaluated and compared to post-intervention fall rates. The educational session comprised of a pre and post-test to assess staff knowledge of the fall prevention program and the willingness of staff to adopt the fall prevention program. The overarching goal was to reduce falls in the elderly population who live in memory care units. The results of the study showed, on average that the fall rate during the implementation period of STEADI (μ=6.79) was significantly lower when compared to the prior 12 months (μ= 9.50) (p=0.02, α = 0.05). The mean staff knowledge scores improved from pretest (μ=77.74%) to post-test (μ=87.42%) (p=0.00, α= 0.05) after the education session. The results of the willingness to adopt a fall prevention program were scored at 100%. In summation, implementing the STEADI fall prevention program can assist in reducing fall rates for residents aged 65 and older who reside in a memory care setting.

Keywords: dementia, elderly, falls, STEADI

Procedia PDF Downloads 119
19649 An Algorithm of Set-Based Particle Swarm Optimization with Status Memory for Traveling Salesman Problem

Authors: Takahiro Hino, Michiharu Maeda

Abstract:

Particle swarm optimization (PSO) is an optimization approach that achieves the social model of bird flocking and fish schooling. PSO works in continuous space and can solve continuous optimization problem with high quality. Set-based particle swarm optimization (SPSO) functions in discrete space by using a set. SPSO can solve combinatorial optimization problem with high quality and is successful to apply to the large-scale problem. In this paper, we present an algorithm of SPSO with status memory to decide the position based on the previous position for solving traveling salesman problem (TSP). In order to show the effectiveness of our approach. We examine SPSOSM for TSP compared to the existing algorithms.

Keywords: combinatorial optimization problems, particle swarm optimization, set-based particle swarm optimization, traveling salesman problem

Procedia PDF Downloads 543
19648 Machine Learning Approach for Stress Detection Using Wireless Physical Activity Tracker

Authors: B. Padmaja, V. V. Rama Prasad, K. V. N. Sunitha, E. Krishna Rao Patro

Abstract:

Stress is a psychological condition that reduces the quality of sleep and affects every facet of life. Constant exposure to stress is detrimental not only for mind but also body. Nevertheless, to cope with stress, one should first identify it. This paper provides an effective method for the cognitive stress level detection by using data provided from a physical activity tracker device Fitbit. This device gathers people’s daily activities of food, weight, sleep, heart rate, and physical activities. In this paper, four major stressors like physical activities, sleep patterns, working hours and change in heart rate are used to assess the stress levels of individuals. The main motive of this system is to use machine learning approach in stress detection with the help of Smartphone sensor technology. Individually, the effect of each stressor is evaluated using logistic regression and then combined model is built and assessed using variants of ordinal logistic regression models like logit, probit and complementary log-log. Then the quality of each model is evaluated using Akaike Information Criterion (AIC) and probit is assessed as the more suitable model for our dataset. This system is experimented and evaluated in a real time environment by taking data from adults working in IT and other sectors in India. The novelty of this work lies in the fact that stress detection system should be less invasive as possible for the users.

Keywords: physical activity tracker, sleep pattern, working hours, heart rate, smartphone sensor

Procedia PDF Downloads 249
19647 Extrudate Swell under the Effect of Radial Flow and Intrinsic Factors to the Polymer Upstream of the Die

Authors: Hela Krir, Abdelhak Ayadi, Chedly Bradaii

Abstract:

The influence of both intrinsic factors, elastic energy and memory effect, and radial flow on the appearance and the evolution of the extrudate swelling are investigated in the present work. The experiments have been performed with linear polydimethylsiloxane (PDMS) via a capillary rheometer in which a convergent radial flow was created upstream the contraction. The correspondence between the effects of radial flow, entry elastic stored energy and memory effect is discussed. In particular, as the influence of the considered radial flow, extrudate photographs showed that when the gap ratio is reduced, the extrudate swell is lessened than what it is when radial flow geometry is not installed. Moreover, with a narrower gap, the polymer stores less energy during its passage through the die which implies a lower extrudate swelling at the outlet of the die. Results previously mentioned may be related both to shear and elongational components of radial flow.

Keywords: elastic energy, extrudate swell, memory effect, radial flow

Procedia PDF Downloads 166
19646 A Dirty Page Migration Method in Process of Memory Migration Based on Pre-copy Technology

Authors: Kang Zijian, Zhang Tingyu, Burra Venkata Durga Kumar

Abstract:

This article investigates the challenges in memory migration during the live migration of virtual machines. We found three challenges probably existing in pre-copy technology. One of the main challenges is the challenge of downtime migration. Decrease the downtime could promise the normal work for a virtual machine. Although pre-copy technology is greatly decreasing the downtime, we still need to shut down the machine in order to finish the last round of data transfer. This paper provides an optimization scheme for the problems existing in pro-copy technology, mainly the optimization of the dirty page migration mechanism. The typical pre-copy technology copy n-1th’s dirty pages in nth turn. However, our idea is to create a double iteration method to solve this problem.

Keywords: virtual machine, pre-copy technology, memory migration process, downtime, dirty pages migration method

Procedia PDF Downloads 135
19645 Effector and Memory Immune Responses Induced by Total Extracts of Naegleria fowleri Co-Administered with Cholera Toxin

Authors: Q. B. Maria de la Luz Ortega Juárez, Saúl Rojas Hernández, Itzel Berenice Rodríguez Mera, María Maricela Carrasco Yépez, Mara Gutierrez Sánchez

Abstract:

Naegleria fowleri is a free-living amoeba found mainly in temperate freshwater and is the etiologic agent of primary amebic meningoencephalitis (PAM), a fatal acute disease with a mortality rate greater than 95%. At present, there are no treatments available for MAP, and the development of effective vaccines that generate long-term immunological memory allowing protection against MAP would be of great importance. The objective of this work was to analyze the effector and memory immune response in BALB/c mice immunized with total extract of N. fowleri co-administered with cholera toxin. In this study, BALB/c mice were immunized four times intranasally with ET of N. fowleri adjuvanted with CT with or without booster at three months and were challenged or not with the lethal dose of N. fowleri, determining survival, the humoral, effector and memory response, by ELISA and flow cytometry techniques. The results obtained showed that the survival of mice immunized with booster had 60% protection compared to the group without booster, which obtained 20% protection. Evaluating the humoral response, it was found that both IgG and IgA levels were higher in sera than in nasal washes in both treatments. In the cellular response, the increase in the percentage of positive cells was found for effector T and B lymphocytes in the nasal passages (NP) in the group with boost and nasopharynx-associated lymphoid tissue (NALT) in the group without boost and lymphocytes only. B in both treatments, as well as in memory cells treatment with boost T lymphocytes in PN and NALT and without boost in cervical lymph nodes (CG) with respect to B lymphocytes, in PN, GC and NALT in treatment with boost and NALT in treatment without booster. Therefore, the involvement of the effector immune response and memory play a fundamental role for protection against N. fowleri and for the development of vaccine candidates.

Keywords: immune response, immunological memory, naegleria fowleri, primary amebic meningoencephalitis

Procedia PDF Downloads 69
19644 A Machine Learning Approach for Earthquake Prediction in Various Zones Based on Solar Activity

Authors: Viacheslav Shkuratskyy, Aminu Bello Usman, Michael O’Dea, Saifur Rahman Sabuj

Abstract:

This paper examines relationships between solar activity and earthquakes; it applied machine learning techniques: K-nearest neighbour, support vector regression, random forest regression, and long short-term memory network. Data from the SILSO World Data Center, the NOAA National Center, the GOES satellite, NASA OMNIWeb, and the United States Geological Survey were used for the experiment. The 23rd and 24th solar cycles, daily sunspot number, solar wind velocity, proton density, and proton temperature were all included in the dataset. The study also examined sunspots, solar wind, and solar flares, which all reflect solar activity and earthquake frequency distribution by magnitude and depth. The findings showed that the long short-term memory network model predicts earthquakes more correctly than the other models applied in the study, and solar activity is more likely to affect earthquakes of lower magnitude and shallow depth than earthquakes of magnitude 5.5 or larger with intermediate depth and deep depth.

Keywords: k-nearest neighbour, support vector regression, random forest regression, long short-term memory network, earthquakes, solar activity, sunspot number, solar wind, solar flares

Procedia PDF Downloads 67
19643 Employing Remotely Sensed Soil and Vegetation Indices and Predicting ‎by Long ‎Short-Term Memory to Irrigation Scheduling Analysis

Authors: Elham Koohikerade, Silvio Jose Gumiere

Abstract:

In this research, irrigation is highlighted as crucial for improving both the yield and quality of ‎potatoes due to their high sensitivity to soil moisture changes. The study presents a hybrid Long ‎Short-Term Memory (LSTM) model aimed at optimizing irrigation scheduling in potato fields in ‎Quebec City, Canada. This model integrates model-based and satellite-derived datasets to simulate ‎soil moisture content, addressing the limitations of field data. Developed under the guidance of the ‎Food and Agriculture Organization (FAO), the simulation approach compensates for the lack of direct ‎soil sensor data, enhancing the LSTM model's predictions. The model was calibrated using indices ‎like Surface Soil Moisture (SSM), Normalized Vegetation Difference Index (NDVI), Enhanced ‎Vegetation Index (EVI), and Normalized Multi-band Drought Index (NMDI) to effectively forecast ‎soil moisture reductions. Understanding soil moisture and plant development is crucial for assessing ‎drought conditions and determining irrigation needs. This study validated the spectral characteristics ‎of vegetation and soil using ECMWF Reanalysis v5 (ERA5) and Moderate Resolution Imaging ‎Spectrometer (MODIS) data from 2019 to 2023, collected from agricultural areas in Dolbeau and ‎Peribonka, Quebec. Parameters such as surface volumetric soil moisture (0-7 cm), NDVI, EVI, and ‎NMDI were extracted from these images. A regional four-year dataset of soil and vegetation moisture ‎was developed using a machine learning approach combining model-based and satellite-based ‎datasets. The LSTM model predicts soil moisture dynamics hourly across different locations and ‎times, with its accuracy verified through cross-validation and comparison with existing soil moisture ‎datasets. The model effectively captures temporal dynamics, making it valuable for applications ‎requiring soil moisture monitoring over time, such as anomaly detection and memory analysis. By ‎identifying typical peak soil moisture values and observing distribution shapes, irrigation can be ‎scheduled to maintain soil moisture within Volumetric Soil Moisture (VSM) values of 0.25 to 0.30 ‎m²/m², avoiding under and over-watering. The strong correlations between parcels suggest that a ‎uniform irrigation strategy might be effective across multiple parcels, with adjustments based on ‎specific parcel characteristics and historical data trends. The application of the LSTM model to ‎predict soil moisture and vegetation indices yielded mixed results. While the model effectively ‎captures the central tendency and temporal dynamics of soil moisture, it struggles with accurately ‎predicting EVI, NDVI, and NMDI.‎

Keywords: irrigation scheduling, LSTM neural network, remotely sensed indices, soil and vegetation ‎monitoring

Procedia PDF Downloads 37
19642 Micro-Rest: Extremely Short Breaks in Post-Learning Interference Support Memory Retention over the Long Term

Authors: R. Marhenke, M. Martini

Abstract:

The distraction of attentional resources after learning hinders long-term memory consolidation compared to several minutes of post-encoding inactivity in form of wakeful resting. We tested whether an 8-minute period of wakeful resting, compared to performing an adapted version of the d2 test of attention after learning, supports memory retention. Participants encoded and immediately recalled a word list followed by either an 8 minute period of wakeful resting (eyes closed, relaxed) or by performing an adapted version of the d2 test of attention (scanning and selecting specific characters while ignoring others). At the end of the experimental session (after 12-24 min) and again after 7 days, participants were required to complete a surprise free recall test of both word lists. Our results showed no significant difference in memory retention between the experimental conditions. However, we found that participants who completed the first lines of the d2 test in less than the given time limit of 20 seconds and thus had short unfilled intervals before switching to the next test line, remembered more words over the 12-24 minute and over the 7 days retention interval than participants who did not complete the first lines. This interaction occurred only for the first test lines, with the highest temporal proximity to the encoding task and not for later test lines. Differences in retention scores between groups (completed first line vs. did not complete) seem to be widely independent of the general performance in the d2 test. Implications and limitations of these exploratory findings are discussed.

Keywords: long-term memory, retroactive interference, attention, forgetting

Procedia PDF Downloads 126
19641 Accuracy of Small Field of View CBCT in Determining Endodontic Working Length

Authors: N. L. S. Ahmad, Y. L. Thong, P. Nambiar

Abstract:

An in vitro study was carried out to evaluate the feasibility of small field of view (FOV) cone beam computed tomography (CBCT) in determining endodontic working length. The objectives were to determine the accuracy of CBCT in measuring the estimated preoperative working lengths (EPWL), endodontic working lengths (EWL) and file lengths. Access cavities were prepared in 27 molars. For each root canal, the baseline electronic working length was determined using an EAL (Raypex 5). The teeth were then divided into overextended, non-modified and underextended groups and the lengths were adjusted accordingly. Imaging and measurements were made using the respective software of the RVG (Kodak RVG 6100) and CBCT units (Kodak 9000 3D). Root apices were then shaved and the apical constrictions viewed under magnification to measure the control working lengths. The paired t-test showed a statistically significant difference between CBCT EPWL and control length but the difference was too small to be clinically significant. From the Bland Altman analysis, the CBCT method had the widest range of 95% limits of agreement, reflecting its greater potential of error. In measuring file lengths, RVG had a bigger window of 95% limits of agreement compared to CBCT. Conclusions: (1) The clinically insignificant underestimation of the preoperative working length using small FOV CBCT showed that it is acceptable for use in the estimation of preoperative working length. (2) Small FOV CBCT may be used in working length determination but it is not as accurate as the currently practiced method of using the EAL. (3) It is also more accurate than RVG in measuring file lengths.

Keywords: accuracy, CBCT, endodontics, measurement

Procedia PDF Downloads 304
19640 Performance Evaluation of the Classic seq2seq Model versus a Proposed Semi-supervised Long Short-Term Memory Autoencoder for Time Series Data Forecasting

Authors: Aswathi Thrivikraman, S. Advaith

Abstract:

The study is aimed at designing encoders for deciphering intricacies in time series data by redescribing the dynamics operating on a lower-dimensional manifold. A semi-supervised LSTM autoencoder is devised and investigated to see if the latent representation of the time series data can better forecast the data. End-to-end training of the LSTM autoencoder, together with another LSTM network that is connected to the latent space, forces the hidden states of the encoder to represent the most meaningful latent variables relevant for forecasting. Furthermore, the study compares the predictions with those of a traditional seq2seq model.

Keywords: LSTM, autoencoder, forecasting, seq2seq model

Procedia PDF Downloads 148
19639 Influence of Nanomaterials on the Properties of Shape Memory Polymeric Materials

Authors: Katielly Vianna Polkowski, Rodrigo Denizarte de Oliveira Polkowski, Cristiano Grings Herbert

Abstract:

The use of nanomaterials in the formulation of polymeric materials modifies their molecular structure, offering an infinite range of possibilities for the development of smart products, being of great importance for science and contemporary industry. Shape memory polymers are generally lightweight, have high shape recovery capabilities, they are easy to process and have properties that can be adapted for a variety of applications. Shape memory materials are active materials that have attracted attention due to their superior damping properties when compared to conventional structural materials. The development of methodologies capable of preparing new materials, which use graphene in their structure, represents technological innovation that transforms low-cost products into advanced materials with high added value. To obtain an improvement in the shape memory effect (SME) of polymeric materials, it is possible to use graphene in its composition containing low concentration by mass of graphene nanoplatelets (GNP), graphene oxide (GO) or other functionalized graphene, via different mixture process. As a result, there was an improvement in the SME, regarding the increase in the values of maximum strain. In addition, the use of graphene contributes to obtaining nanocomposites with superior electrical properties, greater crystallinity, as well as resistance to material degradation. The methodology used in the research is Systematic Review, scientific investigation, gathering relevant studies on influence of nanomaterials on the properties of shape memory polymeric, using the literature database as a source and study methods. In the present study, a systematic reviewwas performed of all papers published from 2014 to 2022 regarding graphene and shape memory polymeric througha search of three databases. This study allows for easy identification of themost relevant fields of study with respect to graphene and shape memory polymeric, as well as the main gaps to beexplored in the literature. The addition of graphene showed improvements in obtaining higher values of maximum deformation of the material, attributed to a possible slip between stacked or agglomerated nanostructures, as well as an increase in stiffness due to the increase in the degree of phase separation that results in a greater amount physical cross-links, referring to the formation of shortrange rigid domains.

Keywords: graphene, shape memory, smart materials, polymers, nanomaterials

Procedia PDF Downloads 79
19638 Emotion Regulation and Executive Functioning Scale for Children and Adolescents (REMEX): Scale Development

Authors: Cristina Costescu, Carmen David, Adrian Roșan

Abstract:

Executive functions (EF) and emotion regulation strategies are processes that allow individuals to function in an adaptative way and to be goal-oriented, which is essential for success in daily living activities, at school, or in social contexts. The Emotion Regulation and Executive Functioning Scale for Children and Adolescents (REMEX) represents an empirically based tool (based on the model of EF developed by Diamond) for evaluating significant dimensions of child and adolescent EFs and emotion regulation strategies, mainly in school contexts. The instrument measures the following dimensions: working memory, inhibition, cognitive flexibility, executive attention, planning, emotional control, and emotion regulation strategies. Building the instrument involved not only a top-down process, as we selected the content in accordance with prominent models of FE, but also a bottom-up one, as we were able to identify valid contexts in which FE and ER are put to use. For the construction of the instrument, we implemented three focus groups with teachers and other professionals since the aim was to develop an accurate, objective, and ecological instrument. We used the focus group method in order to address each dimension and to yield a bank of items to be further tested. Each dimension is addressed through a task that the examiner will apply and through several items derived from the main task. For the validation of the instrument, we plan to use item response theory (IRT), also known as the latent response theory, that attempts to explain the relationship between latent traits (unobservable cognitive processes) and their manifestations (i.e., observed outcomes, responses, or performance). REMEX represents an ecological scale that integrates a current scientific understanding of emotion regulation and EF and is directly applicable to school contexts, and it can be very useful for developing intervention protocols. We plan to test his convergent validity with the Childhood Executive Functioning Inventory (CHEXI) and Emotion Dysregulation Inventory (EDI) and divergent validity between a group of typically developing children and children with neurodevelopmental disorders, aged between 6 and 9 years old. In a previous pilot study, we enrolled a sample of 40 children with autism spectrum disorders and attention-deficit/hyperactivity disorder aged 6 to 12 years old, and we applied the above-mentioned scales (CHEXI and EDI). Our results showed that deficits in planning, bebavior regulation, inhibition, and working memory predict high levels of emotional reactivity, leading to emotional and behavioural problems. Considering previous results, we expect our findings to provide support for the validity and reliability of the REMEX version as an ecological instrument for assessing emotion regulation and EF in children and for key features of its uses in intervention protocols.

Keywords: executive functions, emotion regulation, children, item response theory, focus group

Procedia PDF Downloads 94
19637 Approximate-Based Estimation of Single Event Upset Effect on Statistic Random-Access Memory-Based Field-Programmable Gate Arrays

Authors: Mahsa Mousavi, Hamid Reza Pourshaghaghi, Mohammad Tahghighi, Henk Corporaal

Abstract:

Recently, Statistic Random-Access Memory-based (SRAM-based) Field-Programmable Gate Arrays (FPGAs) are widely used in aeronautics and space systems where high dependability is demanded and considered as a mandatory requirement. Since design’s circuit is stored in configuration memory in SRAM-based FPGAs; they are very sensitive to Single Event Upsets (SEUs). In addition, the adverse effects of SEUs on the electronics used in space are much higher than in the Earth. Thus, developing fault tolerant techniques play crucial roles for the use of SRAM-based FPGAs in space. However, fault tolerance techniques introduce additional penalties in system parameters, e.g., area, power, performance and design time. In this paper, an accurate estimation of configuration memory vulnerability to SEUs is proposed for approximate-tolerant applications. This vulnerability estimation is highly required for compromising between the overhead introduced by fault tolerance techniques and system robustness. In this paper, we study applications in which the exact final output value is not necessarily always a concern meaning that some of the SEU-induced changes in output values are negligible. We therefore define and propose Approximate-based Configuration Memory Vulnerability Factor (ACMVF) estimation to avoid overestimating configuration memory vulnerability to SEUs. In this paper, we assess the vulnerability of configuration memory by injecting SEUs in configuration memory bits and comparing the output values of a given circuit in presence of SEUs with expected correct output. In spite of conventional vulnerability factor calculation methods, which accounts any deviations from the expected value as failures, in our proposed method a threshold margin is considered depending on user-case applications. Given the proposed threshold margin in our model, a failure occurs only when the difference between the erroneous output value and the expected output value is more than this margin. The ACMVF is subsequently calculated by acquiring the ratio of failures with respect to the total number of SEU injections. In our paper, a test-bench for emulating SEUs and calculating ACMVF is implemented on Zynq-7000 FPGA platform. This system makes use of the Single Event Mitigation (SEM) IP core to inject SEUs into configuration memory bits of the target design implemented in Zynq-7000 FPGA. Experimental results for 32-bit adder show that, when 1% to 10% deviation from correct output is considered, the counted failures number is reduced 41% to 59% compared with the failures number counted by conventional vulnerability factor calculation. It means that estimation accuracy of the configuration memory vulnerability to SEUs is improved up to 58% in the case that 10% deviation is acceptable in output results. Note that less than 10% deviation in addition result is reasonably tolerable for many applications in approximate computing domain such as Convolutional Neural Network (CNN).

Keywords: fault tolerance, FPGA, single event upset, approximate computing

Procedia PDF Downloads 192
19636 Unified Structured Process for Health Analytics

Authors: Supunmali Ahangama, Danny Chiang Choon Poo

Abstract:

Health analytics (HA) is used in healthcare systems for effective decision-making, management, and planning of healthcare and related activities. However, user resistance, the unique position of medical data content, and structure (including heterogeneous and unstructured data) and impromptu HA projects have held up the progress in HA applications. Notably, the accuracy of outcomes depends on the skills and the domain knowledge of the data analyst working on the healthcare data. The success of HA depends on having a sound process model, effective project management and availability of supporting tools. Thus, to overcome these challenges through an effective process model, we propose an HA process model with features from the rational unified process (RUP) model and agile methodology.

Keywords: agile methodology, health analytics, unified process model, UML

Procedia PDF Downloads 500
19635 A Case Report on Cognitive-Communication Intervention in Traumatic Brain Injury

Authors: Nikitha Francis, Anjana Hoode, Vinitha George, Jayashree S. Bhat

Abstract:

The interaction between cognition and language, referred as cognitive-communication, is very intricate, involving several mental processes such as perception, memory, attention, lexical retrieval, decision making, motor planning, self-monitoring and knowledge. Cognitive-communication disorders are difficulties in communicative competencies that result from underlying cognitive impairments of attention, memory, organization, information processing, problem solving, and executive functions. Traumatic brain injury (TBI) is an acquired, non - progressive condition, resulting in distinct deficits of cognitive communication abilities such as naming, word-finding, self-monitoring, auditory recognition, attention, perception and memory. Cognitive-communication intervention in TBI is individualized, in order to enhance the person’s ability to process and interpret information for better functioning in their family and community life. The present case report illustrates the cognitive-communicative behaviors and the intervention outcomes of an adult with TBI, who was brought to the Department of Audiology and Speech Language Pathology, with cognitive and communicative disturbances, consequent to road traffic accident. On a detailed assessment, she showed naming deficits along with perseverations and had severe difficulty in recalling the details of the accident, her house address, places she had visited earlier, names of people known to her, as well as the activities she did each day, leading to severe breakdowns in her communicative abilities. She had difficulty in initiating, maintaining and following a conversation. She also lacked orientation to time and place. On administration of the Manipal Manual of Cognitive Linguistic Abilities (MMCLA), she exhibited poor performance on tasks related to visual and auditory perception, short term memory, working memory and executive functions. She attended 20 sessions of cognitive-communication intervention which followed a domain-general, adaptive training paradigm, with tasks relevant to everyday cognitive-communication skills. Compensatory strategies such as maintaining a dairy with reminders of her daily routine, names of people, date, time and place was also recommended. MMCLA was re-administered and her performance in the tasks showed significant improvements. Occurrence of perseverations and word retrieval difficulties reduced. She developed interests to initiate her day-to-day activities at home independently, as well as involve herself in conversations with her family members. Though she lacked awareness about her deficits, she actively involved herself in all the therapy activities. Rehabilitation of moderate to severe head injury patients can be done effectively through a holistic cognitive retraining with a focus on different cognitive-linguistic domains. Selection of goals and activities should have relevance to the functional needs of each individual with TBI, as highlighted in the present case report.

Keywords: cognitive-communication, executive functions, memory, traumatic brain injury

Procedia PDF Downloads 343
19634 Chronic Cognitive Impacts of Mild Traumatic Brain Injury during Aging

Authors: Camille Charlebois-Plante, Marie-Ève Bourassa, Gaelle Dumel, Meriem Sabir, Louis De Beaumont

Abstract:

To the extent of our knowledge, there has been little interest in the chronic effects of mild traumatic brain injury (mTBI) on cognition during normal aging. This is rather surprising considering the impacts on daily and social functioning. In addition, sustaining a mTBI during late adulthood may increase the effect of normal biological aging in individuals who consider themselves normal and healthy. The objective of this study was to characterize the persistent neuropsychological repercussions of mTBI sustained during late adulthood, on average 12 months prior to testing. To this end, 35 mTBI patients and 42 controls between the ages of 50 and 69 completed an exhaustive neuropsychological assessment lasting three hours. All mTBI patients were asymptomatic and all participants had a score ≥ 27 at the MoCA. The evaluation consisted of 20 standardized neuropsychological tests measuring memory, attention, executive and language functions, as well as information processing speed. Performance on tests of visual (Brief Visuospatial Memory Test Revised) and verbal memory (Rey Auditory Verbal Learning Test and WMS-IV Logical Memory subtest), lexical access (Boston Naming Test) and response inhibition (Stroop) revealed to be significantly lower in the mTBI group. These findings suggest that a mTBI sustained during late adulthood induces lasting effects on cognitive function. Episodic memory and executive functions seem to be particularly vulnerable to enduring mTBI effects.

Keywords: cognitive function, late adulthood, mild traumatic brain injury, neuropsychology

Procedia PDF Downloads 166
19633 Study on the Work-Life Balance of Selected Working Single Mothers in the Coastal Community of La Huerta, Paranaque

Authors: Idette Sheirina Biyo, Rhodora Lynn C. Lintag

Abstract:

This paper explores how the work-life balance of selected working single mothers situated in a coastal community is affecting their well-being. Working single mothers carry the responsibility of earning for their family while simultaneously exercising their motherhood. This study utilized a purposeful qualitative research through semi-structured interviews among ten working single mothers living in the coastal community of La Huerta, Parañaque in order to identify the following: a) experiences of the working single mothers, b) problems usually encountered, and c) how these problems are affecting their well-being. Dorothy Smith’s Feminist Standpoint theory is used as a theoretical lens in order to explain their work-life balance. Results have shown that despite their dual roles as the main income earners and heads of the households, they are not neglecting to care for their well-being. They consider getting sufficient rest, eating well, and going to church as forms of caring for their well-being. Other factors that affect their work-life balance include living arrangements, work hours, type of work, and income.

Keywords: coastal community, well-being, work-life balance, Working single mother

Procedia PDF Downloads 200
19632 A Comparative Analysis of Hyper-Parameters Using Neural Networks for E-Mail Spam Detection

Authors: Syed Mahbubuz Zaman, A. B. M. Abrar Haque, Mehedi Hassan Nayeem, Misbah Uddin Sagor

Abstract:

Everyday e-mails are being used by millions of people as an effective form of communication over the Internet. Although e-mails allow high-speed communication, there is a constant threat known as spam. Spam e-mail is often called junk e-mails which are unsolicited and sent in bulk. These unsolicited emails cause security concerns among internet users because they are being exposed to inappropriate content. There is no guaranteed way to stop spammers who use static filters as they are bypassed very easily. In this paper, a smart system is proposed that will be using neural networks to approach spam in a different way, and meanwhile, this will also detect the most relevant features that will help to design the spam filter. Also, a comparison of different parameters for different neural network models has been shown to determine which model works best within suitable parameters.

Keywords: long short-term memory, bidirectional long short-term memory, gated recurrent unit, natural language processing, natural language processing

Procedia PDF Downloads 198
19631 Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

Authors: Yehjune Heo

Abstract:

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Keywords: anti-spoofing, CNN, fingerprint recognition, loss function, optimizer

Procedia PDF Downloads 132
19630 The Predictive Utility of Subjective Cognitive Decline Using Item Level Data from the Everyday Cognition (ECog) Scales

Authors: J. Fox, J. Randhawa, M. Chan, L. Campbell, A. Weakely, D. J. Harvey, S. Tomaszewski Farias

Abstract:

Early identification of individuals at risk for conversion to dementia provides an opportunity for preventative treatment. Many older adults (30-60%) report specific subjective cognitive decline (SCD); however, previous research is inconsistent in terms of what types of complaints predict future cognitive decline. The purpose of this study is to identify which specific complaints from the Everyday Cognition Scales (ECog) scales, a measure of self-reported concerns for everyday abilities across six cognitive domains, are associated with: 1) conversion from a clinical diagnosis of normal to either MCI or dementia (categorical variable) and 2) progressive cognitive decline in memory and executive function (continuous variables). 415 cognitively normal older adults were monitored annually for an average of 5 years. Cox proportional hazards models were used to assess associations between self-reported ECog items and progression to impairment (MCI or dementia). A total of 114 individuals progressed to impairment; the mean time to progression was 4.9 years (SD=3.4 years, range=0.8-13.8). Follow-up models were run controlling for depression. A subset of individuals (n=352) underwent repeat cognitive assessments for an average of 5.3 years. For those individuals, mixed effects models with random intercepts and slopes were used to assess associations between ECog items and change in neuropsychological measures of episodic memory or executive function. Prior to controlling for depression, subjective concerns on five of the eight Everyday Memory items, three of the nine Everyday Language items, one of the seven Everyday Visuospatial items, two of the five Everyday Planning items, and one of the six Everyday Organization items were associated with subsequent diagnostic conversion (HR=1.25 to 1.59, p=0.003 to 0.03). However, after controlling for depression, only two specific complaints of remembering appointments, meetings, and engagements and understanding spoken directions and instructions were associated with subsequent diagnostic conversion. Episodic memory in individuals reporting no concern on ECog items did not significantly change over time (p>0.4). More complaints on seven of the eight Everyday Memory items, three of the nine Everyday Language items, and three of the seven Everyday Visuospatial items were associated with a decline in episodic memory (Interaction estimate=-0.055 to 0.001, p=0.003 to 0.04). Executive function in those reporting no concern on ECog items declined slightly (p <0.001 to 0.06). More complaints on three of the eight Everyday Memory items and three of the nine Everyday Language items were associated with a decline in executive function (Interaction estimate=-0.021 to -0.012, p=0.002 to 0.04). These findings suggest that specific complaints across several cognitive domains are associated with diagnostic conversion. Specific complaints in the domains of Everyday Memory and Language are associated with a decline in both episodic memory and executive function. Increased monitoring and treatment of individuals with these specific SCD may be warranted.

Keywords: alzheimer’s disease, dementia, memory complaints, mild cognitive impairment, risk factors, subjective cognitive decline

Procedia PDF Downloads 75
19629 Shadows and Symbols: The Tri-Level Importance of Memory in Jane Yolen's 'the Devil's Arithmetic' and Soon-To-Be-Published 'Mapping the Bones'

Authors: Kirsten A. Bartels

Abstract:

'Never again' and 'Lest we forget' have long been messages associated with the events of the Shoah. Yet as we attempt to learn from the past, we must find new ways to engage with its memories. The preservation of the culture and the value of tradition are critical factors in Jane Yolen's works of Holocaust fiction, The Devil's Arithmetic and Mapping the Bones, emphasized through the importance of remembering. That word, in its multitude of forms (remember, remembering, memories), occurs no less than ten times in the first four pages and over one hundred times in the one hundred and sixty-four-page narrative The Devil’s Arithmetic. While Yolen takes a different approach to showcasing the importance of memory in Mapping the Bones, it is of equal import in this work and arguably to the future of Holocaust knowing. The idea of remembering, the desire to remember, and the ability to remember, are explored in three divergent ways in The Devil’s Arithmetic. First, in the importance to remember a past which is not her own – to understand history or acquired memories. Second, in the protagonist's actual or initial memories, those of her life in modern-day New York. Third, in a reverse mode of forgetting and trying to reacquire that which has been lost -- as Hannah is processed in the camp and she forgets everything, all worlds prior to the camp are lost to her. As numbers replace names, Yolen stresses the importance of self-identity or owned memories. In addition, the importance of relaying memory, the transitions of memory from perspective, and the ideas of reflective telling are explored in Mapping the Bones -- through the telling of the story through the lens of one of the twins as the events are unfolding; and then the through the reflective telling from the lens of the other twin. Parallel to the exploration of the intersemiosis of memory is the discussion of literary shadows (foreshadowing, backshadowing, and side-shadowing) and their impact on the reader's experience with Yolen's narrative. For in this type of exploration, one cannot look at the events described in Yolen's work and not also contemplate the figurative shadows cast.

Keywords: holocaust literature, memory, narrative, Yolen

Procedia PDF Downloads 232
19628 Prediction on Housing Price Based on Deep Learning

Authors: Li Yu, Chenlu Jiao, Hongrun Xin, Yan Wang, Kaiyang Wang

Abstract:

In order to study the impact of various factors on the housing price, we propose to build different prediction models based on deep learning to determine the existing data of the real estate in order to more accurately predict the housing price or its changing trend in the future. Considering that the factors which affect the housing price vary widely, the proposed prediction models include two categories. The first one is based on multiple characteristic factors of the real estate. We built Convolution Neural Network (CNN) prediction model and Long Short-Term Memory (LSTM) neural network prediction model based on deep learning, and logical regression model was implemented to make a comparison between these three models. Another prediction model is time series model. Based on deep learning, we proposed an LSTM-1 model purely regard to time series, then implementing and comparing the LSTM model and the Auto-Regressive and Moving Average (ARMA) model. In this paper, comprehensive study of the second-hand housing price in Beijing has been conducted from three aspects: crawling and analyzing, housing price predicting, and the result comparing. Ultimately the best model program was produced, which is of great significance to evaluation and prediction of the housing price in the real estate industry.

Keywords: deep learning, convolutional neural network, LSTM, housing prediction

Procedia PDF Downloads 298
19627 Digital Memory in Motion: (Re) Creating and (Re) Posting of “Gaja-gamini walk” Reels as a Collective Feminist Practices on Instagram

Authors: Gazal Khan

Abstract:

This paper investigates the phenomenon of (re) creating and (re) posting of what is popularly known as "gaja-gamini walk" on instagram as a form of digital feminism, examining how these reels (short videos) make meaning in digital spaces. The study analyzes xyz “gaja- gamini walk” reels created by Indian influencers and instagram users, employing qualitative textual analysis, close readings, and digital ethnography to analyze the interplay between media, memory and digital spaces. The research highlights how “gaja-gamini walk” reels, characterized by an assertive presentation, redefines female body aesthetics, re (orients) sexual gaze to provide layered, interwoven and contested narratives. These reels facilitate a unique form of engagement by allowing users to re-share and participate in feminist discourse and allowing reels to function as sites of memory. The paper also discusses the social dynamics of these reels, their intertextuality with cultural narratives, and the limitations of the format for sustained feminist action. Through this analysis, the paper contributes to understanding the role of digital memory in contemporary feminist movements in context of Indian feminism.

Keywords: instagram, gaja-gamni walk, female gaze, digital feminism

Procedia PDF Downloads 26
19626 Lattice Twinning and Detwinning Processes in Phase Transformation in Shape Memory Alloys

Authors: Osman Adiguzel

Abstract:

Shape memory effect is a peculiar property exhibited by certain alloy systems and based on martensitic transformation, and shape memory properties are closely related to the microstructures of the material. Shape memory effect is linked with martensitic transformation, which is a solid state phase transformation and occurs with the cooperative movement of atoms by means of lattice invariant shears on cooling from high-temperature parent phase. Lattice twinning and detwinning can be considered as elementary processes activated during the transformation. Thermally induced martensite occurs as martensite variants, in self-accommodating manner and consists of lattice twins. Also, this martensite is called the twinned martensite or multivariant martensite. Deformation of shape memory alloys in martensitic state proceeds through a martensite variant reorientation. The martensite variants turn into the reoriented single variants with deformation, and the reorientation process has great importance for the shape memory behavior. Copper based alloys exhibit this property in metastable β- phase region, which has DO3 –type ordered lattice in ternary case at high temperature, and these structures martensiticaly turn into the layered complex structures with lattice twinning mechanism, on cooling from high temperature parent phase region. The twinning occurs as martensite variants with lattice invariant shears in two opposite directions, <110 > -type directions on the {110}- type plane of austenite matrix. Lattice invariant shear is not uniform in copper based ternary alloys and gives rise to the formation of unusual layered structures, like 3R, 9R, or 18R depending on the stacking sequences on the close-packed planes of the ordered lattice. The unit cell and periodicity are completed through 18 atomic layers in case of 18R-structure. On the other hand, the deformed material recovers the original shape on heating above the austenite finish temperature. Meanwhile, the material returns to the twinned martensite structures (thermally induced martensite structure) in one way (irreversible) shape memory effect on cooling below the martensite finish temperature, whereas the material returns to the detwinned martensite structure (deformed martensite) in two-way (reversible) shape memory effect. Shortly one can say that the microstructural mechanisms, responsible for the shape memory effect are the twinning and detwinning processes as well as martensitic transformation. In the present contribution, x-ray diffraction, transmission electron microscopy (TEM) and differential scanning calorimetry (DSC) studies were carried out on two copper-based ternary alloys, CuZnAl, and CuAlMn.

Keywords: shape memory effect, martensitic transformation, twinning and detwinning, layered structures

Procedia PDF Downloads 426
19625 Improving Subjective Bias Detection Using Bidirectional Encoder Representations from Transformers and Bidirectional Long Short-Term Memory

Authors: Ebipatei Victoria Tunyan, T. A. Cao, Cheol Young Ock

Abstract:

Detecting subjectively biased statements is a vital task. This is because this kind of bias, when present in the text or other forms of information dissemination media such as news, social media, scientific texts, and encyclopedias, can weaken trust in the information and stir conflicts amongst consumers. Subjective bias detection is also critical for many Natural Language Processing (NLP) tasks like sentiment analysis, opinion identification, and bias neutralization. Having a system that can adequately detect subjectivity in text will boost research in the above-mentioned areas significantly. It can also come in handy for platforms like Wikipedia, where the use of neutral language is of importance. The goal of this work is to identify the subjectively biased language in text on a sentence level. With machine learning, we can solve complex AI problems, making it a good fit for the problem of subjective bias detection. A key step in this approach is to train a classifier based on BERT (Bidirectional Encoder Representations from Transformers) as upstream model. BERT by itself can be used as a classifier; however, in this study, we use BERT as data preprocessor as well as an embedding generator for a Bi-LSTM (Bidirectional Long Short-Term Memory) network incorporated with attention mechanism. This approach produces a deeper and better classifier. We evaluate the effectiveness of our model using the Wiki Neutrality Corpus (WNC), which was compiled from Wikipedia edits that removed various biased instances from sentences as a benchmark dataset, with which we also compare our model to existing approaches. Experimental analysis indicates an improved performance, as our model achieved state-of-the-art accuracy in detecting subjective bias. This study focuses on the English language, but the model can be fine-tuned to accommodate other languages.

Keywords: subjective bias detection, machine learning, BERT–BiLSTM–Attention, text classification, natural language processing

Procedia PDF Downloads 121