Search results for: learning curve
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8257

Search results for: learning curve

3757 Characterization of Printed Reflectarray Elements on Variable Substrate Thicknesses

Authors: M. Y. Ismail, Arslan Kiyani

Abstract:

Narrow bandwidth and high loss performance limits the use of reflectarray antennas in some applications. This article reports on the feasibility of employing strategic reflectarray resonant elements to characterize the reflectivity performance of reflectarrays in X-band frequency range. Strategic reflectarray resonant elements incorporating variable substrate thicknesses ranging from 0.016λ to 0.052λ have been analyzed in terms of reflection loss and reflection phase performance. The effect of substrate thickness has been validated by using waveguide scattering parameter technique. It has been demonstrated that as the substrate thickness is increased from 0.508mm to 1.57mm the measured reflection loss of dipole element decreased from 5.66dB to 3.70dB with increment in 10% bandwidth of 39MHz to 64MHz. Similarly the measured reflection loss of triangular loop element is decreased from 20.25dB to 7.02dB with an increment in 10% bandwidth of 12MHz to 23MHz. The results also show a significant decrease in the slope of reflection phase curve as well. A Figure of Merit (FoM) has also been defined for the comparison of static phase range of resonant elements under consideration. Moreover, a novel numerical model based on analytical equations has been established incorporating the material properties of dielectric substrate and electrical properties of different reflectarray resonant elements to obtain the progressive phase distribution for each individual reflectarray resonant element.

Keywords: numerical model, reflectarray resonant elements, scattering parameter measurements, variable substrate thickness

Procedia PDF Downloads 280
3756 Automation of AAA Game Development using AI and Procedural Generation

Authors: Paul Toprac, Branden Heng, Harsheni Siddharthan, Allison Tseng, Sarah Abraham, Etienne Vouga

Abstract:

The goal of this project was to evaluate and document the capabilities and limitations of AI tools for empowering small teams to create high budget, high profile (AAA) 3D games typically developed by large studios. Two teams of novice game developers attempted to create two different games using AI and Unreal Engine 5.3. First, the teams evaluated 60 AI art, design, sound, and programming tools by considering their capability, ease of use, cost, and license restrictions. Then, the teams used a shortlist of 13 AI tools for game development. During this process, the following tools were found to be the most productive: (1) ChatGPT 4.0 for both game and narrative concepting and documentation; (2) Dall-E 3 and OpenArt for concept art; (3) Beatoven for music drafting; (4) Epic PCG for level design; and (5) ChatGPT 4.0 and Github Copilot for generating simple code and to complement human-made tutorials as an additional learning resource. While current generative AI may appear impressive at first glance, the assets they produce fall short of AAA industry standards. Generative AI tools are helpful when brainstorming ideas such as concept art and basic storylines, but they still cannot replace human input or creativity at this time. Regarding programming, AI can only effectively generate simple code and act as an additional learning resource. Thus, generative AI tools are at best tools to enhance developer productivity rather than as a system to replace developers.

Keywords: AAA games, AI, automation tools, game development

Procedia PDF Downloads 33
3755 Using Machine Learning to Extract Patient Data from Non-standardized Sports Medicine Physician Notes

Authors: Thomas Q. Pan, Anika Basu, Chamith S. Rajapakse

Abstract:

Machine learning requires data that is categorized into features that models train on. This topic is important to the field of sports medicine due to the many tools it provides to physicians such as diagnosis support and risk assessment. Physician note that healthcare professionals take are usually unclean and not suitable for model training. The objective of this study was to develop and evaluate an advanced approach for extracting key features from sports medicine data without the need for extensive model training or data labeling. An LLM (Large Language Model) was given a narrative (Physician’s Notes) and prompted to extract four features (details about the patient). The narrative was found in a datasheet that contained six columns: Case Number, Validation Age, Validation Gender, Validation Diagnosis, Validation Body Part, and Narrative. The validation columns represent the accurate responses that the LLM attempts to output. With the given narrative, the LLM would output its response and extract the age, gender, diagnosis, and injured body part with each category taking up one line. The output would then be cleaned, matched, and added to new columns containing the extracted responses. Five ways of checking the accuracy were used: unclear count, substring comparison, LLM comparison, LLM re-check, and hand-evaluation. The unclear count essentially represented the extractions the LLM missed. This can be also understood as the recall score ([total - false negatives] over total). The rest of these correspond to the precision score ([total - false positives] over total). Substring comparison evaluated the validation (X) and extracted (Y) columns’ likeness by checking if X’s results were a substring of Y's findings and vice versa. LLM comparison directly asked an LLM if the X and Y’s results were similar. LLM Re-check prompted the LLM to see if the extracted results can be found in the narrative. Lastly, A selection of 1,000 random narratives was also selected and hand-evaluated to give an estimate of how well the LLM-based feature extraction model performed. With a selection of 10,000 narratives, the LLM-based approach had a recall score of roughly 98%. However, the precision scores of the substring comparison and LLM comparison models were around 72% and 76% respectively. The reason for these low figures is due to the minute differences between answers. For example, the ‘chest’ is a part of the ‘upper trunk’ however, these models cannot detect that. On the other hand, the LLM re-check and subset of hand-tested narratives showed a precision score of 96% and 95%. If this subset is used to extrapolate the possible outcome of the whole 10,000 narratives, the LLM-based approach would be strong in both precision and recall. These results indicated that an LLM-based feature extraction model could be a useful way for medical data in sports to be collected and analyzed by machine learning models. Wide use of this method could potentially increase the availability of data thus improving machine learning algorithms and supporting doctors with more enhanced tools.

Keywords: AI, LLM, ML, sports

Procedia PDF Downloads 18
3754 Identifying Autism Spectrum Disorder Using Optimization-Based Clustering

Authors: Sharifah Mousli, Sona Taheri, Jiayuan He

Abstract:

Autism spectrum disorder (ASD) is a complex developmental condition involving persistent difficulties with social communication, restricted interests, and repetitive behavior. The challenges associated with ASD can interfere with an affected individual’s ability to function in social, academic, and employment settings. Although there is no effective medication known to treat ASD, to our best knowledge, early intervention can significantly improve an affected individual’s overall development. Hence, an accurate diagnosis of ASD at an early phase is essential. The use of machine learning approaches improves and speeds up the diagnosis of ASD. In this paper, we focus on the application of unsupervised clustering methods in ASD as a large volume of ASD data generated through hospitals, therapy centers, and mobile applications has no pre-existing labels. We conduct a comparative analysis using seven clustering approaches such as K-means, agglomerative hierarchical, model-based, fuzzy-C-means, affinity propagation, self organizing maps, linear vector quantisation – as well as the recently developed optimization-based clustering (COMSEP-Clust) approach. We evaluate the performances of the clustering methods extensively on real-world ASD datasets encompassing different age groups: toddlers, children, adolescents, and adults. Our experimental results suggest that the COMSEP-Clust approach outperforms the other seven methods in recognizing ASD with well-separated clusters.

Keywords: autism spectrum disorder, clustering, optimization, unsupervised machine learning

Procedia PDF Downloads 121
3753 Osteoactivin Is a Specific Biomarker in Bone and Cartilage Metabolism

Authors: Gulnara Azizova, Naila Hasanova, Nazenin Hasanzade

Abstract:

The aim of study is to investigate the role of osteoactivin as a more sensitive and modern diagnostic biomarker that has a prognostic value in metabolic and repair processes occurring in bone and cartilage tissue in osteoporosis and osteoporotic fractures. Osteoactivin (OA) is a new glycoprotein that is highly expressed during osteoblast differentiation. It was first discovered in an osteopetrotic rat model using mRNA . This study was carried out on patients between the ages of 45-83 from the Department of Traumatology and placed in 3 groups: group I - 14 patients with osteoporosis, group II - 15 patients with non-osteoporotic fractures, group III - 25 patients with osteoporotic fractures. The control group consisted of 14 healthy people. To monitor changes in osteoactivin, blood samples were taken at 3 stages: on day 1 before treatment, on day 10 of treatment, and 1 month after treatment. The concentration of OA in the blood serum was determined by ELISA method on the immunoassay analyzer “Mindray MR- 96A” using a set of reagents from the company Boster ( ELISA Kit PicoKine, USA). The statistical evaluation was performed by using SPSS 22.0 program (IBM SPSS Inc., USA). Compared to the control, osteoactivin concentration increased by 66.2% in patients with osteoporosis, 54.1% in patients with non-osteoporotic fractures, and 80.2% in patients with osteoporotic fractures, indicating that it plays an important role in the pathogenesis of osteoporotic fractures. At 1 month after treatment, osteoactivin concentration increased by 81.6% in patients with non-osteoporotic fractures. The lack of a significant change in osteoporotic fractures is explained by the late healing of these fractures. Based on the sensitivity and specificity indicators, the ROC curve was created and it was determined that osteoactivin is a test with high general diagnostic value, specificity and informativeness in the prognosis of osteoporosis and osteoporotic fractures, and can be used throughout the treatment period.

Keywords: osteoactivin, bone, osteoporosis., cartilage

Procedia PDF Downloads 35
3752 Leveraging SHAP Values for Effective Feature Selection in Peptide Identification

Authors: Sharon Li, Zhonghang Xia

Abstract:

Post-database search is an essential phase in peptide identification using tandem mass spectrometry (MS/MS) to refine peptide-spectrum matches (PSMs) produced by database search engines. These engines frequently face difficulty differentiating between correct and incorrect peptide assignments. Despite advances in statistical and machine learning methods aimed at improving the accuracy of peptide identification, challenges remain in selecting critical features for these models. In this study, two machine learning models—a random forest tree and a support vector machine—were applied to three datasets to enhance PSMs. SHAP values were utilized to determine the significance of each feature within the models. The experimental results indicate that the random forest model consistently outperformed the SVM across all datasets. Further analysis of SHAP values revealed that the importance of features varies depending on the dataset, indicating that a feature's role in model predictions can differ significantly. This variability in feature selection can lead to substantial differences in model performance, with false discovery rate (FDR) differences exceeding 50% between different feature combinations. Through SHAP value analysis, the most effective feature combinations were identified, significantly enhancing model performance.

Keywords: peptide identification, SHAP value, feature selection, random forest tree, support vector machine

Procedia PDF Downloads 33
3751 Exploring the Association between Personality Traits and Adolescent Wellbeing in Online Education: A Systematic Review

Authors: Rashmi Motwani, Ritu Raj

Abstract:

The emergence of online educational environments has changed the way adolescents learn, which has benefits and drawbacks for their development. This review has as its goal the examination of how personality traits and adolescents’ well-being are associated in the setting of online education. This review analyses the effects of a variety of personality traits on the mental, emotional, and social health of online school-going adolescents by looking at a wide range of previous research. This research explores the mechanisms that mediate or regulate the connection between one's personality traits and well-being in an online educational environment. The elements can be broken down into two categories: technological, like internet availability and digital literacy, and social, including social support, peer interaction, and teacher-student connections. To improve the well-being of adolescents in online learning environments, it is essential to understand factors that moderate the effects of interventions and support systems. This review concludes by emphasising the complex nature of the association between individual differences in personality and the success of online students aged 13 to 18. This review contributes to the development of evidence-based strategies for promoting positive mental health and overall well-being among adolescents engaged in online educational settings by shedding light on the impact of personality traits on various dimensions of well-being and by identifying the mediating or moderating factors. Educators, governments, and parents can use the findings of this review to create an online learning environment that is safe and well-being for adolescents.

Keywords: personality traits, adolescent, wellbeing, online education

Procedia PDF Downloads 54
3750 Increase in Specificity of MicroRNA Detection by RT-qPCR Assay Using a Specific Extension Sequence

Authors: Kyung Jin Kim, Jiwon Kwak, Jae-Hoon Lee, Soo Suk Lee

Abstract:

We describe an innovative method for highly specific detection of miRNAs using a specially modified method of poly(A) adaptor RT-qPCR. We use uniquely designed specific extension sequence, which plays important role in providing an opportunity to affect high specificity of miRNA detection. This method involves two steps of reactions as like previously reported and which are poly(A) tailing and reverse-transcription followed by real-time PCR. Firstly, miRNAs are extended by a poly(A) tailing reaction and then converted into cDNA. Here, we remarkably reduced the reaction time by the application of short length of poly(T) adaptor. Next, cDNA is hybridized to the 3’-end of a specific extension sequence which contains miRNA sequence and results in producing a novel PCR template. Thereafter, the SYBR Green-based RT-qPCR progresses with a universal poly(T) adaptor forward primer and a universal reverse primer. The target miRNA, miR-106b in human brain total RNA, could be detected quantitatively in the range of seven orders of magnitude, which demonstrate that the assay displays a dynamic range of at least 7 logs. In addition, the better specificity of this novel extension-based assay against well known poly(A) tailing method for miRNA detection was confirmed by melt curve analysis of real-time PCR product, clear gel electrophoresis and sequence chromatogram images of amplified DNAs.

Keywords: microRNA(miRNA), specific extension sequence, RT-qPCR, poly(A) tailing assay, reverse transcription

Procedia PDF Downloads 310
3749 Soil Penetration Resistance and Water Content Spatial Distribution Following Different Tillage and Crop Rotation in a Chinese Mollisol

Authors: Xuewen Chen, Aizhen Liang, Xiaoping Zhang

Abstract:

To better understand the spatial variability of soil penetration resistance (SPR) and soil water content (SWC) induced by different tillage and crop rotation in a Mollisol of Northeast China, the soil was sampled from the tillage experiment which was established in Dehui County, Jilin Province, Northeast China, in 2001. Effect of no-tillage (NT), moldboard plow (MP) and ridge tillage (RT) under corn-soybean rotation (C-S) and continuous corn (C-C) system on SPR and SWC were compared with horizontal and vertical variations. The results showed that SPR and SWC spatially varied across the ridge. SPR in the rows was higher than inter-rows, especially in topsoil (2.5-15 cm) of NT and RT plots. SPR of MP changed in the trend with the curve-shaped ridge. In contrast to MP, NT, and RT resulted in average increment of 166.3% and 152.3% at a depth of 2.5-17.5 cm in the row positions, respectively. The mean SPR in topsoil in the rows means soil compaction is not the main factor limiting plant growth and crop yield. SPR in the row of RT soil was lower than NT at a depth of 2.5-12.5 cm. The SWC in NT and RT soil was highest in the inter-rows and least in the rows or shoulders, respectively. However, the lateral variation trend of MP was opposite to NT. From the profile view of SWC, MP was greater than NT and RT in 0-20 cm of the rows. SWC in RT soil was higher than NT in the row of 0-20 cm. Crop rotation did not have a marked impact on SPR and SWC. In addition to the tillage practices, the factor which affects SPR greatly was depth but not position. These two factors have significant effects on SWC. These results indicated that the adoption of RT was a more suitable conservation tillage practices than NT in the black soil of Northeast China.

Keywords: row, soil penetration resistance, spatial variability, tillage practice

Procedia PDF Downloads 140
3748 Effectiveness of Using Phonemic Awareness Based Activities in Improving Decoding Skills of Third Grade Students Referred for Reading Disabilities in Oman

Authors: Mahmoud Mohamed Emam

Abstract:

In Oman the number of students referred for reading disabilities is on the rise. Schools serve these students by placement in the so-called learning disabilities unit. Recently the author led a strategic project to train teachers on the use of curriculum based measurement to identify students with reading disabilities in Oman. Additional the project involved training teachers to use phonemic awareness based activities to improve reading skills of those students. Phonemic awareness refers to the ability to notice, think about, and work with the individual sounds in words. We know that a student's skill in phonemic awareness is a good predictor of later reading success or difficulty. Using multiple baseline design across four participants the current studies investigated the effectiveness of using phonemic awareness based activities to improve decoding skills of third grade students referred for reading disabilities in Oman. During treatment students received phonemic awareness based activities that were designed to fulfill the idiosyncratic characteristics of Arabic language phonology as well as orthography. Results indicated that the phonemic awareness based activities were effective in substantially increasing the number of correctly decoded word for all four participants. Maintenance of strategy effects was evident for the weeks following the termination of intervention for the four students. In addition, the effects of intervention generalized to decoding novel words for all four participants.

Keywords: learning disabilities, phonemic awareness, third graders, Oman

Procedia PDF Downloads 645
3747 Charting Sentiments with Naive Bayes and Logistic Regression

Authors: Jummalla Aashrith, N. L. Shiva Sai, K. Bhavya Sri

Abstract:

The swift progress of web technology has not only amassed a vast reservoir of internet data but also triggered a substantial surge in data generation. The internet has metamorphosed into one of the dynamic hubs for online education, idea dissemination, as well as opinion-sharing. Notably, the widely utilized social networking platform Twitter is experiencing considerable expansion, providing users with the ability to share viewpoints, participate in discussions spanning diverse communities, and broadcast messages on a global scale. The upswing in online engagement has sparked a significant curiosity in subjective analysis, particularly when it comes to Twitter data. This research is committed to delving into sentiment analysis, focusing specifically on the realm of Twitter. It aims to offer valuable insights into deciphering information within tweets, where opinions manifest in a highly unstructured and diverse manner, spanning a spectrum from positivity to negativity, occasionally punctuated by neutrality expressions. Within this document, we offer a comprehensive exploration and comparative assessment of modern approaches to opinion mining. Employing a range of machine learning algorithms such as Naive Bayes and Logistic Regression, our investigation plunges into the domain of Twitter data streams. We delve into overarching challenges and applications inherent in the realm of subjectivity analysis over Twitter.

Keywords: machine learning, sentiment analysis, visualisation, python

Procedia PDF Downloads 59
3746 Protection of Human Rights in Polish Centres for Foreigners – in the Context of the European Human Rights System

Authors: Oktawia Braniewicz

Abstract:

The phenomenon of emigration and migration increasingly affects Poland's borders as well. For this reason, it is necessary to examine the level of protection of Human Rights in Polish Centres for Foreigners. The field study covered 11 centers for Foreigners in the provinces Kujawsko-Pomorskie Region, Lubelskie Region, Lodzkie Region, Mazowieckie Region and Podlaskie Region. Photographic documentation of living and social conditions, conversations with center employees and refugees allow to show a comprehensive picture of the situation prevailing in Centres for Foreigners. The object of reflection will be, in particular, the standards resulting from art. 8 and 13 of the Convention for the Protection of Human Rights and Fundamental Freedoms and article 2 of Protocol No. 1 to the Convention for the Protection of Human Rights and Fundamental Freedoms. The degree of realization of the right to education and the right to respect for family and private life will be shown. Issues related to learning the Polish language, access to a professional translator and psychological help will also be approximated. Learning Polish is not obligatory, which causes problems with assimilation and integration with other members of the new community. In centers for foreigners, there are no translators - a translator from an external company is rented if necessary. The waiting time for an interpreter makes the refugees feel anxious, unable to communicate with the employees of the centers (this is a situation in which the refugees do not know either English, Polish or Russian). Psychologist's help is available on designated days of the week. There is no separate specialist in child psychology, which is a serious problem.

Keywords: human rights, Polish centres, foreigners, fundamental freedoms

Procedia PDF Downloads 136
3745 Video Compression Using Contourlet Transform

Authors: Delara Kazempour, Mashallah Abasi Dezfuli, Reza Javidan

Abstract:

Video compression used for channels with limited bandwidth and storage devices has limited storage capabilities. One of the most popular approaches in video compression is the usage of different transforms. Discrete cosine transform is one of the video compression methods that have some problems such as blocking, noising and high distortion inappropriate effect in compression ratio. wavelet transform is another approach is better than cosine transforms in balancing of compression and quality but the recognizing of curve curvature is so limit. Because of the importance of the compression and problems of the cosine and wavelet transforms, the contourlet transform is most popular in video compression. In the new proposed method, we used contourlet transform in video image compression. Contourlet transform can save details of the image better than the previous transforms because this transform is multi-scale and oriented. This transform can recognize discontinuity such as edges. In this approach we lost data less than previous approaches. Contourlet transform finds discrete space structure. This transform is useful for represented of two dimension smooth images. This transform, produces compressed images with high compression ratio along with texture and edge preservation. Finally, the results show that the majority of the images, the parameters of the mean square error and maximum signal-to-noise ratio of the new method based contourlet transform compared to wavelet transform are improved but in most of the images, the parameters of the mean square error and maximum signal-to-noise ratio in the cosine transform is better than the method based on contourlet transform.

Keywords: video compression, contourlet transform, discrete cosine transform, wavelet transform

Procedia PDF Downloads 449
3744 Application to Monitor the Citizens for Corona and Get Medical Aids or Assistance from Hospitals

Authors: Vathsala Kaluarachchi, Oshani Wimalarathna, Charith Vandebona, Gayani Chandrarathna, Lakmal Rupasinghe, Windhya Rankothge

Abstract:

It is the fundamental function of a monitoring system to allow users to collect and process data. A worldwide threat, the corona outbreak has wreaked havoc in Sri Lanka, and the situation has gotten out of hand. Since the epidemic, the Sri Lankan government has been unable to establish a systematic system for monitoring corona patients and providing emergency care in the event of an outbreak. Most patients have been held at home because of the high number of patients reported in the nation, but they do not yet have access to a functioning medical system. It has resulted in an increase in the number of patients who have been left untreated because of a lack of medical care. The absence of competent medical monitoring is the biggest cause of mortality for many people nowadays, according to our survey. As a result, a smartphone app for analyzing the patient's state and determining whether they should be hospitalized will be developed. Using the data supplied, we are aiming to send an alarm letter or SMS to the hospital once the system recognizes them. Since we know what those patients need and when they need it, we will put up a desktop program at the hospital to monitor their progress. Deep learning, image processing and application development, natural language processing, and blockchain management are some of the components of the research solution. The purpose of this research paper is to introduce a mechanism to connect hospitals and patients even when they are physically apart. Further data security and user-friendliness are enhanced through blockchain and NLP.

Keywords: blockchain, deep learning, NLP, monitoring system

Procedia PDF Downloads 137
3743 Improving Topic Quality of Scripts by Using Scene Similarity Based Word Co-Occurrence

Authors: Yunseok Noh, Chang-Uk Kwak, Sun-Joong Kim, Seong-Bae Park

Abstract:

Scripts are one of the basic text resources to understand broadcasting contents. Since broadcast media wields lots of influence over the public, tools for understanding broadcasting contents are more required. Topic modeling is the method to get the summary of the broadcasting contents from its scripts. Generally, scripts represent contents descriptively with directions and speeches. Scripts also provide scene segments that can be seen as semantic units. Therefore, a script can be topic modeled by treating a scene segment as a document. Because scripts consist of speeches mainly, however, relatively small co-occurrences among words in the scene segments are observed. This causes inevitably the bad quality of topics based on statistical learning method. To tackle this problem, we propose a method of learning with additional word co-occurrence information obtained using scene similarities. The main idea of improving topic quality is that the information that two or more texts are topically related can be useful to learn high quality of topics. In addition, by using high quality of topics, we can get information more accurate whether two texts are related or not. In this paper, we regard two scene segments are related if their topical similarity is high enough. We also consider that words are co-occurred if they are in topically related scene segments together. In the experiments, we showed the proposed method generates a higher quality of topics from Korean drama scripts than the baselines.

Keywords: broadcasting contents, scripts, text similarity, topic model

Procedia PDF Downloads 324
3742 A Complex Network Approach to Structural Inequality of Educational Deprivation

Authors: Harvey Sanchez-Restrepo, Jorge Louca

Abstract:

Equity and education are major focus of government policies around the world due to its relevance for addressing the sustainable development goals launched by Unesco. In this research, we developed a primary analysis of a data set of more than one hundred educational and non-educational factors associated with learning, coming from a census-based large-scale assessment carried on in Ecuador for 1.038.328 students, their families, teachers, and school directors, throughout 2014-2018. Each participating student was assessed by a standardized computer-based test. Learning outcomes were calibrated through item response theory with two-parameters logistic model for getting raw scores that were re-scaled and synthetized by a learning index (LI). Our objective was to develop a network for modelling educational deprivation and analyze the structure of inequality gaps, as well as their relationship with socioeconomic status, school financing, and student's ethnicity. Results from the model show that 348 270 students did not develop the minimum skills (prevalence rate=0.215) and that Afro-Ecuadorian, Montuvios and Indigenous students exhibited the highest prevalence with 0.312, 0.278 and 0.226, respectively. Regarding the socioeconomic status of students (SES), modularity class shows clearly that the system is out of equilibrium: the first decile (the poorest) exhibits a prevalence rate of 0.386 while rate for decile ten (the richest) is 0.080, showing an intense negative relationship between learning and SES given by R= –0.58 (p < 0.001). Another interesting and unexpected result is the average-weighted degree (426.9) for both private and public schools attending Afro-Ecuadorian students, groups that got the highest PageRank (0.426) and pointing out that they suffer the highest educational deprivation due to discrimination, even belonging to the richest decile. The model also found the factors which explain deprivation through the highest PageRank and the greatest degree of connectivity for the first decile, they are: financial bonus for attending school, computer access, internet access, number of children, living with at least one parent, books access, read books, phone access, time for homework, teachers arriving late, paid work, positive expectations about schooling, and mother education. These results provide very accurate and clear knowledge about the variables affecting poorest students and the inequalities that it produces, from which it might be defined needs profiles, as well as actions on the factors in which it is possible to influence. Finally, these results confirm that network analysis is fundamental for educational policy, especially linking reliable microdata with social macro-parameters because it allows us to infer how gaps in educational achievements are driven by students’ context at the time of assigning resources.

Keywords: complex network, educational deprivation, evidence-based policy, large-scale assessments, policy informatics

Procedia PDF Downloads 128
3741 Techniques to Teach Reading at Pre-Reading Stage

Authors: Anh Duong

Abstract:

The three-phase reading lesson has been put forth around the world as the new and innovative framework which is corresponding to the learner-centered trend in English language teaching and learning. Among three stages, pre-reading attracts many teachers’ and researchers’ attention for its vital role in preparing students with knowledge and interest in reading class. The researcher’s desire to exemplify effectiveness of activities prior to text reading has provoked the current study. Three main aspects were investigated in this paper, i.e. teachers’ and student’s perception of pre-reading stage, teachers’ exploitation of pre-reading techniques and teachers’ recommendation of effective pre-reading activities. Aiming at pre-reading techniques for first-year students at English Department, this study involved 200 fresh-men and 10 teachers from Division 1 to participate in the questionnaire survey. Interviews with the teachers and classroom observation were employed as a tool to take an insight into the responses gained from the early instrument. After a detailed procedure of analyzing data, the researcher discovered that thanks to the participants’ acclamation of pre-reading stage, this phase was frequently conducted by the surveyed teachers. Despite the fact that pre-reading activities apparently put a hand in motivating students to read and creating a joyful learning atmosphere, they did not fulfill another function as supporting students’ reading comprehension. Therefore, a range of techniques and notices when preparing and conducting pre-reading phase was detected from the interviewed teachers. The findings assisted the researcher to propose some related pedagogical implications concerning teachers’ source of pre-reading techniques, variations of suggested activities and first-year reading syllabus.

Keywords: pre-reading stage, pre-reading techniques, teaching reading, language teaching

Procedia PDF Downloads 488
3740 The Practices and Challenges of Secondary School Cluster Supervisors in Implementing School Improvement Program in Saesie Tsaeda Emba Woreda, Eastern Zone of Tigray Region

Authors: Haftom Teshale Gebre

Abstract:

According to the ministry of education’s school improvement program blueprint document (2007), the timely and basic aim of the program is to improve students’ academic achievement through creating conducive teaching and learning environments and with the active involvement of parents in the teaching and learning process. The general objective of the research is to examine the practices of cluster school supervisors in implementing school improvement programs and the major factors affecting the study area. The study used both primary and secondary sources, and the sample size was 93. Twelve people are chosen from each of the two clusters (Edaga Hamus and Adi-kelebes). And cluster ferewyni are Tekli suwaat, Edaga robue, and Kiros Alemayo. In the analysis stage, several interrelated pieces of information were summarized and arranged to make the analysis easily manageable by using statistics and data (STATA). Study findings revealed that the major four domains impacted by school improvement programs through their mean, standard deviation, and variance were 2.688172, 1.052724, and 1.108228, respectively. And also, the researcher can conclude that the major factors of the school improvement program and mostly cluster supervisors were inadequate attention given to supervision service and no experience in the practice of supervision in the study area.

Keywords: cluster, eastern Tigray, Saesie Tsaeda Emba, SPI

Procedia PDF Downloads 36
3739 Finite-Sum Optimization: Adaptivity to Smoothness and Loopless Variance Reduction

Authors: Bastien Batardière, Joon Kwon

Abstract:

For finite-sum optimization, variance-reduced gradient methods (VR) compute at each iteration the gradient of a single function (or of a mini-batch), and yet achieve faster convergence than SGD thanks to a carefully crafted lower-variance stochastic gradient estimator that reuses past gradients. Another important line of research of the past decade in continuous optimization is the adaptive algorithms such as AdaGrad, that dynamically adjust the (possibly coordinate-wise) learning rate to past gradients and thereby adapt to the geometry of the objective function. Variants such as RMSprop and Adam demonstrate outstanding practical performance that have contributed to the success of deep learning. In this work, we present AdaLVR, which combines the AdaGrad algorithm with loopless variance-reduced gradient estimators such as SAGA or L-SVRG that benefits from a straightforward construction and a streamlined analysis. We assess that AdaLVR inherits both good convergence properties from VR methods and the adaptive nature of AdaGrad: in the case of L-smooth convex functions we establish a gradient complexity of O(n + (L + √ nL)/ε) without prior knowledge of L. Numerical experiments demonstrate the superiority of AdaLVR over state-of-the-art methods. Moreover, we empirically show that the RMSprop and Adam algorithm combined with variance-reduced gradients estimators achieve even faster convergence.

Keywords: convex optimization, variance reduction, adaptive algorithms, loopless

Procedia PDF Downloads 74
3738 An Electrocardiography Deep Learning Model to Detect Atrial Fibrillation on Clinical Application

Authors: Jui-Chien Hsieh

Abstract:

Background:12-lead electrocardiography(ECG) is one of frequently-used tools to detect atrial fibrillation (AF), which might degenerate into life-threaten stroke, in clinical Practice. Based on this study, the AF detection by the clinically-used 12-lead ECG device has only 0.73~0.77 positive predictive value (ppv). Objective: It is on great demand to develop a new algorithm to improve the precision of AF detection using 12-lead ECG. Due to the progress on artificial intelligence (AI), we develop an ECG deep model that has the ability to recognize AF patterns and reduce false-positive errors. Methods: In this study, (1) 570-sample 12-lead ECG reports whose computer interpretation by the ECG device was AF were collected as the training dataset. The ECG reports were interpreted by 2 senior cardiologists, and confirmed that the precision of AF detection by the ECG device is 0.73.; (2) 88 12-lead ECG reports whose computer interpretation generated by the ECG device was AF were used as test dataset. Cardiologist confirmed that 68 cases of 88 reports were AF, and others were not AF. The precision of AF detection by ECG device is about 0.77; (3) A parallel 4-layer 1 dimensional convolutional neural network (CNN) was developed to identify AF based on limb-lead ECGs and chest-lead ECGs. Results: The results indicated that this model has better performance on AF detection than traditional computer interpretation of the ECG device in 88 test samples with 0.94 ppv, 0.98 sensitivity, 0.80 specificity. Conclusions: As compared to the clinical ECG device, this AI ECG model promotes the precision of AF detection from 0.77 to 0.94, and can generate impacts on clinical applications.

Keywords: 12-lead ECG, atrial fibrillation, deep learning, convolutional neural network

Procedia PDF Downloads 117
3737 Automatic Identification and Classification of Contaminated Biodegradable Plastics using Machine Learning Algorithms and Hyperspectral Imaging Technology

Authors: Nutcha Taneepanichskul, Helen C. Hailes, Mark Miodownik

Abstract:

Plastic waste has emerged as a critical global environmental challenge, primarily driven by the prevalent use of conventional plastics derived from petrochemical refining and manufacturing processes in modern packaging. While these plastics serve vital functions, their persistence in the environment post-disposal poses significant threats to ecosystems. Addressing this issue necessitates approaches, one of which involves the development of biodegradable plastics designed to degrade under controlled conditions, such as industrial composting facilities. It is imperative to note that compostable plastics are engineered for degradation within specific environments and are not suited for uncontrolled settings, including natural landscapes and aquatic ecosystems. The full benefits of compostable packaging are realized when subjected to industrial composting, preventing environmental contamination and waste stream pollution. Therefore, effective sorting technologies are essential to enhance composting rates for these materials and diminish the risk of contaminating recycling streams. In this study, it leverage hyperspectral imaging technology (HSI) coupled with advanced machine learning algorithms to accurately identify various types of plastics, encompassing conventional variants like Polyethylene terephthalate (PET), Polypropylene (PP), Low density polyethylene (LDPE), High density polyethylene (HDPE) and biodegradable alternatives such as Polybutylene adipate terephthalate (PBAT), Polylactic acid (PLA), and Polyhydroxyalkanoates (PHA). The dataset is partitioned into three subsets: a training dataset comprising uncontaminated conventional and biodegradable plastics, a validation dataset encompassing contaminated plastics of both types, and a testing dataset featuring real-world packaging items in both pristine and contaminated states. Five distinct machine learning algorithms, namely Partial Least Squares Discriminant Analysis (PLS-DA), Support Vector Machine (SVM), Convolutional Neural Network (CNN), Logistic Regression, and Decision Tree Algorithm, were developed and evaluated for their classification performance. Remarkably, the Logistic Regression and CNN model exhibited the most promising outcomes, achieving a perfect accuracy rate of 100% for the training and validation datasets. Notably, the testing dataset yielded an accuracy exceeding 80%. The successful implementation of this sorting technology within recycling and composting facilities holds the potential to significantly elevate recycling and composting rates. As a result, the envisioned circular economy for plastics can be established, thereby offering a viable solution to mitigate plastic pollution.

Keywords: biodegradable plastics, sorting technology, hyperspectral imaging technology, machine learning algorithms

Procedia PDF Downloads 85
3736 Analyzing of Speed Disparity in Mixed Vehicle Technologies on Horizontal Curves

Authors: Tahmina Sultana, Yasser Hassan

Abstract:

Vehicle technologies rapidly evolving due to their multifaceted advantages. Adapted different vehicle technologies like connectivity and automation on the same roads with conventional vehicles controlled by human drivers may increase speed disparity in mixed vehicle technologies. Identifying relationships between speed distribution measures of different vehicles and road geometry can be an indicator of speed disparity in mixed technologies. Previous studies proved that speed disparity measures and traffic accidents are inextricably related. Horizontal curves from three geographic areas were selected based on relevant criteria, and speed data were collected at the midpoint of the preceding tangent and starting, ending, and middle point of the curve. Multiple linear mixed effect models (LME) were developed using the instantaneous speed measures representing the speed of vehicles at different points of horizontal curves to recognize relationships between speed variance (standard deviation) and road geometry. A simulation-based framework (Monte Carlo) was introduced to check the speed disparity on horizontal curves in mixed vehicle technologies when consideration is given to the interactions among connected vehicles (CVs), autonomous vehicles (AVs), and non-connected vehicles (NCVs) on horizontal curves. The Monte Carlo method was used in the simulation to randomly sample values for the various parameters from their respective distributions. Theresults show that NCVs had higher speed variation than CVs and AVs. In addition, AVs and CVs contributed to reduce speed disparity in the mixed vehicle technologies in any penetration rates.

Keywords: autonomous vehicles, connected vehicles, non-connected vehicles, speed variance

Procedia PDF Downloads 147
3735 Effect of Particle Shape on Monotonic and Cyclic Biaxial Behaviour of Sand Using Discrete Element Method

Authors: Raj Banerjee, Y. M. Parulekar, Aniruddha Sengupta, J. Chattopadhyay

Abstract:

This study proposes a Discrete Element Method (DEM) simulation using a commercial software PFC 2D (2019) for quantitatively simulating the monotonic and cyclic behaviour of sand using irregular shapes of sand grains. A preliminary analysis of the number of particles for optimal Representative Element Volume (REV) simulation of dimension 35mm x 35mm x 70mm using the scaled Grain Size Distribution (GSD) of sand is carried out. Subsequently, the effect of particle shape on the performance of sand during monotonic and cyclic bi-axial tests is assessed using numerical simulation. The validation of the numerical simulation for one case is carried out using the test results from the literature. Further numerical studies are performed in which the particles in REV are simulated by mixing round discs with irregular clumps (100% round disc, 75% round disc 25% irregular clump, 50% round disc 50% irregular clump, 25% round disc 75% irregular clump, 100% irregular clump) in different proportions using Dry Deposition (DD) method. The macro response for monotonic loading shows that irregular sand has a higher strength than round particles and that the Mohr-Coulomb failure envelope depends on the shape of the grains. During cyclic loading, it is observed that the liquefaction resistance curve (Cyclic Stress Ratio (CSR)-Number of cycles (N)) of sand is dependent on the combination of particle shapes with different proportions.

Keywords: biaxial test, particle shape, monotonic, cyclic

Procedia PDF Downloads 76
3734 The Preceptorship Experience and Clinical Competence of Final Year Nursing Students

Authors: Susan Ka Yee Chow

Abstract:

Effective clinical preceptorship is affecting students’ competence and fostering their growth in applying theoretical knowledge and skills in clinical settings. Any difference between the expected and actual learning experience will reduce nursing students’ interest in clinical practices and having a negative consequence with their clinical performance. This cross-sectional study is an attempt to compare the differences between preferred and actual preceptorship experience of final year nursing students, and to examine the relationship between the actual preceptorship experience and perceived clinical competence of the students in a tertiary institution. Participants of the study were final year bachelor nursing students of a self-financing tertiary institution in Hong Kong. The instruments used to measure the effectiveness of clinical preceptorship was developed by the participating institution. The scale consisted of five items in a 5-point likert scale. The questions including goals development, critical thinking, learning objectives, asking questions and providing feedback to students. The “Clinical Competence Questionnaire” by Liou & Cheng (2014) was used to examine students’ perceived clinical competences. The scale consisted of 47 items categorized into four domains, namely nursing professional behaviours; skill competence: general performance; skill competence: core nursing skills and skill competence: advanced nursing skills. There were 193 questionnaires returned with a response rate of 89%. The paired t-test was used to compare the differences between preferred and actual preceptorship experiences of students. The results showed significant differences (p<0.001) for the five questions. The mean for the preferred scores is higher than the actual scores resulting statistically significance. The maximum mean difference was accepted goal and the highest mean different was giving feedback. The Pearson Correlation Coefficient was used to examine the relationship. The results showed moderate correlations between nursing professional behaviours with asking questions and providing feedback. Providing useful feedback to students is having moderate correlations with all domains of the Clinical Competence Questionnaire (r=0.269 – 0.345). It is concluded that nursing students do not have a positive perception of the clinical preceptorship. Their perceptions are significantly different from their expected preceptorship. If students were given more opportunities to ask questions in a pedagogical atmosphere, their perceived clinical competence and learning outcomes could be improved as a result.

Keywords: clinical preceptor, clinical competence, clinical practicum, nursing students

Procedia PDF Downloads 131
3733 Deep Learning-Based Automated Structure Deterioration Detection for Building Structures: A Technological Advancement for Ensuring Structural Integrity

Authors: Kavita Bodke

Abstract:

Structural health monitoring (SHM) is experiencing growth, necessitating the development of distinct methodologies to address its expanding scope effectively. In this study, we developed automatic structure damage identification, which incorporates three unique types of a building’s structural integrity. The first pertains to the presence of fractures within the structure, the second relates to the issue of dampness within the structure, and the third involves corrosion inside the structure. This study employs image classification techniques to discern between intact and impaired structures within structural data. The aim of this research is to find automatic damage detection with the probability of each damage class being present in one image. Based on this probability, we know which class has a higher probability or is more affected than the other classes. Utilizing photographs captured by a mobile camera serves as the input for an image classification system. Image classification was employed in our study to perform multi-class and multi-label classification. The objective was to categorize structural data based on the presence of cracks, moisture, and corrosion. In the context of multi-class image classification, our study employed three distinct methodologies: Random Forest, Multilayer Perceptron, and CNN. For the task of multi-label image classification, the models employed were Rasnet, Xceptionet, and Inception.

Keywords: SHM, CNN, deep learning, multi-class classification, multi-label classification

Procedia PDF Downloads 44
3732 Learner's Difficulties Acquiring English: The Case of Native Speakers of Rio de La Plata Spanish Towards Justifying the Need for Corpora

Authors: Maria Zinnia Bardas Hoffmann

Abstract:

Contrastive Analysis (CA) is the systematic comparison between two languages. It stems from the notion that errors are caused by interference of the L1 system in the acquisition process of an L2. CA represents a useful tool to understand the nature of learning and acquisition. Also, this particular method promises a path to un-derstand the nature of underlying cognitive processes, even when other factors such as intrinsic motivation and teaching strategies were found to best explain student’s problems in acquisition. CA study is justified not only from the need to get a deeper understanding of the nature of SLA, but as an invaluable source to provide clues, at a cognitive level, for those general processes involved in rule formation and abstract thought. It is relevant for cross disciplinary studies and the fields of Computational Thought, Natural Language processing, Applied Linguistics, Cognitive Linguistics and Math Theory. That being said, this paper intends to address here as well its own set of constraints and limitations. Finally, this paper: (a) aims at identifying some of the difficulties students may find in their learning process due to the nature of their specific variety of L1, Rio de la Plata Spanish (RPS), (b) represents an attempt to discuss the necessity for specific models to approach CA.

Keywords: second language acquisition, applied linguistics, contrastive analysis, applied contrastive analysis English language department, meta-linguistic rules, cross-linguistics studies, computational thought, natural language processing

Procedia PDF Downloads 155
3731 The Perceptions of High School English Home Language Learners on Fostering 21st Century Skills Through the Use of Technology in the Classroom

Authors: Lisa Naudine Parrock, Geoffrey Lautenbach

Abstract:

The changes brought to society by the technological development in the Fourth Industrial Revolution are also reaching the sphere of education and the education system needs to respond. Students need skills such as communication, collaboration, creativity and critical thinking in order to be successful in the 21st Century, which could be developed through the meaningful use of technology. This study is theorized by the 21st Century Framework for Learning and examines the student perceptions of grade 10 and 11 English Home language learners on how the technology used in their English classroom contributes to the development of 21st Century skills. The researcher adopted a constructivist paradigm and presented findings based on a general qualitative method. The study found that students perceived the use of technology in the classroom positively contributed to their development of communication, collaboration, creativity and critical thinking. Students also perceived technology as contributing to their access to information, a positive classroom atmosphere, heightened engagement in learning and developing skills necessary for their future. In addition, this study highlighted certain pedagogical strategies and digital tools that support the development of 21st Century skills. The findings suggest that the meaningful integration of technology fosters the development of 21st Century skills in grade 10 and 11 learners.

Keywords: educational technology, 21st century skills, fourth industrial revolution, affordances of technology

Procedia PDF Downloads 138
3730 Climate Safe House: A Community Housing Project Tackling Catastrophic Sea Level Rise in Coastal Communities

Authors: Chris Fersterer, Col Fay, Tobias Danielmeier, Kat Achterberg, Scott Willis

Abstract:

New Zealand, an island nation, has an extensive coastline peppered with small communities of iconic buildings known as Bachs. Post WWII, these modest buildings were constructed by their owners as retreats and generally were small, low cost, often using recycled material and often they fell below current acceptable building standards. In the latter part of the 20th century, real estate prices in many of these communities remained low and these areas became permanent residences for people attracted to this affordable lifestyle choice. The Blueskin Resilient Communities Trust (BRCT) is an organisation that recognises the vulnerability of communities in low lying settlements as now being prone to increased flood threat brought about by climate change and sea level rise. Some of the inhabitants of Blueskin Bay, Otago, NZ have already found their properties to be un-insurable because of increased frequency of flood events and property values have slumped accordingly. Territorial authorities also acknowledge this increased risk and have created additional compliance measures for new buildings that are less than 2 m above tidal peaks. Community resilience becomes an additional concern where inhabitants are attracted to a lifestyle associated with a specific location and its people when this lifestyle is unable to be met in a suburban or city context. Traditional models of social housing fail to provide the sense of community connectedness and identity enjoyed by the current residents of Blueskin Bay. BRCT have partnered with the Otago Polytechnic Design School to design a new form of community housing that can react to this environmental change. It is a longitudinal project incorporating participatory approaches as a means of getting people ‘on board’, to understand complex systems and co-develop solutions. In the first period, they are seeking industry support and funding to develop a transportable and fully self-contained housing model that exploits current technologies. BRCT also hope that the building will become an educational tool to highlight climate change issues facing us today. This paper uses the Climate Safe House (CSH) as a case study for education in architectural sustainability through experiential learning offered as part of the Otago Polytechnics Bachelor of Design. Students engage with the project with research methodologies, including site surveys, resident interviews, data sourced from government agencies and physical modelling. The process involves collaboration across design disciplines including product and interior design but also includes connections with industry, both within the education institution and stakeholder industries introduced through BRCT. This project offers a rich learning environment where students become engaged through project based learning within a community of practice, including architecture, construction, energy and other related fields. The design outcomes are expressed in a series of public exhibitions and forums where community input is sought in a truly participatory process.

Keywords: community resilience, problem based learning, project based learning, case study

Procedia PDF Downloads 293
3729 Entrepreneurial Leadership in Malaysian Public University: Competency and Behavior in the Face of Institutional Adversity

Authors: Noorlizawati Abd Rahim, Zainai Mohamed, Zaidatun Tasir, Astuty Amrin, Haliyana Khalid, Nina Diana Nawi

Abstract:

Entrepreneurial leaders have been sought as in-demand talents to lead profit-driven organizations during turbulent and unprecedented times. However, research regarding the pertinence of their roles in the public sector has been limited. This paper examined the characteristics of the challenging experiences encountered by senior leaders in public universities that require them to embrace entrepreneurialism in their leadership. Through a focus group interview with five Malaysian university top senior leaders with experience being Vice-Chancellor, we explored and developed a framework of institutional adversity characteristics and exemplary entrepreneurial leadership competency in the face of adversity. Complexity of diverse stakeholders, multiplicity of academic disciplines, unfamiliarity to lead different and broader roles, leading new directions, and creating change in high velocity and uncertain environment are among the dimensions that characterise institutional adversities. Our findings revealed that learning agility, opportunity recognition capacity, and bridging capability are among the characteristics of entrepreneurial university leaders. The findings reinforced that the presence of specific attributes in institutional adversity and experiences in overcoming those challenges may contribute to the development of entrepreneurial leadership capabilities.

Keywords: bridging capability, entrepreneurial leadership, leadership development, learning agility, opportunity recognition, university leaders

Procedia PDF Downloads 115
3728 Evaluation of a Potential Metabolism-Mediated Drug-Drug Interaction between Carvedilol and Fluvoxamine in Rats

Authors: Ana-Maria Gheldiu, Bianca M. Abrudan, Maria A. Neag, Laurian Vlase, Dana M. Muntean

Abstract:

Background information: The objective of this study was to investigate the effect of multiple-dose fluvoxamine on the pharmacokinetic profile of single-dose carvedilol in rats, in order to evaluate this possible drug-drug pharmacokinetic interaction. Methods: A preclinical study, in 28 white male Wistar rats, was conducted. Each rat was cannulated on the femoral vein, prior to being connected to BASi Culex ABC®. Carvedilol was orally administrated in rats (3.57 mg/kg body mass (b.m.)) in the absence of fluvoxamine or after a pre-treatment with multiple oral doses of fluvoxamine (14.28 mg/kg b.m.). The plasma concentrations of carvedilol were estimated by high performance liquid chromatography-tandem mass spectrometry. The pharmacokinetic parameters of carvedilol were analyzed by non-compartmental method. Results: After carvediol co-administration with fluvoxamine, an approximately 2-fold increase in the exposure of carvedilol was observed, considering the significantly elevated value of the total area under the concentration versus time curve (AUC₀₋∞). Moreover, an increase by approximately 145% of the peak plasma concentration was found, as well as an augmentation by approximately 230% of the half life time of carvedilol was observed. Conclusion: Fluvoxamine co-administration led to a significant alteration of carvedilol’s pharmacokinetic profile in rats, these effects could be explained by the existence of a drug-drug interaction mediated by CYP2D6 inhibition. Acknowledgement: This work was supported by CNCS Romania – project PNII-RU-TE-2014-4-0242.

Keywords: carvedilol, fluvoxamine, drug-drug pharmacokinetic interaction, rats

Procedia PDF Downloads 278