Search results for: supervised machine learning algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11237

Search results for: supervised machine learning algorithm

8957 Book Exchange System with a Hybrid Recommendation Engine

Authors: Nilki Upathissa, Torin Wirasinghe

Abstract:

This solution addresses the challenges faced by traditional bookstores and the limitations of digital media, striking a balance between the tactile experience of printed books and the convenience of modern technology. The book exchange system offers a sustainable alternative, empowering users to access a diverse range of books while promoting community engagement. The user-friendly interfaces incorporated into the book exchange system ensure a seamless and enjoyable experience for users. Intuitive features for book management, search, and messaging facilitate effortless exchanges and interactions between users. By streamlining the process, the system encourages readers to explore new books aligned with their interests, enhancing the overall reading experience. Central to the system's success is the hybrid recommendation engine, which leverages advanced technologies such as Long Short-Term Memory (LSTM) models. By analyzing user input, the engine accurately predicts genre preferences, enabling personalized book recommendations. The hybrid approach integrates multiple technologies, including user interfaces, machine learning models, and recommendation algorithms, to ensure the accuracy and diversity of the recommendations. The evaluation of the book exchange system with the hybrid recommendation engine demonstrated exceptional performance across key metrics. The high accuracy score of 0.97 highlights the system's ability to provide relevant recommendations, enhancing users' chances of discovering books that resonate with their interests. The commendable precision, recall, and F1score scores further validate the system's efficacy in offering appropriate book suggestions. Additionally, the curve classifications substantiate the system's effectiveness in distinguishing positive and negative recommendations. This metric provides confidence in the system's ability to navigate the vast landscape of book choices and deliver recommendations that align with users' preferences. Furthermore, the implementation of this book exchange system with a hybrid recommendation engine has the potential to revolutionize the way readers interact with printed books. By facilitating book exchanges and providing personalized recommendations, the system encourages a sense of community and exploration within the reading community. Moreover, the emphasis on sustainability aligns with the growing global consciousness towards eco-friendly practices. With its robust technical approach and promising evaluation results, this solution paves the way for a more inclusive, accessible, and enjoyable reading experience for book lovers worldwide. In conclusion, the developed book exchange system with a hybrid recommendation engine represents a progressive solution to the challenges faced by traditional bookstores and the limitations of digital media. By promoting sustainability, widening access to printed books, and fostering engagement with reading, this system addresses the evolving needs of book enthusiasts. The integration of user-friendly interfaces, advanced machine learning models, and recommendation algorithms ensure accurate and diverse book recommendations, enriching the reading experience for users.

Keywords: recommendation systems, hybrid recommendation systems, machine learning, data science, long short-term memory, recurrent neural network

Procedia PDF Downloads 74
8956 Automatic Aggregation and Embedding of Microservices for Optimized Deployments

Authors: Pablo Chico De Guzman, Cesar Sanchez

Abstract:

Microservices are a software development methodology in which applications are built by composing a set of independently deploy-able, small, modular services. Each service runs a unique process and it gets instantiated and deployed in one or more machines (we assume that different microservices are deployed into different machines). Microservices are becoming the de facto standard for developing distributed cloud applications due to their reduced release cycles. In principle, the responsibility of a microservice can be as simple as implementing a single function, which can lead to the following issues: - Resource fragmentation due to the virtual machine boundary. - Poor communication performance between microservices. Two composition techniques can be used to optimize resource fragmentation and communication performance: aggregation and embedding of microservices. Aggregation allows the deployment of a set of microservices on the same machine using a proxy server. Aggregation helps to reduce resource fragmentation, and is particularly useful when the aggregated services have a similar scalability behavior. Embedding deals with communication performance by deploying on the same virtual machine those microservices that require a communication channel (localhost bandwidth is reported to be about 40 times faster than cloud vendor local networks and it offers better reliability). Embedding can also reduce dependencies on load balancer services since the communication takes place on a single virtual machine. For example, assume that microservice A has two instances, a1 and a2, and it communicates with microservice B, which also has two instances, b1 and b2. One embedding can deploy a1 and b1 on machine m1, and a2 and b2 are deployed on a different machine m2. This deployment configuration allows each pair (a1-b1), (a2-b2) to communicate using the localhost interface without the need of a load balancer between microservices A and B. Aggregation and embedding techniques are complex since different microservices might have incompatible runtime dependencies which forbid them from being installed on the same machine. There is also a security concern since the attack surface between microservices can be larger. Luckily, container technology allows to run several processes on the same machine in an isolated manner, solving the incompatibility of running dependencies and the previous security concern, thus greatly simplifying aggregation/embedding implementations by just deploying a microservice container on the same machine as the aggregated/embedded microservice container. Therefore, a wide variety of deployment configurations can be described by combining aggregation and embedding to create an efficient and robust microservice architecture. This paper presents a formal method that receives a declarative definition of a microservice architecture and proposes different optimized deployment configurations by aggregating/embedding microservices. The first prototype is based on i2kit, a deployment tool also submitted to ICWS 2018. The proposed prototype optimizes the following parameters: network/system performance, resource usage, resource costs and failure tolerance.

Keywords: aggregation, deployment, embedding, resource allocation

Procedia PDF Downloads 188
8955 Deciphering Orangutan Drawing Behavior Using Artificial Intelligence

Authors: Benjamin Beltzung, Marie Pelé, Julien P. Renoult, Cédric Sueur

Abstract:

To this day, it is not known if drawing is specifically human behavior or if this behavior finds its origins in ancestor species. An interesting window to enlighten this question is to analyze the drawing behavior in genetically close to human species, such as non-human primate species. A good candidate for this approach is the orangutan, who shares 97% of our genes and exhibits multiple human-like behaviors. Focusing on figurative aspects may not be suitable for orangutans’ drawings, which may appear as scribbles but may have meaning. A manual feature selection would lead to an anthropocentric bias, as the features selected by humans may not match with those relevant for orangutans. In the present study, we used deep learning to analyze the drawings of a female orangutan named Molly († in 2011), who has produced 1,299 drawings in her last five years as part of a behavioral enrichment program at the Tama Zoo in Japan. We investigate multiple ways to decipher Molly’s drawings. First, we demonstrate the existence of differences between seasons by training a deep learning model to classify Molly’s drawings according to the seasons. Then, to understand and interpret these seasonal differences, we analyze how the information spreads within the network, from shallow to deep layers, where early layers encode simple local features and deep layers encode more complex and global information. More precisely, we investigate the impact of feature complexity on classification accuracy through features extraction fed to a Support Vector Machine. Last, we leverage style transfer to dissociate features associated with drawing style from those describing the representational content and analyze the relative importance of these two types of features in explaining seasonal variation. Content features were relevant for the classification, showing the presence of meaning in these non-figurative drawings and the ability of deep learning to decipher these differences. The style of the drawings was also relevant, as style features encoded enough information to have a classification better than random. The accuracy of style features was higher for deeper layers, demonstrating and highlighting the variation of style between seasons in Molly’s drawings. Through this study, we demonstrate how deep learning can help at finding meanings in non-figurative drawings and interpret these differences.

Keywords: cognition, deep learning, drawing behavior, interpretability

Procedia PDF Downloads 141
8954 A Hierarchical Method for Multi-Class Probabilistic Classification Vector Machines

Authors: P. Byrnes, F. A. DiazDelaO

Abstract:

The Support Vector Machine (SVM) has become widely recognised as one of the leading algorithms in machine learning for both regression and binary classification. It expresses predictions in terms of a linear combination of kernel functions, referred to as support vectors. Despite its popularity amongst practitioners, SVM has some limitations, with the most significant being the generation of point prediction as opposed to predictive distributions. Stemming from this issue, a probabilistic model namely, Probabilistic Classification Vector Machines (PCVM), has been proposed which respects the original functional form of SVM whilst also providing a predictive distribution. As physical system designs become more complex, an increasing number of classification tasks involving industrial applications consist of more than two classes. Consequently, this research proposes a framework which allows for the extension of PCVM to a multi class setting. Additionally, the original PCVM framework relies on the use of type II maximum likelihood to provide estimates for both the kernel hyperparameters and model evidence. In a high dimensional multi class setting, however, this approach has been shown to be ineffective due to bad scaling as the number of classes increases. Accordingly, we propose the application of Markov Chain Monte Carlo (MCMC) based methods to provide a posterior distribution over both parameters and hyperparameters. The proposed framework will be validated against current multi class classifiers through synthetic and real life implementations.

Keywords: probabilistic classification vector machines, multi class classification, MCMC, support vector machines

Procedia PDF Downloads 213
8953 Grating Scale Thermal Expansion Error Compensation for Large Machine Tools Based on Multiple Temperature Detection

Authors: Wenlong Feng, Zhenchun Du, Jianguo Yang

Abstract:

To decrease the grating scale thermal expansion error, a novel method which based on multiple temperature detections is proposed. Several temperature sensors are installed on the grating scale and the temperatures of these sensors are recorded. The temperatures of every point on the grating scale are calculated by interpolating between adjacent sensors. According to the thermal expansion principle, the grating scale thermal expansion error model can be established by doing the integral for the variations of position and temperature. A novel compensation method is proposed in this paper. By applying the established error model, the grating scale thermal expansion error is decreased by 90% compared with no compensation. The residual positioning error of the grating scale is less than 15um/10m and the accuracy of the machine tool is significant improved.

Keywords: thermal expansion error of grating scale, error compensation, machine tools, integral method

Procedia PDF Downloads 351
8952 A Segmentation Method for Grayscale Images Based on the Firefly Algorithm and the Gaussian Mixture Model

Authors: Donatella Giuliani

Abstract:

In this research, we propose an unsupervised grayscale image segmentation method based on a combination of the Firefly Algorithm and the Gaussian Mixture Model. Firstly, the Firefly Algorithm has been applied in a histogram-based research of cluster means. The Firefly Algorithm is a stochastic global optimization technique, centered on the flashing characteristics of fireflies. In this context it has been performed to determine the number of clusters and the related cluster means in a histogram-based segmentation approach. Successively these means are used in the initialization step for the parameter estimation of a Gaussian Mixture Model. The parametric probability density function of a Gaussian Mixture Model is represented as a weighted sum of Gaussian component densities, whose parameters are evaluated applying the iterative Expectation-Maximization technique. The coefficients of the linear super-position of Gaussians can be thought as prior probabilities of each component. Applying the Bayes rule, the posterior probabilities of the grayscale intensities have been evaluated, therefore their maxima are used to assign each pixel to the clusters, according to their gray-level values. The proposed approach appears fairly solid and reliable when applied even to complex grayscale images. The validation has been performed by using different standard measures, more precisely: the Root Mean Square Error (RMSE), the Structural Content (SC), the Normalized Correlation Coefficient (NK) and the Davies-Bouldin (DB) index. The achieved results have strongly confirmed the robustness of this gray scale segmentation method based on a metaheuristic algorithm. Another noteworthy advantage of this methodology is due to the use of maxima of responsibilities for the pixel assignment that implies a consistent reduction of the computational costs.

Keywords: clustering images, firefly algorithm, Gaussian mixture model, meta heuristic algorithm, image segmentation

Procedia PDF Downloads 205
8951 Advancement of Computer Science Research in Nigeria: A Bibliometric Analysis of the Past Three Decades

Authors: Temidayo O. Omotehinwa, David O. Oyewola, Friday J. Agbo

Abstract:

This study aims to gather a proper perspective of the development landscape of Computer Science research in Nigeria. Therefore, a bibliometric analysis of 4,333 bibliographic records of Computer Science research in Nigeria in the last 31 years (1991-2021) was carried out. The bibliographic data were extracted from the Scopus database and analyzed using VOSviewer and the bibliometrix R package through the biblioshiny web interface. The findings of this study revealed that Computer Science research in Nigeria has a growth rate of 24.19%. The most developed and well-studied research areas in the Computer Science field in Nigeria are machine learning, data mining, and deep learning. The social structure analysis result revealed that there is a need for improved international collaborations. Sparsely established collaborations are largely influenced by geographic proximity. The funding analysis result showed that Computer Science research in Nigeria is under-funded. The findings of this study will be useful for researchers conducting Computer Science related research. Experts can gain insights into how to develop a strategic framework that will advance the field in a more impactful manner. Government agencies and policymakers can also utilize the outcome of this research to develop strategies for improved funding for Computer Science research.

Keywords: bibliometric analysis, biblioshiny, computer science, Nigeria, science mapping

Procedia PDF Downloads 93
8950 Explaining E-Learning Systems Usage in Higher Education Institutions: UTAUT Model

Authors: Muneer Abbad

Abstract:

This research explains the e-learning usage in a university in Jordan. Unified theory of acceptance and use of technology (UTAUT) model has been used as a base model to explain the usage. UTAUT is a model of individual acceptance that is compiled mainly from different models of technology acceptance. This research is the initial part from full explanations of the users' acceptance model that use Structural Equation Modelling (SEM) method to explain the users' acceptance of the e-learning systems based on UTAUT model. In this part data has been collected and prepared for further analysis. The main factors of UTAUT model has been tested as different factors using exploratory factor analysis (EFA). The second phase will be confirmatory factor analysis (CFA) and SEM to explain the users' acceptance of e-learning systems.

Keywords: e-learning, moodle, adoption, Unified Theory of Acceptance and Use of Technology (UTAUT)

Procedia PDF Downloads 387
8949 A Study to Explore the Views of Students regarding E-Learning as an Instructional Tool at University Level

Authors: Zafar Iqbal

Abstract:

This study involved students of 6th semester enrolled in a Bachelor of Computer Science Program at university level. In this era of science and technology, e-learning can be helpful for grassroots in providing them access to education tenant in less developed areas. It is a potential substitute of face-to-face teaching being used in different countries. The purpose of the study was to explore the views of students about e-learning (Facebook) as an instructional tool. By using purposive sampling technique an intact class of 30 students included both male and female were selected where e-learning was used as an instructional tool. The views of students were explored through qualitative approach by using focus group interviews. The approach was helpful to develop comprehensive understanding of students’ views towards e- learning. In addition, probing questions were also asked and recorded. Data was transcribed, generated nodes and then coded text against these nodes. For this purpose and further analysis, NVivo 10 software was used. Themes were generated and tangibly presented through cluster analysis. Findings were interesting and provide sufficient evidence that face book is a subsequent e-learning source for students of higher education. Students acknowledged it as best source of learning and it was aligned with their academic and social behavior. It was not time specific and therefore, feasible for students who work day time and can get on line access to the material when they got free time. There were some distracters (time wasters) reported by the students but can be minimized by little effort. In short, e-learning is need of the day and potential learning source for every individual who have access to internet living at any part of the globe.

Keywords: e-learning, facebook, instructional tool, higher education

Procedia PDF Downloads 363
8948 The Data-Driven Localized Wave Solution of the Fokas-Lenells Equation using PINN

Authors: Gautam Kumar Saharia, Sagardeep Talukdar, Riki Dutta, Sudipta Nandy

Abstract:

The physics informed neural network (PINN) method opens up an approach for numerically solving nonlinear partial differential equations leveraging fast calculating speed and high precession of modern computing systems. We construct the PINN based on strong universal approximation theorem and apply the initial-boundary value data and residual collocation points to weekly impose initial and boundary condition to the neural network and choose the optimization algorithms adaptive moment estimation (ADAM) and Limited-memory Broyden-Fletcher-Golfard-Shanno (L-BFGS) algorithm to optimize learnable parameter of the neural network. Next, we improve the PINN with a weighted loss function to obtain both the bright and dark soliton solutions of Fokas-Lenells equation (FLE). We find the proposed scheme of adjustable weight coefficients into PINN has a better convergence rate and generalizability than the basic PINN algorithm. We believe that the PINN approach to solve the partial differential equation appearing in nonlinear optics would be useful to study various optical phenomena.

Keywords: deep learning, optical Soliton, neural network, partial differential equation

Procedia PDF Downloads 106
8947 Auditory and Visual Perceptual Category Learning in Adults with ADHD: Implications for Learning Systems and Domain-General Factors

Authors: Yafit Gabay

Abstract:

Attention deficit hyperactivity disorder (ADHD) has been associated with both suboptimal functioning in the striatum and prefrontal cortex. Such abnormalities may impede the acquisition of perceptual categories, which are important for fundamental abilities such as object recognition and speech perception. Indeed, prior research has supported this possibility, demonstrating that children with ADHD have similar visual category learning performance as their neurotypical peers but use suboptimal learning strategies. However, much less is known about category learning processes in the auditory domain or among adults with ADHD in which prefrontal functions are more mature compared to children. Here, we investigated auditory and visual perceptual category learning in adults with ADHD and neurotypical individuals. Specifically, we examined learning of rule-based categories – presumed to be optimally learned by a frontal cortex-mediated hypothesis testing – and information-integration categories – hypothesized to be optimally learned by a striatally-mediated reinforcement learning system. Consistent with striatal and prefrontal cortical impairments observed in ADHD, our results show that across sensory modalities, both rule-based and information-integration category learning is impaired in adults with ADHD. Computational modeling analyses revealed that individuals with ADHD were slower to shift to optimal strategies than neurotypicals, regardless of category type or modality. Taken together, these results suggest that both explicit, frontally mediated and implicit, striatally mediated category learning are impaired in ADHD. These results suggest impairments across multiple learning systems in young adults with ADHD that extend across sensory modalities and likely arise from domain-general mechanisms.

Keywords: ADHD, category learning, modality, computational modeling

Procedia PDF Downloads 32
8946 Optimizing PharmD Education: Quantifying Curriculum Complexity to Address Student Burnout and Cognitive Overload

Authors: Frank Fan

Abstract:

PharmD (Doctor of Pharmacy) education has confronted an increasing challenge — curricular overload, a phenomenon resulting from the expansion of curricular requirements, as PharmD education strives to produce graduates who are practice-ready. The aftermath of the global pandemic has amplified the need for healthcare professionals, leading to a growing trend of assigning more responsibilities to them to address the global healthcare shortage. For instance, the pharmacist’s role has expanded to include not only compounding and distributing medication but also providing clinical services, including minor ailments management, patient counselling and vaccination. Consequently, PharmD programs have responded by continually expanding their curricula adding more requirements. While these changes aim to enhance the education and training of future professionals, they have also led to unintended consequences, including curricular overload, student burnout, and a potential decrease in program quality. To address the issue and ensure program quality, there is a growing need for evidence-based curriculum reforms. My research seeks to integrate Cognitive Load Theory, emerging machine learning algorithms within artificial intelligence (AI), and statistical approaches to develop a quantitative framework for optimizing curriculum design within the PharmD program at the University of Toronto, the largest PharmD program within Canada, to provide quantification and measurement of issues that currently are only discussed in terms of anecdote rather than data. This research will serve as a guide for curriculum planners, administrators, and educators, aiding in the comprehension of how the pharmacy degree program compares to others within and beyond the field of pharmacy. It will also shed light on opportunities to reduce the curricular load while maintaining its quality and rigor. Given that pharmacists constitute the third-largest healthcare workforce, their education shares similarities and challenges with other health education programs. Therefore, my evidence-based, data-driven curriculum analysis framework holds significant potential for training programs in other healthcare professions, including medicine, nursing, and physiotherapy.

Keywords: curriculum, curriculum analysis, health professions education, reflective writing, machine learning

Procedia PDF Downloads 48
8945 A Multiobjective Damping Function for Coordinated Control of Power System Stabilizer and Power Oscillation Damping

Authors: Jose D. Herrera, Mario A. Rios

Abstract:

This paper deals with the coordinated tuning of the Power System Stabilizer (PSS) controller and Power Oscillation Damping (POD) Controller of Flexible AC Transmission System (FACTS) in a multi-machine power systems. The coordinated tuning is based on the critical eigenvalues of the power system and a model reduction technique where the Hankel Singular Value method is applied. Through the linearized system model and the parameter-constrained nonlinear optimization algorithm, it can compute the parameters of both controllers. Moreover, the parameters are optimized simultaneously obtaining the gains of both controllers. Then, the nonlinear simulation to observe the time response of the controller is performed.

Keywords: electromechanical oscillations, power system stabilizers, power oscillation damping, hankel singular values

Procedia PDF Downloads 574
8944 Evaluating the Role of Multisensory Elements in Foreign Language Acquisition

Authors: Sari Myréen

Abstract:

The aim of this study was to evaluate the role of multisensory elements in enhancing and facilitating foreign language acquisition among adult students in a language classroom. The use of multisensory elements enables the creation of a student-centered classroom, where the focus is on individual learner’s language learning process, perceptions and motivation. Multisensory language learning is a pedagogical approach where the language learner uses all the senses more effectively than in a traditional in-class environment. Language learning is facilitated due to multisensory stimuli which increase the number of cognitive connections in the learner and take into consideration different types of learners. A living lab called Multisensory Space creates a relaxed and receptive state in the learners through various multisensory stimuli, and thus promotes their natural foreign language acquisition. Qualitative and quantitative data were collected in two questionnaire inquiries among the Finnish students of a higher education institute at the end of their basic French courses in December 2014 and 2016. The inquiries discussed the effects of multisensory elements on the students’ motivation to study French as well as their learning outcomes. The results show that the French classes in the Multisensory Space provide the students with an encouraging and pleasant learning environment, which has a positive impact on their motivation to study the foreign language as well as their language learning outcomes.

Keywords: foreign language acquisition, pedagogical approach, multisensory learning, transcultural learning

Procedia PDF Downloads 372
8943 The Need for the Utilization of Instructional Materials on the Teaching and Learning of Agricultural Science Education in Developing Countries

Authors: Ogoh Andrew Enokela

Abstract:

This paper dwelt on the need for the utilization of instructional materials with highlights on the type of instructional materials, selection, uses and their importance on the learning and teaching of Agricultural Science Education in developing countries. It further discussed the concept of improvisation with some recommendation in terms of availability, utilization on the teaching and learning of Agricultural Science Education.

Keywords: instructional materials, agricultural science education, improvisation, teaching and learning

Procedia PDF Downloads 305
8942 A Statistical Approach to Predict and Classify the Commercial Hatchability of Chickens Using Extrinsic Parameters of Breeders and Eggs

Authors: M. S. Wickramarachchi, L. S. Nawarathna, C. M. B. Dematawewa

Abstract:

Hatchery performance is critical for the profitability of poultry breeder operations. Some extrinsic parameters of eggs and breeders cause to increase or decrease the hatchability. This study aims to identify the affecting extrinsic parameters on the commercial hatchability of local chicken's eggs and determine the most efficient classification model with a hatchability rate greater than 90%. In this study, seven extrinsic parameters were considered: egg weight, moisture loss, breeders age, number of fertilised eggs, shell width, shell length, and shell thickness. Multiple linear regression was performed to determine the most influencing variable on hatchability. First, the correlation between each parameter and hatchability were checked. Then a multiple regression model was developed, and the accuracy of the fitted model was evaluated. Linear Discriminant Analysis (LDA), Classification and Regression Trees (CART), k-Nearest Neighbors (kNN), Support Vector Machines (SVM) with a linear kernel, and Random Forest (RF) algorithms were applied to classify the hatchability. This grouping process was conducted using binary classification techniques. Hatchability was negatively correlated with egg weight, breeders' age, shell width, shell length, and positive correlations were identified with moisture loss, number of fertilised eggs, and shell thickness. Multiple linear regression models were more accurate than single linear models regarding the highest coefficient of determination (R²) with 94% and minimum AIC and BIC values. According to the classification results, RF, CART, and kNN had performed the highest accuracy values 0.99, 0.975, and 0.972, respectively, for the commercial hatchery process. Therefore, the RF is the most appropriate machine learning algorithm for classifying the breeder outcomes, which are economically profitable or not, in a commercial hatchery.

Keywords: classification models, egg weight, fertilised eggs, multiple linear regression

Procedia PDF Downloads 78
8941 Examination Scheduling System with Proposed Algorithm

Authors: Tabrej Khan

Abstract:

Examination Scheduling System (ESS) is a scheduling system that targets as an exam committee in any academic institute to help them in managing the exams automatically. We present an algorithm for Examination Scheduling System. Nowadays, many universities have challenges with creating examination schedule fast with less confliction compared to hand works. Our aims are to develop a computerized system that can be used in examination scheduling in an academic institute versus available resources (Time, Hall, Invigilator and instructor) with no contradiction and achieve fairness among students. ESS was developed using HTML, C# language, Crystal Report and ASP.NET through Microsoft Visual Studio 2010 as developing tools with integrated SQL server database. This application can produce some benefits such as reducing the time spent in creating an exam schedule and achieving fairness among students

Keywords: examination scheduling system (ESS), algorithm, ASP.NET, crystal report

Procedia PDF Downloads 385
8940 The Implementation of Teaching and Learning Quality Assurance System at the Chaoyang University of Technology for Academic Year 2013-2015

Authors: Ting Hsiang Chang

Abstract:

Nowadays in Taiwan, higher education, which was previously more emphasized on teaching-oriented approaches, has gradually shifted to an approach more focusing on students learning outcomes. With student employment rate as an important indicator for University Program Evaluation periodically held by the Ministry of Education, it becomes extremely critical for a university to build up a teaching and learning quality assurance system to bridge the gap between learning and practice. Teaching and Learning Quality Assurance System has been built and implemented at Chaoyang University of Technology for years and has received substantial results. By employing various forms of evaluation and performance appraisals, the effectiveness of teaching and learning can consistently be tracked as a means of ensuring teaching and learning quality. This study aims to explore the evaluation system of teaching and learning quality assurance system at the Chaoyang University of Technology by means of content analysis. The study contents the evaluation reports on the teaching and learning quality assurance at the Chaoyang University of Technology in the Academic Year 2013-2015. The quantitative results of the assessment were analyzed using the five-point Likert Scale. Quality assurance Committee meetings were further held for examining and discussions on the results. To the end, the annual evaluation report is to be produced as references used to improve approaches in both teaching and learning. The findings indicate that there is a respective relationship between the overall teaching evaluation items and the teaching goals and core competencies. In addition, graduates’ feedbacks were also collected for further analysis to examine if the current educational planning is able to achieve the university’s teaching goal and cultivation of core competencies.

Keywords: core competencies, teaching and learning quality assurance system, teaching goals, university program evaluation

Procedia PDF Downloads 278
8939 Mobile Mediated Learning and Teachers Education in Less Resourced Region

Authors: Abdul Rashid Ahmadi, Samiullah Paracha, Hamidullah Sokout, Mohammad Hanif Gharana

Abstract:

Conventional educational practices, do not offer all the required skills for teachers to successfully survive in today’s workplace. Due to poor professional training, a big gap exists across the curriculum plan and the teacher practices in the classroom. As such, raising the quality of teaching through ICT-enabled training and professional development of teachers should be an urgent priority. ‘Mobile Learning’, in that vein, is an increasingly growing field of educational research and practice across schools and work places. In this paper, we propose a novel Mobile learning system that allows the users to learn through an intelligent mobile learning in cooperatively every-time and every-where. The system will reduce the training cost and increase consistency, efficiency, and data reliability. To establish that our system will display neither functional nor performance failure, the evaluation strategy is based on formal observation of users interacting with system followed by questionnaires and structured interviews.

Keywords: computer assisted learning, intelligent tutoring system, learner centered design, mobile mediated learning and teacher education

Procedia PDF Downloads 275
8938 Implications of Humanizing Pedagogy on Learning Design in a Technology-Enhanced Language Learning Environment: Critical Reflections on Student Identity and Agency

Authors: Mukhtar Raban

Abstract:

Nelson Mandela University subscribes to a humanizing pedagogy (HP), as housed under broader critical pedagogy, that underpins and informs learning and teaching activities at the institution. The investigation sought to explore the implications of humanizing and critical pedagogical considerations for a technology-enhanced language learning (TELL) environment in a university course. The paper inquires into the design of a learning resource in an online learning environment of an English communication module, that applied HP principles. With an objective of creating agentive spaces for foregrounding identity, student voice, critical self-reflection, and recognition of others’ humanity; a flexible and open 'My Presence' feature was added to the TELL environment that allowed students and lecturers to share elements of their backgrounds in a ‘mutually vulnerable’ manner as a way of establishing digital identity and a more ‘human’ presence in the online language learning encounter, serving as a catalyst for the recognition of the ‘other’. Following a qualitative research design, the study adopted an auto-ethnographic approach, complementing the critical inquiry nature embedded into the activity’s practices. The study’s findings provide critical reflections and deductions on the possibilities of leveraging digital human expression within a humanizing pedagogical framework to advance the realization of HP-adoption in language learning and teaching encounters. It was found that the consideration of humanizing pedagogical principles in the design of online learning was more effective when the critical outcomes were explicated to students and lecturers prior to the completion of the activities. The integration of humanizing pedagogy also led to a contextual advancement of ‘affective’ language learning. Upon critical reflection and analysis, student identity and agency can flourish in a technology-enhanced learning environment when humanizing, and critical pedagogy influences the learning design.

Keywords: critical reflection, humanizing pedagogy, student identity, technology-enhanced language learning

Procedia PDF Downloads 115
8937 Impact of Team-Based Learning Approach in English Language Learning Process: A Case Study of Universidad Federico Santa Maria

Authors: Yessica A. Aguilera

Abstract:

English is currently the only foreign language included in the national educational curriculum in Chile. The English curriculum establishes that once completed secondary education, students are expected to reach B1 level according to the Common European Reference Framework (CEFR) scale. However, the objective has not been achieved, and to the author’s best knowledge, there is still a severe lack of English language skills among students who have completed their secondary education studies. In order to deal with the fact that students do not manage English as expected, team-based learning (TBL) was introduced in English language lessons at the Universidad Federico Santa María (USM). TBL is a collaborative teaching-learning method which enhances active learning by combining individual and team work. This approach seeks to help students achieve course objectives while learning how to function in teams. The purpose of the research was to assess the implementation and effectiveness of TBL in English language classes at USM technical training education. Quantitative and qualitative data were collected from teachers and students about their experience through TBL. Research findings show that both teachers and students are satisfied with the method and that students’ engagement and participation in class is higher. Additionally, students score higher on examinations improving academic outcomes. The findings of the research have the potential to guide how TBL could be included in future English language courses.

Keywords: collaborative learning, college education, English language learning, team-based learning

Procedia PDF Downloads 174
8936 Improve Divers Tracking and Classification in Sonar Images Using Robust Diver Wake Detection Algorithm

Authors: Mohammad Tarek Al Muallim, Ozhan Duzenli, Ceyhun Ilguy

Abstract:

Harbor protection systems are so important. The need for automatic protection systems has increased over the last years. Diver detection active sonar has great significance. It used to detect underwater threats such as divers and autonomous underwater vehicle. To automatically detect such threats the sonar image is processed by algorithms. These algorithms used to detect, track and classify of underwater objects. In this work, divers tracking and classification algorithm is improved be proposing a robust wake detection method. To detect objects the sonar images is normalized then segmented based on fixed threshold. Next, the centroids of the segments are found and clustered based on distance metric. Then to track the objects linear Kalman filter is applied. To reduce effect of noise and creation of false tracks, the Kalman tracker is fine tuned. The tuning is done based on our active sonar specifications. After the tracks are initialed and updated they are subjected to a filtering stage to eliminate the noisy and unstable tracks. Also to eliminate object with a speed out of the diver speed range such as buoys and fast boats. Afterwards the result tracks are subjected to a classification stage to deiced the type of the object been tracked. Here the classification stage is to deice wither if the tracked object is an open circuit diver or a close circuit diver. At the classification stage, a small area around the object is extracted and a novel wake detection method is applied. The morphological features of the object with his wake is extracted. We used support vector machine to find the best classifier. The sonar training images and the test images are collected by ARMELSAN Defense Technologies Company using the portable diver detection sonar ARAS-2023. After applying the algorithm to the test sonar data, we get fine and stable tracks of the divers. The total classification accuracy achieved with the diver type is 97%.

Keywords: harbor protection, diver detection, active sonar, wake detection, diver classification

Procedia PDF Downloads 221
8935 A Novel Guided Search Based Multi-Objective Evolutionary Algorithm

Authors: A. Baviskar, C. Sandeep, K. Shankar

Abstract:

Solving Multi-objective Optimization Problems requires faster convergence and better spread. Though existing Evolutionary Algorithms (EA's) are able to achieve this, the computation effort can further be reduced by hybridizing them with innovative strategies. This study is focuses on converging to the pareto front faster while adapting the advantages of Strength Pareto Evolutionary Algorithm-II (SPEA-II) for a better spread. Two different approaches based on optimizing the objective functions independently are implemented. In the first method, the decision variables corresponding to the optima of individual objective functions are strategically used to guide the search towards the pareto front. In the second method, boundary points of the pareto front are calculated and their decision variables are seeded to the initial population. Both the methods are applied to different constrained and unconstrained multi-objective test functions. It is observed that proposed guided search based algorithm gives better convergence and diversity than several well-known existing algorithms (such as NSGA-II and SPEA-II) in considerably less number of iterations.

Keywords: boundary points, evolutionary algorithms (EA's), guided search, strength pareto evolutionary algorithm-II (SPEA-II)

Procedia PDF Downloads 258
8934 The Impact of Gamification on Self-Assessment for English Language Learners in Saudi Arabia

Authors: Wala A. Bagunaid, Maram Meccawy, Arwa Allinjawi, Zilal Meccawy

Abstract:

Continuous self-assessment becomes crucial in self-paced online learning environments. Students often depend on themselves to assess their progress; which is considered an essential requirement for any successful learning process. Today’s education institutions face major problems around student motivation and engagement. Thus, personalized e-learning systems aim to help and guide the students. Gamification provides an opportunity to help students for self-assessment and social comparison with other students through attempting to harness the motivational power of games and apply it to the learning environment. Furthermore, Open Social Student Modeling (OSSM) as considered as the latest user modeling technologies is believed to improve students’ self-assessment and to allow them to social comparison with other students. This research integrates OSSM approach and gamification concepts in order to provide self-assessment for English language learners at King Abdulaziz University (KAU). This is achieved through an interactive visual representation of their learning progress.

Keywords: e-learning system, gamification, motivation, social comparison, visualization

Procedia PDF Downloads 135
8933 A Comparative Analysis of Asymmetric Encryption Schemes on Android Messaging Service

Authors: Mabrouka Algherinai, Fatma Karkouri

Abstract:

Today, Short Message Service (SMS) is an important means of communication. SMS is not only used in informal environment for communication and transaction, but it is also used in formal environments such as institutions, organizations, companies, and business world as a tool for communication and transactions. Therefore, there is a need to secure the information that is being transmitted through this medium to ensure security of information both in transit and at rest. But, encryption has been identified as a means to provide security to SMS messages in transit and at rest. Several past researches have proposed and developed several encryption algorithms for SMS and Information Security. This research aims at comparing the performance of common Asymmetric encryption algorithms on SMS security. The research employs the use of three algorithms, namely RSA, McEliece, and RABIN. Several experiments were performed on SMS of various sizes on android mobile device. The experimental results show that each of the three techniques has different key generation, encryption, and decryption times. The efficiency of an algorithm is determined by the time that it takes for encryption, decryption, and key generation. The best algorithm can be chosen based on the least time required for encryption. The obtained results show the least time when McEliece size 4096 is used. RABIN size 4096 gives most time for encryption and so it is the least effective algorithm when considering encryption. Also, the research shows that McEliece size 2048 has the least time for key generation, and hence, it is the best algorithm as relating to key generation. The result of the algorithms also shows that RSA size 1024 is the most preferable algorithm in terms of decryption as it gives the least time for decryption.

Keywords: SMS, RSA, McEliece, RABIN

Procedia PDF Downloads 148
8932 Assessment of Ultra-High Cycle Fatigue Behavior of EN-GJL-250 Cast Iron Using Ultrasonic Fatigue Testing Machine

Authors: Saeedeh Bakhtiari, Johannes Depessemier, Stijn Hertelé, Wim De Waele

Abstract:

High cycle fatigue comprising up to 107 load cycles has been the subject of many studies, and the behavior of many materials was recorded adequately in this regime. However, many applications involve larger numbers of load cycles during the lifetime of machine components. In this ultra-high cycle regime, other failure mechanisms play, and the concept of a fatigue endurance limit (assumed for materials such as steel) is often an oversimplification of reality. When machine component design demands a high geometrical complexity, cast iron grades become interesting candidate materials. Grey cast iron is known for its low cost, high compressive strength, and good damping properties. However, the ultra-high cycle fatigue behavior of cast iron is poorly documented. The current work focuses on the ultra-high cycle fatigue behavior of EN-GJL-250 (GG25) grey cast iron by developing an ultrasonic (20 kHz) fatigue testing system. Moreover, the testing machine is instrumented to measure the temperature and the displacement of  the specimen, and to control the temperature. The high resonance frequency allowed to assess the  behavior of the cast iron of interest within a matter of days for ultra-high numbers of cycles, and repeat the tests to quantify the natural scatter in fatigue resistance.

Keywords: GG25, cast iron, ultra-high cycle fatigue, ultrasonic test

Procedia PDF Downloads 156
8931 A Unified Deep Framework for Joint 3d Pose Estimation and Action Recognition from a Single Color Camera

Authors: Huy Hieu Pham, Houssam Salmane, Louahdi Khoudour, Alain Crouzil, Pablo Zegers, Sergio Velastin

Abstract:

We present a deep learning-based multitask framework for joint 3D human pose estimation and action recognition from color video sequences. Our approach proceeds along two stages. In the first, we run a real-time 2D pose detector to determine the precise pixel location of important key points of the body. A two-stream neural network is then designed and trained to map detected 2D keypoints into 3D poses. In the second, we deploy the Efficient Neural Architecture Search (ENAS) algorithm to find an optimal network architecture that is used for modeling the Spatio-temporal evolution of the estimated 3D poses via an image-based intermediate representation and performing action recognition. Experiments on Human3.6M, Microsoft Research Redmond (MSR) Action3D, and Stony Brook University (SBU) Kinect Interaction datasets verify the effectiveness of the proposed method on the targeted tasks. Moreover, we show that our method requires a low computational budget for training and inference.

Keywords: human action recognition, pose estimation, D-CNN, deep learning

Procedia PDF Downloads 129
8930 A Method for Compression of Short Unicode Strings

Authors: Masoud Abedi, Abbas Malekpour, Peter Luksch, Mohammad Reza Mojtabaei

Abstract:

The use of short texts in communication has been greatly increasing in recent years. Applying different languages in short texts has led to compulsory use of Unicode strings. These strings need twice the space of common strings, hence, applying algorithms of compression for the purpose of accelerating transmission and reducing cost is worthwhile. Nevertheless, other compression methods like gzip, bzip2 or PAQ due to high overhead data size are not appropriate. The Huffman algorithm is one of the rare algorithms effective in reducing the size of short Unicode strings. In this paper, an algorithm is proposed for compression of very short Unicode strings. At first, every new character to be sent to a destination is inserted in the proposed mapping table. At the beginning, every character is new. In case the character is repeated for the same destination, it is not considered as a new character. Next, the new characters together with the mapping value of repeated characters are arranged through a specific technique and specially formatted to be transmitted. The results obtained from an assessment made on a set of short Persian and Arabic strings indicate that this proposed algorithm outperforms the Huffman algorithm in size reduction.

Keywords: Algorithms, Data Compression, Decoding, Encoding, Huffman Codes, Text Communication

Procedia PDF Downloads 335
8929 Non-Dominated Sorting Genetic Algorithm (NSGA-II) for the Redistricting Problem in Mexico

Authors: Antonin Ponsich, Eric Alfredo Rincon Garcia, Roman Anselmo Mora Gutierrez, Miguel Angel Gutierrez Andrade, Sergio Gerardo De Los Cobos Silva, Pedro Lara Velzquez

Abstract:

The electoral zone design problem consists in redrawing the boundaries of legislative districts for electoral purposes in such a way that federal or state requirements are fulfilled. In Mexico, this process has been historically carried out by the National Electoral Institute (INE), by optimizing an integer nonlinear programming model, in which population equality and compactness of the designed districts are considered as two conflicting objective functions, while contiguity is included as a hard constraint. The solution technique used by the INE is a Simulated Annealing (SA) based algorithm, which handles the multi-objective nature of the problem through an aggregation function. The present work represents the first intent to apply a classical Multi-Objective Evolutionary Algorithm (MOEA), the second version of the Non-dominated Sorting Genetic Algorithm (NSGA-II), to this hard combinatorial problem. First results show that, when compared with the SA algorithm, the NSGA-II obtains promising results. The MOEA manages to produce well-distributed solutions over a wide-spread front, even though some convergence troubles for some instances constitute an issue, which should be corrected in future adaptations of MOEAs to the redistricting problem.

Keywords: multi-objective optimization, NSGA-II, redistricting, zone design problem

Procedia PDF Downloads 353
8928 Application of Hybrid Honey Bees Mating Optimization Algorithm in Multiuser Detection of Wireless Communication Systems

Authors: N. Larbi, F. Debbat

Abstract:

Wireless communication systems have changed dramatically and shown spectacular evolution over the past two decades. These radio technologies are engaged in a quest endless high-speed transmission coupled to a constant need to improve transmission quality. Various radio communication systems being developed use code division multiple access (CDMA) technique. This work analyses a hybrid honey bees mating optimization algorithm (HBMO) applied to multiuser detection (MuD) in CDMA communication systems. The HBMO is a swarm-based optimization algorithm, which simulates the mating process of real honey bees. We apply a hybridization of HBMO with simulated annealing (SA) in order to improve the solution generated by the HBMO. Simulation results show that the detection based on Hybrid HBMO, in term of bit error rate (BER), is viable option when compared with the classic detectors from literature under Rayleigh flat fading channel.

Keywords: BER, DS-CDMA multiuser detection, genetic algorithm, hybrid HBMO, simulated annealing

Procedia PDF Downloads 416