Search results for: statistical machine learning
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3913

Search results for: statistical machine learning

433 A Study of Distinctive Models for Pre-hospital EMS in Thailand: Knowledge Capture

Authors: R. Sinthavalai, N. Memongkol, N. Patthanaprechawong, J. Viriyanantavong, C. Choosuk

Abstract:

In Thailand, the practice of pre-hospital Emergency Medical Service (EMS) in each area reveals the different growth rates and effectiveness of the practices. Those can be found as the diverse quality and quantity. To shorten the learning curve prior to speed-up the practices in other areas, story telling and lessons learnt from the effective practices are valued as meaningful knowledge. To this paper, it was to ascertain the factors, lessons learnt and best practices that have impact as contributing to the success of prehospital EMS system. Those were formulized as model prior to speedup the practice in other areas. To develop the model, Malcolm Baldrige National Quality Award (MBNQA), which is widely recognized as a framework for organizational quality assessment and improvement, was chosen as the discussion framework. Remarkably, this study was based on the consideration of knowledge capture; however it was not to complete the loop of knowledge activities. Nevertheless, it was to highlight the recognition of knowledge capture, which is the initiation of knowledge management.

Keywords: Emergency Medical Service, Modeling, MBNQA, Thailand.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1558
432 An Auxiliary Technique for Coronary Heart Disease Prediction by Analyzing ECG Based on ResNet and Bi-LSTM

Authors: Yang Zhang, Jian He

Abstract:

Heart disease is one of the leading causes of death in the world, and coronary heart disease (CHD) is one of the major heart diseases. Electrocardiogram (ECG) is widely used in the detection of heart diseases, but the traditional manual method for CHD prediction by analyzing ECG requires lots of professional knowledge for doctors. This paper presents sliding window and continuous wavelet transform (CWT) to transform ECG signals into images, and then ResNet and Bi-LSTM are introduced to build the ECG feature extraction network (namely ECGNet). At last, an auxiliary system for CHD prediction was developed based on modified ResNet18 and Bi-LSTM, and the public ECG dataset of CHD from MIMIC-3 was used to train and test the system. The experimental results show that the accuracy of the method is 83%, and the F1-score is 83%. Compared with the available methods for CHD prediction based on ECG, such as kNN, decision tree, VGGNet, etc., this method not only improves the prediction accuracy but also could avoid the degradation phenomenon of the deep learning network.

Keywords: Bi-LSTM, CHD, coronary heart disease, ECG, electrocardiogram, ResNet, sliding window.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 334
431 Estimating Saturated Hydraulic Conductivity from Soil Physical Properties using Neural Networks Model

Authors: B. Ghanbarian-Alavijeh, A.M. Liaghat, S. Sohrabi

Abstract:

Saturated hydraulic conductivity is one of the soil hydraulic properties which is widely used in environmental studies especially subsurface ground water. Since, its direct measurement is time consuming and therefore costly, indirect methods such as pedotransfer functions have been developed based on multiple linear regression equations and neural networks model in order to estimate saturated hydraulic conductivity from readily available soil properties e.g. sand, silt, and clay contents, bulk density, and organic matter. The objective of this study was to develop neural networks (NNs) model to estimate saturated hydraulic conductivity from available parameters such as sand and clay contents, bulk density, van Genuchten retention model parameters (i.e. r θ , α , and n) as well as effective porosity. We used two methods to calculate effective porosity: : (1) eff s FC φ =θ -θ , and (2) inf φ =θ -θ eff s , in which s θ is saturated water content, FC θ is water content retained at -33 kPa matric potential, and inf θ is water content at the inflection point. Total of 311 soil samples from the UNSODA database was divided into three groups as 187 for the training, 62 for the validation (to avoid over training), and 62 for the test of NNs model. A commercial neural network toolbox of MATLAB software with a multi-layer perceptron model and back propagation algorithm were used for the training procedure. The statistical parameters such as correlation coefficient (R2), and mean square error (MSE) were also used to evaluate the developed NNs model. The best number of neurons in the middle layer of NNs model for methods (1) and (2) were calculated 44 and 6, respectively. The R2 and MSE values of the test phase were determined for method (1), 0.94 and 0.0016, and for method (2), 0.98 and 0.00065, respectively, which shows that method (2) estimates saturated hydraulic conductivity better than method (1).

Keywords: Neural network, Saturated hydraulic conductivity, Soil physical properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2557
430 Application of Neural Networks to Predict Changing the Diameters of Bubbles in Pool Boiling Distilled Water

Authors: V. Nikkhah Rashidabad, M. Manteghian, M. Masoumi, S. Mousavian, D. Ashouri

Abstract:

In this research, the capability of neural networks in  modeling and learning complicated and nonlinear relations has been  used to develop a model for the prediction of changes in the diameter  of bubbles in pool boiling distilled water. The input parameters used  in the development of this network include element temperature, heat  flux, and retention time of bubbles. The test data obtained from the  experiment of the pool boiling of distilled water, and the  measurement of the bubbles form on the cylindrical element. The  model was developed based on training algorithm, which is  typologically of back-propagation type. Considering the correlation  coefficient obtained from this model is 0.9633. This shows that this  model can be trusted for the simulation and modeling of the size of  bubble and thermal transfer of boiling.

Keywords: Bubble Diameter, Heat Flux, Neural Network, Training Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1456
429 Reading Strategy Awareness of English Major Students

Authors: Hsin-Yi Lien

Abstract:

The study explored the role of metacognition in foreign language anxiety on a sample of 411 Taiwanese students of English as a Foreign Language. The reading strategy inventory was employed to evaluate the tertiary learners’ level of metacognitive awareness and a semi-structured background questionnaire was also used to examine the learners’ perceptions of their English proficiency and satisfaction of their current English learning. In addition, gender and academic level differences in employment of reading strategies were investigated. The results showed the frequency of reading strategy use increase slightly along with academic years and males and females actually employ different reading strategies. The EFL tertiary learners in the present study utilized cognitive strategies more frequently than metacognitive strategies or support strategies. Male students use metacognitive strategy more often while female students use cognitive and support strategy more frequently.

Keywords: Cognitive strategy, gender differences, metacognitive strategy, support strategy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3427
428 Dengue Disease Mapping with Standardized Morbidity Ratio and Poisson-gamma Model: An Analysis of Dengue Disease in Perak, Malaysia

Authors: N. A. Samat, S. H. Mohd Imam Ma’arof

Abstract:

Dengue disease is an infectious vector-borne viral disease that is commonly found in tropical and sub-tropical regions, especially in urban and semi-urban areas, around the world and including Malaysia. There is no currently available vaccine or chemotherapy for the prevention or treatment of dengue disease. Therefore prevention and treatment of the disease depend on vector surveillance and control measures. Disease risk mapping has been recognized as an important tool in the prevention and control strategies for diseases. The choice of statistical model used for relative risk estimation is important as a good model will subsequently produce a good disease risk map. Therefore, the aim of this study is to estimate the relative risk for dengue disease based initially on the most common statistic used in disease mapping called Standardized Morbidity Ratio (SMR) and one of the earliest applications of Bayesian methodology called Poisson-gamma model. This paper begins by providing a review of the SMR method, which we then apply to dengue data of Perak, Malaysia. We then fit an extension of the SMR method, which is the Poisson-gamma model. Both results are displayed and compared using graph, tables and maps. Results of the analysis shows that the latter method gives a better relative risk estimates compared with using the SMR. The Poisson-gamma model has been demonstrated can overcome the problem of SMR when there is no observed dengue cases in certain regions. However, covariate adjustment in this model is difficult and there is no possibility for allowing spatial correlation between risks in adjacent areas. The drawbacks of this model have motivated many researchers to propose other alternative methods for estimating the risk.

Keywords: Dengue disease, Disease mapping, Standardized Morbidity Ratio, Poisson-gamma model, Relative risk.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3294
427 The Effect of Cyclic Speed on the Wear Properties of Molybdenum Disulfide Greases under Extreme Pressure Loading Using 4 Balls Wear Tests

Authors: Gabi Nehme

Abstract:

The relationship between different types of Molybdenum disulfide greases under extreme pressure loading and different speed situations have been studied using Design of Experiment (DOE) under 1200rpm steady state rotational speed and cyclic frequencies between 2400 and 1200rpm using a Plint machine software to set up the different rotational speed situations.  Research described here is aimed at providing good friction and wear performance while optimizing cyclic frequencies and MoS2 concentration due to the recent concern about grease behavior in extreme pressure applications. Extreme load of 785 Newton was used in conjunction with different cyclic frequencies (2400rpm -3.75min, 1200rpm -7.5min, 2400rpm -3.75min, 1200rpm -7.5min), to examine lithium based grease with and without MoS2 for equal number of revolutions, and a total run of 36000 revolutions; then compared to 1200rpm steady speed for the same total number of revolutions. 4 Ball wear tester was utilized to run large number of experiments randomly selected by the DOE software. The grease was combined with fine grade MoS2 or technical grade then heated to 750C and the wear scar width was collected at the end of each test. DOE model validation results verify that the data were very significant and can be applied to a wide range of extreme pressure applications. Based on simulation results and Scanning Electron images (SEM), it has been found that wear was largely dependent on the cyclic frequency condition. It is believed that technical grade MoS2 greases under faster cyclic speeds perform better and provides antiwear film that can resist extreme pressure loadings. Figures showed reduced wear scars width and improved frictional values.

 

Keywords: MoS2 grease, wear, friction, extreme load, cyclic frequencies, aircraft grade bearing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1898
426 Semantic Enhanced Social Media Sentiments for Stock Market Prediction

Authors: K. Nirmala Devi, V. Murali Bhaskaran

Abstract:

Traditional document representation for classification follows Bag of Words (BoW) approach to represent the term weights. The conventional method uses the Vector Space Model (VSM) to exploit the statistical information of terms in the documents and they fail to address the semantic information as well as order of the terms present in the documents. Although, the phrase based approach follows the order of the terms present in the documents rather than semantics behind the word. Therefore, a semantic concept based approach is used in this paper for enhancing the semantics by incorporating the ontology information. In this paper a novel method is proposed to forecast the intraday stock market price directional movement based on the sentiments from Twitter and money control news articles. The stock market forecasting is a very difficult and highly complicated task because it is affected by many factors such as economic conditions, political events and investor’s sentiment etc. The stock market series are generally dynamic, nonparametric, noisy and chaotic by nature. The sentiment analysis along with wisdom of crowds can automatically compute the collective intelligence of future performance in many areas like stock market, box office sales and election outcomes. The proposed method utilizes collective sentiments for stock market to predict the stock price directional movements. The collective sentiments in the above social media have powerful prediction on the stock price directional movements as up/down by using Granger Causality test.

Keywords: Bag of Words, Collective Sentiments, Ontology, Semantic relations, Sentiments, Social media, Stock Prediction, Twitter, Vector Space Model and wisdom of crowds.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2800
425 Statistical Analysis and Optimization of a Process for CO2 Capture

Authors: Muftah H. El-Naas, Ameera F. Mohammad, Mabruk I. Suleiman, Mohamed Al Musharfy, Ali H. Al-Marzouqi

Abstract:

CO2 capture and storage technologies play a significant role in contributing to the control of climate change through the reduction of carbon dioxide emissions into the atmosphere. The present study evaluates and optimizes CO2 capture through a process, where carbon dioxide is passed into pH adjusted high salinity water and reacted with sodium chloride to form a precipitate of sodium bicarbonate. This process is based on a modified Solvay process with higher CO2 capture efficiency, higher sodium removal, and higher pH level without the use of ammonia. The process was tested in a bubble column semi-batch reactor and was optimized using response surface methodology (RSM). CO2 capture efficiency and sodium removal were optimized in terms of major operating parameters based on four levels and variables in Central Composite Design (CCD). The operating parameters were gas flow rate (0.5–1.5 L/min), reactor temperature (10 to 50 oC), buffer concentration (0.2-2.6%) and water salinity (25-197 g NaCl/L). The experimental data were fitted to a second-order polynomial using multiple regression and analyzed using analysis of variance (ANOVA). The optimum values of the selected variables were obtained using response optimizer. The optimum conditions were tested experimentally using desalination reject brine with salinity ranging from 65,000 to 75,000 mg/L. The CO2 capture efficiency in 180 min was 99% and the maximum sodium removal was 35%. The experimental and predicted values were within 95% confidence interval, which demonstrates that the developed model can successfully predict the capture efficiency and sodium removal using the modified Solvay method.

Keywords: Bubble column reactor, CO2 capture, Response Surface Methodology, water desalination.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1844
424 Director Compensation, CEO Duality, State Ownership, and Firm Performance in China: Proof from Panel Data of Publicly Listed Enterprises from 1999 to 2020

Authors: Wanda Luen-Wun Siu, Xiaowen Zhang

Abstract:

This paper offered the primary methodical proof on how director remuneration related to enterprise earnings in listed firms in China in light of most evidence focusing on cross-sectional data or data in a short span of time. Using full economic and business panel data on China’s publicly listed enterprise from 1999 to 2020 over two decades in the China Stock Market & Accounting Research database, we found statistically significant positive associations between director pay and firm performance in privately owned firms over this period, supporting the agency theory. In contrast, among the state-owned enterprises, there was a reverse relation between director compensation and firm financial performance, contributing to the existing literature. But the results also revealed that state-owned enterprises financially performed as well as private enterprises. Such findings suggested that state ownership might line up officials’ career incentives with party prime concern rather than pecuniary incentives. Also, CEO duality enhanced firm performance. As such, allegiance to the party and possible advancement to an upper-level political position would motivate company directors in state-owned enterprises. On the other hand, directors in privately owned enterprises might be motivated by monetary incentives. In addition, a statistical regression model was proposed and tested to get the results of the performance of state-owned enterprises. Finally, some suggestions were made about how to improve the institutional management of government-owned corporations in China.

Keywords: China’s listed Firm, director compensation, CEO duality, firm performance, panel analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 498
423 The Strategies for Teaching Digital Art in the Classroom as a Way of Enhancing Pupils’ Artistic Creativity

Authors: Aber Salem Aboalgasm, Rupert Ward

Abstract:

Teaching art by digital means is a big challenge for the majority of teachers of art and design in primary schools, yet it allows relationships between art, technology and creativity to be clearly identified. The aim of this article is to present a modern way of teaching art, using digital tools in the art classroom to improve creative ability in pupils aged between nine and eleven years. It also presents a conceptual model for creativity based on digital art. The model could be useful for pupils interested in learning to draw by using an e-drawing package, and for teachers who are interested in teaching modern digital art in order to improve children’s creativity. By illustrating the strategy of teaching art through technology, this model may also help education providers to make suitable choices about which technological approaches are most effective in enhancing students’ creative ability, and which digital art tools can benefit children by developing their technical skills. It is also expected that use of this model will help to develop skills of social interaction, which may in turn improve intellectual ability.

Keywords: Digital tools, motivation, creative activity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3102
422 Revising the Student Experiment Materials and Practices at the National University of Laos

Authors: Syhalath Xaphakdy, Toshio Nagata, Saykham Phommathat, Pavy Souwannavong, Vilayvanh Srithilat, Phoxay Sengdala, Bounaom Phetarnousone, Boualay Siharath, Xaya Chemcheng

Abstract:

The National University of Laos (NUOL) invited a group of volunteers from the Japan International Cooperation Agency (JICA) to revise the physics experiments to utilize the materials that were already available to students. The intension was to review and revise the materials regularly utilized in physics class. The project had access to limited materials and a small budget for the class in the unit; however, by developing experimental textbooks related to mechanics, electricity, and wave and vibration, the group found a way to apply them in the classroom and enhance the students teaching activities. The aim was to introduce a way to incorporate the materials and practices in the classroom to enhance the students learning and teaching skills, particularly when they graduate and begin working as high school teachers.

Keywords: NUOL, JICA, physics experiment materials, small budget, mechanics, electricity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1246
421 Using Serious Games to Improve the Preparation of Pre-Service Teachers in Bulgaria

Authors: Rumyana Peytcheva-Forsyth, Blagovesna Yovkova

Abstract:

This paper presents the outcomes of a qualitative study which aims to investigate the pedagogical potentials of serious games in the preparation of future teachers. The authors discuss the existing problems and barriers associated with the organization of teaching practices in Bulgaria as part of the pre-service teacher training, as well as the attitudes and perceptions of the interviewed academics, teachers and trainees concerning the integration of serious games in the teaching practicum. The study outcomes strongly confirm the positive attitudes of the respondents to the introduction of virtual learning environments for the development of professional skills of future teachers as a supplement to the traditional forms of education. Through the inclusion of serious games it is expected to improve the quality of practical training of pre-service teachers as they overcome many of the problems identified in the existing teaching practices. The outcomes of the study will inform the design of the educational simulation software which is part of the project SimAula Tomorrow's Teachers Training.

Keywords: pre-service teacher training, serious games, virtual practicum, simulations

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672
420 Ordinal Regression with Fenton-Wilkinson Order Statistics: A Case Study of an Orienteering Race

Authors: Joonas Pääkkönen

Abstract:

In sports, individuals and teams are typically interested in final rankings. Final results, such as times or distances, dictate these rankings, also known as places. Places can be further associated with ordered random variables, commonly referred to as order statistics. In this work, we introduce a simple, yet accurate order statistical ordinal regression function that predicts relay race places with changeover-times. We call this function the Fenton-Wilkinson Order Statistics model. This model is built on the following educated assumption: individual leg-times follow log-normal distributions. Moreover, our key idea is to utilize Fenton-Wilkinson approximations of changeover-times alongside an estimator for the total number of teams as in the notorious German tank problem. This original place regression function is sigmoidal and thus correctly predicts the existence of a small number of elite teams that significantly outperform the rest of the teams. Our model also describes how place increases linearly with changeover-time at the inflection point of the log-normal distribution function. With real-world data from Jukola 2019, a massive orienteering relay race, the model is shown to be highly accurate even when the size of the training set is only 5% of the whole data set. Numerical results also show that our model exhibits smaller place prediction root-mean-square-errors than linear regression, mord regression and Gaussian process regression.

Keywords: Fenton-Wilkinson approximation, German tank problem, log-normal distribution, order statistics, ordinal regression, orienteering, sports analytics, sports modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 833
419 An Efficient Approach for Shear Behavior Definition of Plant Stalk

Authors: M. R. Kamandar, J. Massah

Abstract:

The information of the impact cutting behavior of plants stalk plays an important role in the design and fabrication of plants cutting equipment. It is difficult to investigate a theoretical method for defining cutting properties of plants stalks because the cutting process is complex. Thus, it is necessary to set up an experimental approach to determine cutting parameters for a single stalk. To measure the shear force, shear energy and shear strength of plant stalk, a special impact cutting tester was fabricated. It was similar to an Izod impact cutting tester for metals but a cutting blade and data acquisition system were attached to the end of pendulum's arm. The apparatus was included four strain gages and a digital indicator to show the real-time cutting force of plant stalk. To measure the shear force and also testing the apparatus, two plants’ stalks, like buxus and privet, were selected. The samples (buxus and privet stalks) were cut under impact cutting process at four loading rates 1, 2, 3 and 4 m.s-1 and three internodes fifth, tenth and fifteenth by the apparatus. At buxus cutting analysis: the minimum value of cutting energy was obtained at fifth internode and loading rate 4 m.s-1 and the maximum value of shear energy was obtained at fifteenth internode and loading rate 1 m.s-1. At privet cutting analysis: the minimum value of shear consumption energy was obtained at fifth internode and loading rate: 4 m.s-1 and the maximum value of shear energy was obtained at fifteenth internode and loading rate: 1 m.s-1. The statistical analysis at both plants showed that the increase of impact cutting speed would decrease the shear consumption energy and shear strength. In two scenarios, the results showed that with increase the cutting speed, shear force would decrease.

Keywords: Buxus, privet, impact cutting, shear energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 829
418 Evaluating the Validity of Computational Fluid Dynamics Model of Dispersion in a Complex Urban Geometry Using Two Sets of Experimental Measurements

Authors: Mohammad R. Kavian Nezhad, Carlos F. Lange, Brian A. Fleck

Abstract:

This research presents the validation study of a computational fluid dynamics (CFD) model developed to simulate the scalar dispersion emitted from rooftop sources around the buildings at the University of Alberta North Campus. The ANSYS CFX code was used to perform the numerical simulation of the wind regime and pollutant dispersion by solving the 3D steady Reynolds-averaged Navier-Stokes (RANS) equations on a building-scale high-resolution grid. The validation study was performed in two steps. First, the CFD model performance in 24 cases (eight wind directions and three wind speeds) was evaluated by comparing the predicted flow fields with the available data from the previous measurement campaign designed at the North Campus, using the standard deviation method (SDM), while the estimated results of the numerical model showed maximum average percent errors of approximately 53% and 37% for wind incidents from the North and Northwest, respectively. Good agreement with the measurements was observed for the other six directions, with an average error of less than 30%. In the second step, the reliability of the implemented turbulence model, numerical algorithm, modeling techniques, and the grid generation scheme was further evaluated using the Mock Urban Setting Test (MUST) dispersion dataset. Different statistical measures, including the fractional bias (FB), the mean geometric bias (MG), and the normalized mean square error (NMSE), were used to assess the accuracy of the predicted dispersion field. Our CFD results are in very good agreement with the field measurements.

Keywords: CFD, plume dispersion, complex urban geometry, validation study, wind flow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 373
417 Generating Qualitative Causal Graph using Modeling Constructs of Qualitative Process Theory for Explaining Organic Chemistry Reactions

Authors: Alicia Y. C. Tang, Rukaini Abdullah, Sharifuddin M. Zain, Noorsaadah A. Rahman

Abstract:

This paper discusses the causal explanation capability of QRIOM, a tool aimed at supporting learning of organic chemistry reactions. The development of the tool is based on the hybrid use of Qualitative Reasoning (QR) technique and Qualitative Process Theory (QPT) ontology. Our simulation combines symbolic, qualitative description of relations with quantity analysis to generate causal graphs. The pedagogy embedded in the simulator is to both simulate and explain organic reactions. Qualitative reasoning through a causal chain will be presented to explain the overall changes made on the substrate; from initial substrate until the production of final outputs. Several uses of the QPT modeling constructs in supporting behavioral and causal explanation during run-time will also be demonstrated. Explaining organic reactions through causal graph trace can help improve the reasoning ability of learners in that their conceptual understanding of the subject is nurtured.

Keywords: Qualitative reasoning, causal graph, organicreactions, explanation, QPT, modeling constructs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1419
416 Task-Based Language Teaching: A Paradigm Shift in ESL/EFL Teaching and Learning: A Case Study-Based Approach

Authors: Zehra Sultan

Abstract:

The study is based on the Task-based Language Teaching (TBLT) approach which is found to be very effective in the EFL/ESL classroom. This approach engages learners to acquire the usage of authentic language skills by interacting with the real world through a sequence of pedagogical tasks. The use of technology enhances the effectiveness of this approach. This study throws light on the historical background of TBLT, and its efficacy in the EFL /ESL classroom. In addition, this study precisely talks about the implementation of this approach in the General Foundation Program (GFP) of Muscat College, Oman. It furnishes the list of the pedagogical tasks embedded in the language curriculum of the GFP which are skillfully allied to the College graduate attributes. Moreover, the study also discusses the challenges pertaining to this approach from the point of view of teachers, students and its classroom application. Additionally, the operational success of this methodology is gauged through formative assessments of the GFP which is apparent in the students’ progress.

Keywords: Task-based language teaching, authentic language, communicative approach, real world activities, ESL/EFL activities.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 954
415 Roadmapping as a Collaborative Strategic Decision-Making Process: Shaping Social Dialogue Options for the European Banking Sector

Authors: Christos A. Ioannou, Panagiotis Panagiotopoulos, Lampros Stergioulas

Abstract:

The new status generated by technological advancements and changes in the global economy raises important issues on how communities and organisations need to innovate upon their traditional processes in order to adapt to the challenges of the Knowledge Society. The DialogoS+ European project aims to study the role of and promote social dialogue in the banking sector, strengthen the link between old and new members and make social dialogue at the European level a force for innovation and change, also given the context of the international crisis emerging in 2008- 2009. Under the scope of DialogoS+, this paper describes how the community of Europe-s banking sector trade unions attempted to adapt to the challenges of the Knowledge Society by exploiting the benefits of new channels of communication, learning, knowledge generation and diffusion focusing on the concept of roadmapping. Important dimensions of social dialogue such as collective bargaining and working conditions are addressed.

Keywords: Banking sector, knowledge society, road mapping, social dialogue.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2131
414 Predicting Extrusion Process Parameters Using Neural Networks

Authors: Sachin Man Bajimaya, SangChul Park, Gi-Nam Wang

Abstract:

The objective of this paper is to estimate realistic principal extrusion process parameters by means of artificial neural network. Conventionally, finite element analysis is used to derive process parameters. However, the finite element analysis of the extrusion model does not consider the manufacturing process constraints in its modeling. Therefore, the process parameters obtained through such an analysis remains highly theoretical. Alternatively, process development in industrial extrusion is to a great extent based on trial and error and often involves full-size experiments, which are both expensive and time-consuming. The artificial neural network-based estimation of the extrusion process parameters prior to plant execution helps to make the actual extrusion operation more efficient because more realistic parameters may be obtained. And so, it bridges the gap between simulation and real manufacturing execution system. In this work, a suitable neural network is designed which is trained using an appropriate learning algorithm. The network so trained is used to predict the manufacturing process parameters.

Keywords: Artificial Neural Network (ANN), Indirect Extrusion, Finite Element Analysis, MES.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2367
413 A Quantitative Study Identifying the Prevalence of Anxiety in Dyslexic Students in Higher Education

Authors: Amanda Abbott-Jones

Abstract:

Adult students with dyslexia in higher education can receive support for their cognitive needs but may also experience negative emotion such as anxiety due to their dyslexia in connection with their studies. This paper aims to test the hypothesis that adult dyslexic learners have a higher prevalence of academic and social anxiety than their non-dyslexic peers. A quantitative approach was used to measure differences in academic and social anxiety between 102 students with a formal diagnosis of dyslexia compared to 72 students with no history of learning difficulties. Academic and social anxiety was measured in a questionnaire based on the State-Trait Anxiety Inventory. Findings showed that dyslexic students showed statistically significant higher levels of academic, but not social anxiety in comparison to the non-dyslexic sample. Dyslexic students in higher education show academic anxiety levels that are well above what is shown by students without dyslexia. The implications of this for the dyslexia practitioner is that delivery of strategies to deal with anxiety should be seen equally as important, if not more so, than interventions to deal with cognitive difficulties.

Keywords: Academic, anxiety, dyslexia, quantitative, students, university.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1325
412 Development of Innovative Islamic Web Applications

Authors: Farrukh Shahzad

Abstract:

The rich Islamic resources related to religious text, Islamic sciences, and history are widely available in print and in electronic format online. However, most of these works are only available in Arabic language. In this research, an attempt is made to utilize these resources to create interactive web applications in Arabic, English and other languages. The system utilizes the Pattern Recognition, Knowledge Management, Data Mining, Information Retrieval and Management, Indexing, storage and data-analysis techniques to parse, store, convert and manage the information from authentic Arabic resources. These interactive web Apps provide smart multi-lingual search, tree based search, on-demand information matching and linking. In this paper, we provide details of application architecture, design, implementation and technologies employed. We also presented the summary of web applications already developed. We have also included some screen shots from the corresponding web sites. These web applications provide an Innovative On-line Learning Systems (eLearning and computer based education).

Keywords: Islamic resources, Muslim scholars, hadith, narrators, history, fiqh.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1302
411 Detecting Remote Protein Evolutionary Relationships via String Scoring Method

Authors: Nazar Zaki, Safaai Deris

Abstract:

The amount of the information being churned out by the field of biology has jumped manifold and now requires the extensive use of computer techniques for the management of this information. The predominance of biological information such as protein sequence similarity in the biological information sea is key information for detecting protein evolutionary relationship. Protein sequence similarity typically implies homology, which in turn may imply structural and functional similarities. In this work, we propose, a learning method for detecting remote protein homology. The proposed method uses a transformation that converts protein sequence into fixed-dimensional representative feature vectors. Each feature vector records the sensitivity of a protein sequence to a set of amino acids substrings generated from the protein sequences of interest. These features are then used in conjunction with support vector machines for the detection of the protein remote homology. The proposed method is tested and evaluated on two different benchmark protein datasets and it-s able to deliver improvements over most of the existing homology detection methods.

Keywords: Protein homology detection; support vectormachine; string kernel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1392
410 Normalizing Flow to Augmented Posterior: Conditional Density Estimation with Interpretable Dimension Reduction for High Dimensional Data

Authors: Cheng Zeng, George Michailidis, Hitoshi Iyatomi, Leo L Duan

Abstract:

The conditional density characterizes the distribution of a response variable y given other predictor x, and plays a key role in many statistical tasks, including classification and outlier detection. Although there has been abundant work on the problem of Conditional Density Estimation (CDE) for a low-dimensional response in the presence of a high-dimensional predictor, little work has been done for a high-dimensional response such as images. The promising performance of normalizing flow (NF) neural networks in unconditional density estimation acts a motivating starting point. In this work, we extend NF neural networks when external x is present. Specifically, they use the NF to parameterize a one-to-one transform between a high-dimensional y and a latent z that comprises two components [zP , zN]. The zP component is a low-dimensional subvector obtained from the posterior distribution of an elementary predictive model for x, such as logistic/linear regression. The zN component is a high-dimensional independent Gaussian vector, which explains the variations in y not or less related to x. Unlike existing CDE methods, the proposed approach, coined Augmented Posterior CDE (AP-CDE), only requires a simple modification on the common normalizing flow framework, while significantly improving the interpretation of the latent component, since zP represents a supervised dimension reduction. In image analytics applications, AP-CDE shows good separation of x-related variations due to factors such as lighting condition and subject id, from the other random variations. Further, the experiments show that an unconditional NF neural network, based on an unsupervised model of z, such as Gaussian mixture, fails to generate interpretable results.

Keywords: Conditional density estimation, image generation, normalizing flow, supervised dimension reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 166
409 Clinical Parameters Response to Low-Level Laser versus Monochromatic Near-Infrared Photo Energy in Diabetic Patients with Peripheral Neuropathy

Authors: Abeer A. Abdelhamed

Abstract:

Background: Diabetic sensorimotor polyneuropathy (DSP) is one of the most common microvascular complications of type 2 diabetes. Loss of sensation is thought to contribute to a lack of static and dynamic stability and increased risk of falling. Purpose: The purpose of this study was to compare the effects of low-level laser (LLL) and monochromatic near-infrared photo energy (MIRE) on pain, cutaneous sensation, static stability, and index of lower limb blood flow in diabetic patients with peripheral neuropathy. Methods: Forty diabetic patients with peripheral neuropathy were recruited for participation in this study. They were divided into two groups: The MIRE group, which contained 20 patients, and the LLL group, which contained 20 patients. All patients who participated in the study had been subjected to various physical assessment procedures, including pain, cutaneous sensation, Doppler flow meter, and static stability assessments. The baseline measurements were followed by treatment sessions that were conducted twice a week for six successive weeks. Results: The statistical analysis of the data revealed significant improvement of pain in both groups, with significant improvement in cutaneous sensation and static balance in the MIRE group compared to the LLL group; on the other hand, the results showed no significant differences in lower limb blood flow between the groups. Conclusion: LLL and MIRE can improve painful symptoms in patients with diabetic neuropathy. On the other hand, MIRE is also useful in improving cutaneous sensation and static stability in patients with diabetic neuropathy.

Keywords: Diabetic neuropathy, Doppler flow meter, –Lowlevel laser, Monochromatic near-infrared photo energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1887
408 Digital Content Strategy: Detailed Review of the Key Content Components

Authors: Oksana Razina, Shakeel Ahmad, Jessie Qun Ren, Olufemi Isiaq

Abstract:

The modern life of businesses is categorically reliant on their established position online, where digital (and particularly website) content plays a significant role as the first point of information. Digital content, therefore, becomes essential – from making the first impression through to the building and development of client relationships. Despite a number of valuable papers suggesting a strategic approach when dealing with digital data, other sources often do not view or accept the approach to digital content as a holistic or continuous process. Associations are frequently made with merely a one-off marketing campaign or similar. The challenge is in establishing an agreed definition for the notion of Digital Content Strategy (DCS), which currently does not exist, as it is viewed from an excessive number of angles. A strategic approach to content, nonetheless, is required, both practically and contextually. We, therefore, aimed at attempting to identify the key content components, comprising a DCS, to ensure all the aspects were covered and strategically applied – from the company’s understanding of the content value to the ability to display flexibility of content and advances in technology. This conceptual project evaluated existing literature on the topic of DCS and related aspects, using PRISMA Systematic Review Method, Document Analysis, Inclusion and Exclusion Criteria, Scoping Review, Snow-Balling Technique and Thematic Analysis. The data were collected from academic and statistical sources, government and relevant trade publications. Based on the suggestions from academics and trading sources, related to the issues discussed, we revealed the key actions for content creation and attempted to define the notion of DCS. The major finding of the study presented Key Content Components of DCS and can be considered for implementation in a business retail setting.

Keywords: Digital content strategy, digital marketing strategy, key content components, websites.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 228
407 Digital Content Strategy: Detailed Review of the Key Content Components

Authors: Oksana Razina, Shakeel Ahmad, Jessie Qun Ren, Olufemi Isiaq

Abstract:

The modern life of businesses is categorically reliant on their established position online, where digital (and particularly website) content plays a significant role as the first point of information. Digital content, therefore, becomes essential – from making the first impression through to the building and development of client relationships. Despite a number of valuable papers suggesting a strategic approach when dealing with digital data, other sources often do not view or accept the approach to digital content as a holistic or continuous process. Associations are frequently made with merely a one-off marketing campaign or similar. The challenge is in establishing an agreed definition for the notion of Digital Content Strategy (DCS), which currently does not exist, as it is viewed from an excessive number of angles. A strategic approach to content, nonetheless, is required, both practically and contextually. We, therefore, aimed at attempting to identify the key content components, comprising a DCS, to ensure all the aspects were covered and strategically applied – from the company’s understanding of the content value to the ability to display flexibility of content and advances in technology. This conceptual project evaluated existing literature on the topic of DCS and related aspects, using PRISMA Systematic Review Method, Document Analysis, Inclusion and Exclusion Criteria, Scoping Review, Snow-Balling Technique and Thematic Analysis. The data were collected from academic and statistical sources, government and relevant trade publications. Based on the suggestions from academics and trading sources, related to the issues discussed, we revealed the key actions for content creation and attempted to define the notion of DCS. The major finding of the study presented Key Content Components of DCS and can be considered for implementation in a business retail setting.

Keywords: Digital content strategy, digital marketing strategy, key content components, websites.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 214
406 Use of Corn Stover for the Production of 2G Bioethanol, Enzymes and Xylitol under a Biorefinery Concept

Authors: Astorga-Trejo Rebeca, Fonseca-Peralta Héctor Manuel, Beltrán-Arredondo Laura Ivonne, Castro-Martínez Claudia

Abstract:

The use of biomass as feedstock for the production of fuels and other chemicals of interest is an ever growing accepted option in the way to the development of biorefinery complexes. In the Mexican state of Sinaloa, a significant amount of residues from corn crops are produced every year, most of which can be converted to bioethanol and other products through biotechnological conversion using yeast and other microorganisms. Therefore, the objective of this work was to take advantage of corn stover and evaluate its potential as a substrate for the production of second generation bioethanol (2G), enzymes and xylitol. To produce bioethanol 2G, an acid-alkaline pretreatment was carried out prior to saccharification and fermentation. The microorganisms used for the production of enzymes, as well as for the production of xylitol, were isolated and characterized in our work group. Statistical analysis was performed using Design Expert version 11.0. The results showed that it is possible to obtain 2G bioethanol employing corn stover as a carbon source and Saccharomyces cerevisiae ItVer01 and Candida intermedia CBE002 with yields of 0.42 g and 0.31 g, respectively. It was also shown that C. intermedia has the ability to produce xylitol with a good yield (0.46 g/g). On the other hand, qualitative and quantitative studies showed that the native strains of Fusarium equiseti (0.4 IU/mL - xylanase), Bacillus velezensis (1.2 IU/mL – xylanase and 0.4 UI/mL - amylase) and Penicillium funiculosum (1.5 IU/mL - cellulases) have the capacity to produce xylanases, amylases or cellulases using corn stover as raw material. This study allowed us to demonstrate that it is possible to use corn stover as a carbon source, a low-cost raw material with high availability in our country, to obtain bioproducts of industrial interest, using processes that are more environmentally friendly and sustainable. It is necessary to continue the optimization of each bioprocess.

Keywords: Biomass, corn stover, biorefinery, bioethanol 2G, enzymes, xylitol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 474
405 Using the Monte Carlo Simulation to Predict the Assembly Yield

Authors: C. Chahin, M. C. Hsu, Y. H. Lin, C. Y. Huang

Abstract:

Electronics Products that achieve high levels of integrated communications, computing and entertainment, multimedia features in small, stylish and robust new form factors are winning in the market place. Due to the high costs that an industry may undergo and how a high yield is directly proportional to high profits, IC (Integrated Circuit) manufacturers struggle to maximize yield, but today-s customers demand miniaturization, low costs, high performance and excellent reliability making the yield maximization a never ending research of an enhanced assembly process. With factors such as minimum tolerances, tighter parameter variations a systematic approach is needed in order to predict the assembly process. In order to evaluate the quality of upcoming circuits, yield models are used which not only predict manufacturing costs but also provide vital information in order to ease the process of correction when the yields fall below expectations. For an IC manufacturer to obtain higher assembly yields all factors such as boards, placement, components, the material from which the components are made of and processes must be taken into consideration. Effective placement yield depends heavily on machine accuracy and the vision of the system which needs the ability to recognize the features on the board and component to place the device accurately on the pads and bumps of the PCB. There are currently two methods for accurate positioning, using the edge of the package and using solder ball locations also called footprints. The only assumption that a yield model makes is that all boards and devices are completely functional. This paper will focus on the Monte Carlo method which consists in a class of computational algorithms (information processed algorithms) which depends on repeated random samplings in order to compute the results. This method utilized in order to recreate the simulation of placement and assembly processes within a production line.

Keywords: Monte Carlo simulation, placement yield, PCBcharacterization, electronics assembly

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2166
404 A Retrospective Cross-Sectional Study on the Prevalence and Factors Associated with Virological Non-Suppression among HIV-Positive Adult Patients on Antiretroviral Therapy in Woliso Town, Oromia, Ethiopia

Authors: Teka Haile, Behailu Hawulte, Solomon Alemayehu

Abstract:

Background: HIV virological failure still remains a problem in HV/AIDS treatment and care. This study aimed to describe the prevalence and identify the factors associated with viral non-suppression among HIV-positive adult patients on antiretroviral therapy in Woliso Town, Oromia, Ethiopia. Methods: A retrospective cross-sectional study was conducted among 424 HIV-positive patient’s attending antiretroviral therapy (ART) in Woliso Town during the period from August 25, 2020 to August 30, 2020. Data collected from patient medical records were entered into Epi Info version 2.3.2.1 and exported to SPSS version 21.0 for analysis. Logistic regression analysis was done to identify factors associated with viral load non-suppression, and statistical significance of odds ratios were declared using 95% confidence interval and p-value < 0.05. Results: A total of 424 patients were included in this study. The mean age (± SD) of the study participants was 39.88 (± 9.995) years. The prevalence of HIV viral load non-suppression was 55 (13.0%) with 95% CI (9.9-16.5). Second-line ART treatment regimen (Adjusted Odds Ratio (AOR) = 8.98, 95% Confidence Interval (CI): 2.64, 30.58) and routine viral load testing (AOR = 0.01, 95% CI: 0.001, 0.02) were significantly associated with virological non-suppression. Conclusion: Virological non-suppression was high, which hinders the achievement of the third global 95 target. The second-line regimen and routine viral load testing were significantly associated with virological non-suppression. It suggests the need to assess the effectiveness of antiretroviral drugs for epidemic control. It also clearly shows the need to decentralize third-line ART treatment for those patients in need.

Keywords: Virological non-suppression, HIV-positive, ART, Woliso Town, Ethiopia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 587