Search results for: language error
2708 The Effects of Prosthetic Leg Stiffness on Gait, Comfort, and Satisfaction: A Review of Mechanical Engineering Approaches
Authors: Kourosh Fatehi, Niloofar Hanafi
Abstract:
One of the challenges in providing optimal prosthetic legs for lower limb amputees is to select the appropriate foot stiffness that suits their individual needs and preferences. Foot stiffness affects various aspects of walking, such as stability, comfort, and energy expenditure. However, the current prescription process is largely based on trial-and-error, manufacturer recommendations, or clinician judgment, which may not reflect the prosthesis user’s subjective experience or psychophysical sensitivity. Therefore, there is a need for more scientific and technological tools to measure and understand how prosthesis users perceive and prefer different foot stiffness levels, and how this preference relates to clinical outcomes. This review covers how to measure and design lower leg prostheses based on user preference and foot stiffness. It also explores how these factors affect walking outcomes and quality of life, and identifies the current challenges and gaps in this field from a mechanical engineering standpoint.Keywords: perception, preference, prosthetics, stiffness
Procedia PDF Downloads 862707 Quantitative, Preservative Methodology for Review of Interview Transcripts Using Natural Language Processing
Authors: Rowan P. Martnishn
Abstract:
During the execution of a National Endowment of the Arts grant, approximately 55 interviews were collected from professionals across various fields. These interviews were used to create deliverables – historical connections for creations that began as art and evolved entirely into computing technology. With dozens of hours’ worth of transcripts to be analyzed by qualitative coders, a quantitative methodology was created to sift through the documents. The initial step was to both clean and format all the data. First, a basic spelling and grammar check was applied, as well as a Python script for normalized formatting which used an open-source grammatical formatter to make the data as coherent as possible. 10 documents were randomly selected to manually review, where words often incorrectly translated during the transcription were recorded and replaced throughout all other documents. Then, to remove all banter and side comments, the transcripts were spliced into paragraphs (separated by change in speaker) and all paragraphs with less than 300 characters were removed. Secondly, a keyword extractor, a form of natural language processing where significant words in a document are selected, was run on each paragraph for all interviews. Every proper noun was put into a data structure corresponding to that respective interview. From there, a Bidirectional and Auto-Regressive Transformer (B.A.R.T.) summary model was then applied to each paragraph that included any of the proper nouns selected from the interview. At this stage the information to review had been sent from about 60 hours’ worth of data to 20. The data was further processed through light, manual observation – any summaries which proved to fit the criteria of the proposed deliverable were selected, as well their locations within the document. This narrowed that data down to about 5 hours’ worth of processing. The qualitative researchers were then able to find 8 more connections in addition to our previous 4, exceeding our minimum quota of 3 to satisfy the grant. Major findings of the study and subsequent curation of this methodology raised a conceptual finding crucial to working with qualitative data of this magnitude. In the use of artificial intelligence there is a general trade off in a model between breadth of knowledge and specificity. If the model has too much knowledge, the user risks leaving out important data (too general). If the tool is too specific, it has not seen enough data to be useful. Thus, this methodology proposes a solution to this tradeoff. The data is never altered outside of grammatical and spelling checks. Instead, the important information is marked, creating an indicator of where the significant data is without compromising the purity of it. Secondly, the data is chunked into smaller paragraphs, giving specificity, and then cross-referenced with the keywords (allowing generalization over the whole document). This way, no data is harmed, and qualitative experts can go over the raw data instead of using highly manipulated results. Given the success in deliverable creation as well as the circumvention of this tradeoff, this methodology should stand as a model for synthesizing qualitative data while maintaining its original form.Keywords: B.A.R.T.model, keyword extractor, natural language processing, qualitative coding
Procedia PDF Downloads 342706 Hybrid Subspace Approach for Time Delay Estimation in MIMO Systems
Authors: Mojtaba Saeedinezhad, Sarah Yousefi
Abstract:
In this paper, we present a hybrid subspace approach for Time Delay Estimation (TDE) in multivariable systems. While several methods have been proposed for time delay estimation in SISO systems, delay estimation in MIMO systems were always a big challenge. In these systems the existing TDE methods have significant limitations because most of procedures are just based on system response estimation or correlation analysis. We introduce a new hybrid method for TDE in MIMO systems based on subspace identification and explicit output error method; and compare its performance with previously introduced procedures in presence of different noise levels and in a statistical manner. Then the best method is selected with multi objective decision making technique. It is shown that the performance of new approach is much better than the existing methods, even in low signal-to-noise conditions.Keywords: system identification, time delay estimation, ARX, OE, merit ratio, multi variable decision making
Procedia PDF Downloads 3492705 Improved Pitch Detection Using Fourier Approximation Method
Authors: Balachandra Kumaraswamy, P. G. Poonacha
Abstract:
Automatic Music Information Retrieval has been one of the challenging topics of research for a few decades now with several interesting approaches reported in the literature. In this paper we have developed a pitch extraction method based on a finite Fourier series approximation to the given window of samples. We then estimate pitch as the fundamental period of the finite Fourier series approximation to the given window of samples. This method uses analysis of the strength of harmonics present in the signal to reduce octave as well as harmonic errors. The performance of our method is compared with three best known methods for pitch extraction, namely, Yin, Windowed Special Normalization of the Auto-Correlation Function and Harmonic Product Spectrum methods of pitch extraction. Our study with artificially created signals as well as music files show that Fourier Approximation method gives much better estimate of pitch with less octave and harmonic errors.Keywords: pitch, fourier series, yin, normalization of the auto- correlation function, harmonic product, mean square error
Procedia PDF Downloads 4152704 Low-Cost Embedded Biometric System Based on Fingervein Modality
Authors: Randa Boukhris, Alima Damak, Dorra Sellami
Abstract:
Fingervein biometric authentication is one of the most popular and accurate technologies. However, low cost embedded solution is still an open problem. In this paper, a real-time implementation of fingervein recognition process embedded in Raspberry-Pi has been proposed. The use of Raspberry-Pi reduces overall system cost and size while allowing an easy user interface. Implementation of a target technology has guided to opt some specific parallel and simple processing algorithms. In the proposed system, we use four structural directional kernel elements for filtering finger vein images. Then, a Top-Hat and Bottom-Hat kernel filters are used to enhance the visibility and the appearance of venous images. For feature extraction step, a simple Local Directional Code (LDC) descriptor is applied. The proposed system presents an Error Equal Rate (EER) and Identification Rate (IR), respectively, equal to 0.02 and 98%. Furthermore, experimental results show that real-time operations have good performance.Keywords: biometric, Bottom-Hat, Fingervein, LDC, Rasberry-Pi, ROI, Top-Hat
Procedia PDF Downloads 2072703 The Impact of Artificial Intelligence on Qualty Conrol and Quality
Authors: Mary Moner Botros Fanawel
Abstract:
Many companies use the statistical tool named as statistical quality control, and which can have a high cost for the companies interested on these statistical tools. The evaluation of the quality of products and services is an important topic, but the reduction of the cost of the implantation of the statistical quality control also has important benefits for the companies. For this reason, it is important to implement a economic design for the various steps included into the statistical quality control. In this paper, we describe some relevant aspects related to the economic design of a quality control chart for the proportion of defective items. They are very important because the suggested issues can reduce the cost of implementing a quality control chart for the proportion of defective items. Note that the main purpose of this chart is to evaluate and control the proportion of defective items of a production process.Keywords: model predictive control, hierarchical control structure, genetic algorithm, water quality with DBPs objectives proportion, type I error, economic plan, distribution function bootstrap control limit, p-value method, out-of-control signals, p-value, quality characteristics
Procedia PDF Downloads 662702 Exploring Academic Writing Challenges of First Year English as an Additional Language Students at an ODeL Institution in South Africa
Authors: Tumelo Jaquiline Ntsopi
Abstract:
This study explored the academic writing challenges of first-year students who use English as an Additional Language (EAL) registered in the EAW101 module at an ODeL institution. Research shows that academic writing is a challenge for EAL teaching and learning contexts across the globe in higher education institutions (HEIs). Academic writing is an important aspect of academic literacy in any institution of higher learning, more so in an ODeL institution. This has probed research that shows that academic writing is and continues to pose challenges for EAL teaching and learning contexts in higher education institutions. This study stems from the researcher’s experience in teaching academic writing to first-year students in the EAW101 module. The motivation for this study emerged from the fact that EAW101 is a writing module that has a high number of students in the Department of English Studies with an average of between 50-80 percent pass rate. These statistics elaborate on the argument that most students registered in this module struggle with academic writing, and they need intervention to assist and support them in achieving competence in the module. This study is underpinned by Community of Inquiry (CoI) framework and Transactional distance theory. This study adopted a qualitative research methodology and utilised a case study approach as a research design. Furthermore, the study gathered data from first year students and the EAW101 module’s student support initiatives. To collect data, focus group discussions, structured open-ended evaluation questions, and an observation schedule were used to gather data. The study is vital towards exploring academic writing challenges that first-year students in EAW101 encounter so that lecturers in the module may consider re-evaluating their methods of teaching to improve EAL students’ academic writing skills. This study may help lecturers towards enhancing academic writing in a ODeL context by assisting first year students through using student support interventions.Keywords: academic writing, academic writing challenge, ODeL, EAL
Procedia PDF Downloads 1092701 Modeling Revolution Shell Structures by MATLAB Programming-Axisymmetric and Nonaxisymmetric Shells
Authors: Hamadi Djamal, Labiodh Bachir, Ounis Abdelhafid, Chaalane Mourad
Abstract:
The objective of this work is setting numerically operational finite element CAXI_L for the axisymmetric and nonaxisymmetric shells. This element is based on the Reissner-Mindlin theory and mixed model formulation. The MATLAB language is used for the programming. In order to test the elaborated program, some applications are carried out.Keywords: axisymmetric shells, nonaxisymmetric behaviour, finite element, MATLAB programming
Procedia PDF Downloads 3222700 High Speed Motion Tracking with Magnetometer in Nonuniform Magnetic Field
Authors: Jeronimo Cox, Tomonari Furukawa
Abstract:
Magnetometers have become more popular in inertial measurement units (IMU) for their ability to correct estimations using the earth's magnetic field. Accelerometer and gyroscope-based packages fail with dead-reckoning errors accumulated over time. Localization in robotic applications with magnetometer-inclusive IMUs has become popular as a way to track the odometry of slower-speed robots. With high-speed motions, the accumulated error increases over smaller periods of time, making them difficult to track with IMU. Tracking a high-speed motion is especially difficult with limited observability. Visual obstruction of motion leaves motion-tracking cameras unusable. When motions are too dynamic for estimation techniques reliant on the observability of the gravity vector, the use of magnetometers is further justified. As available magnetometer calibration methods are limited with the assumption that background magnetic fields are uniform, estimation in nonuniform magnetic fields is problematic. Hard iron distortion is a distortion of the magnetic field by other objects that produce magnetic fields. This kind of distortion is often observed as the offset from the origin of the center of data points when a magnetometer is rotated. The magnitude of hard iron distortion is dependent on proximity to distortion sources. Soft iron distortion is more related to the scaling of the axes of magnetometer sensors. Hard iron distortion is more of a contributor to the error of attitude estimation with magnetometers. Indoor environments or spaces inside ferrite-based structures, such as building reinforcements or a vehicle, often cause distortions with proximity. As positions correlate to areas of distortion, methods of magnetometer localization include the production of spatial mapping of magnetic field and collection of distortion signatures to better aid location tracking. The goal of this paper is to compare magnetometer methods that don't need pre-productions of magnetic field maps. Mapping the magnetic field in some spaces can be costly and inefficient. Dynamic measurement fusion is used to track the motion of a multi-link system with us. Conventional calibration by data collection of rotation at a static point, real-time estimation of calibration parameters each time step, and using two magnetometers for determining local hard iron distortion are compared to confirm the robustness and accuracy of each technique. With opposite-facing magnetometers, hard iron distortion can be accounted for regardless of position, Rather than assuming that hard iron distortion is constant regardless of positional change. The motion measured is a repeatable planar motion of a two-link system connected by revolute joints. The links are translated on a moving base to impulse rotation of the links. Equipping the joints with absolute encoders and recording the motion with cameras to enable ground truth comparison to each of the magnetometer methods. While the two-magnetometer method accounts for local hard iron distortion, the method fails where the magnetic field direction in space is inconsistent.Keywords: motion tracking, sensor fusion, magnetometer, state estimation
Procedia PDF Downloads 882699 New Methodology for Monitoring Alcoholic Fermentation Processes Using Refractometry
Authors: Boukhiar Aissa, Iguergaziz Nadia, Halladj Fatima, Lamrani Yasmina, Benamara Salem
Abstract:
Determining the alcohol content in alcoholic fermentation bioprocess has a great importance. In fact, it is a key indicator for monitoring this fermentation bioprocess. Several methodologies (chemical, spectrophotometric, chromatographic...) are used to the determination of this parameter. However, these techniques are very long and require: rigorous preparations, sometimes dangerous chemical reagents, and/or expensive equipment. In the present study, the date juice is used as a substrate of alcoholic fermentation. The extracted juice undergoes an alcoholic fermentation by Saccharomyces cerevisiae. The study of the possible use of refractometry as a sole means for the in situ control of this process revealed a good correlation (R2 = 0.98) between initial and final ° Brix: ° Brix f = 0.377× ° Brixi. In addition, we verified the relationship between the variation in final and initial ° Brix (Δ ° Brix) and alcoholic rate produced (A exp): CΔ° Brix / A exp = 1.1. This allows the tracing of abacus isoresponses that permit to determine the alcoholic and residual sugar rates with a mean relative error (MRE) of 5.35%.Keywords: refractometry, alcohol, residual sugar, fermentation, brix, date, juice
Procedia PDF Downloads 4832698 Teacher Education: Teacher Development and Support
Authors: Khadem Hichem
Abstract:
With the new technology challenges, dynamics and challenges of the contemporary world, most teachers are struggling to maintain effective and successful teaching /learning environment for learners. Teachers as a key to the success of reforms in the educational setting, they must improve their competencies to teach effectively. Many researchers emphasis on the ongoing professional development of the teacher by enhancing their experiences and encouraging their responsibility for learning, and thus promoting self-reliance, collaboration, and reflection. In short, teachers are considered as learners and they need to learn together. The educational system must support, both conceptually and financially, the teachers’ development as lifelong learners Teachers need opportunities to grow in language proficiency and in knowledge. Changing nature of language and culture in the world, all teachers must have opportunities to update their knowledge and practices. Many researchers in the field of foreign or additional languages indicate that teachers keep side by side of effective instructional practices and they need special support with the challenging task of developing and administering proficiency tests to their students. For significant change to occur, each individual teacher’s needs must be addressed. The teacher must be involved experientially in the process of development, since, by itself, knowledge of how to change does not mean change will be initiated. For improvement to occur, new skills have to be guided, practiced, and reflected upon in collaboration with colleagues. Clearly, teachers are at different places developmentally; therefore, allowances for various entry levels and individual differences need to be built into the professional development structure. Objectives must be meaningful to the participant and teacher improvement must be stated terms of student knowledge, student performance, and motivation. The most successful professional development process acknowledges the student-centered nature of good teaching. This paper highlights the importance of teacher professional development process and institutional supports as way to enhance good teaching and learning environment.Keywords: teacher professional development, teacher competencies, institutional support, teacher education
Procedia PDF Downloads 3572697 E-Portfolios as a Means of Perceiving Students’ Listening and Speaking Progress
Authors: Heba Salem
Abstract:
This paper aims to share the researcher’s experience of using e-Portfolios as an assessment tool to follow up on students’ learning experiences and performance throughout the semester. It also aims at highlighting the importance of students’ self-reflection in the process of language learning. The paper begins by introducing the advanced media course, with its focus on listening and speaking skills, and introduces the students’ profiles. Then it explains the students’ role in the e-portfolio process as they are given the option to choose a listening text they studied throughout the semester and to choose a recorded oral production of their collection of artifacts throughout the semester. Students showcase and reflect on their progress in both listening comprehension and speaking. According to the research, re-listening to work given to them and to their production is a means of reflecting on both their progress and achievement. And choosing the work students want to showcase is a means to promote independent learning as well as self-expression. Students are encouraged to go back to the class learning outcomes in the process of choosing the work. In their reflections, students express how they met the specific learning outcome. While giving their presentations, students expressed how useful the experience of returning and going over what they covered to select one and going over their production as well. They also expressed how beneficial it was to listen to themselves and literally see their progress in both listening comprehension and speaking. Students also reported that they grasped more details from the texts than they did when first having it as an assignment, which coincided with one of the class learning outcomes. They also expressed the fact that they had more confidence speaking as well as they were able to use a variety of vocabulary and idiomatic expressions that students have accumulated. For illustration, this paper includes practical samples of students’ tasks and instructions as well as samples of their reflections. The results of students’ reflections coincide with what the research confirms about the effectiveness of the e-portfolios as a means of assessment. The employment of e-Portfolios has two-folded benefits; students are able to measure the achievement of the targeted learning outcomes, and teachers receive constructive feedback on their teaching methods.Keywords: e-portfolios, assessment, self assessment, listening and speaking progress, foreign language, reflection, learning out comes, sharing experience
Procedia PDF Downloads 1022696 A Detection Method of Faults in Railway Pantographs Based on Dynamic Phase Plots
Authors: G. Santamato, M. Solazzi, A. Frisoli
Abstract:
Systems for detection of damages in railway pantographs effectively reduce the cost of maintenance and improve time scheduling. In this paper, we present an approach to design a monitoring tool fitting strong customer requirements such as portability and ease of use. Pantograph has been modeled to estimate its dynamical properties, since no data are available. With the aim to focus on suspensions health, a two Degrees of Freedom (DOF) scheme has been adopted. Parameters have been calculated by means of analytical dynamics. A Finite Element Method (FEM) modal analysis verified the former model with an acceptable error. The detection strategy seeks phase-plots topology alteration, induced by defects. In order to test the suitability of the method, leakage in the dashpot was simulated on the lumped model. Results are interesting because changes in phase plots are more appreciable than frequency-shift. Further calculations as well as experimental tests will support future developments of this smart strategy.Keywords: pantograph models, phase plots, structural health monitoring, damage detection
Procedia PDF Downloads 3652695 Predicting Seoul Bus Ridership Using Artificial Neural Network Algorithm with Smartcard Data
Authors: Hosuk Shin, Young-Hyun Seo, Eunhak Lee, Seung-Young Kho
Abstract:
Currently, in Seoul, users have the privilege to avoid riding crowded buses with the installation of Bus Information System (BIS). BIS has three levels of on-board bus ridership level information (spacious, normal, and crowded). However, there are flaws in the system due to it being real time which could provide incomplete information to the user. For example, a bus comes to the station, and on the BIS it shows that the bus is crowded, but on the stop that the user is waiting many people get off, which would mean that this station the information should show as normal or spacious. To fix this problem, this study predicts the bus ridership level using smart card data to provide more accurate information about the passenger ridership level on the bus. An Artificial Neural Network (ANN) is an interconnected group of nodes, that was created based on the human brain. Forecasting has been one of the major applications of ANN due to the data-driven self-adaptive methods of the algorithm itself. According to the results, the ANN algorithm was stable and robust with somewhat small error ratio, so the results were rational and reasonable.Keywords: smartcard data, ANN, bus, ridership
Procedia PDF Downloads 1702694 Approximations of Fractional Derivatives and Its Applications in Solving Non-Linear Fractional Variational Problems
Authors: Harendra Singh, Rajesh Pandey
Abstract:
The paper presents a numerical method based on operational matrix of integration and Ryleigh method for the solution of a class of non-linear fractional variational problems (NLFVPs). Chebyshev first kind polynomials are used for the construction of operational matrix. Using operational matrix and Ryleigh method the NLFVP is converted into a system of non-linear algebraic equations, and solving these equations we obtained approximate solution for NLFVPs. Convergence analysis of the proposed method is provided. Numerical experiment is done to show the applicability of the proposed numerical method. The obtained numerical results are compared with exact solution and solution obtained from Chebyshev third kind. Further the results are shown graphically for different fractional order involved in the problems.Keywords: non-linear fractional variational problems, Rayleigh-Ritz method, convergence analysis, error analysis
Procedia PDF Downloads 3032693 Robust Pattern Recognition via Correntropy Generalized Orthogonal Matching Pursuit
Authors: Yulong Wang, Yuan Yan Tang, Cuiming Zou, Lina Yang
Abstract:
This paper presents a novel sparse representation method for robust pattern classification. Generalized orthogonal matching pursuit (GOMP) is a recently proposed efficient sparse representation technique. However, GOMP adopts the mean square error (MSE) criterion and assign the same weights to all measurements, including both severely and slightly corrupted ones. To reduce the limitation, we propose an information-theoretic GOMP (ITGOMP) method by exploiting the correntropy induced metric. The results show that ITGOMP can adaptively assign small weights on severely contaminated measurements and large weights on clean ones, respectively. An ITGOMP based classifier is further developed for robust pattern classification. The experiments on public real datasets demonstrate the efficacy of the proposed approach.Keywords: correntropy induced metric, matching pursuit, pattern classification, sparse representation
Procedia PDF Downloads 3592692 Functioning of Public Distribution System and Calories Intake in the State of Maharashtra
Authors: Balasaheb Bansode, L. Ladusingh
Abstract:
The public distribution system is an important component of food security. It is a massive welfare program undertaken by Government of India and implemented by state government since India being a federal state; for achieving multiple objectives like eliminating hunger, reduction in malnutrition and making food consumption affordable. This program reaches at the community level through the various agencies of the government. The paper focuses on the accessibility of PDS at household level and how the present policy framework results in exclusion and inclusion errors. It tries to explore the sanctioned food grain quantity received by differentiated ration cards according to income criterion at household level, and also it has highlighted on the type of corruption in food distribution that is generated by the PDS system. The data used is of secondary nature from NSSO 68 round conducted in 2012. Bivariate and multivariate techniques have been used to understand the working and consumption of food for this paper.Keywords: calories intake, entitle food quantity, poverty aliviation through PDS, target error
Procedia PDF Downloads 3392691 Continual Learning Using Data Generation for Hyperspectral Remote Sensing Scene Classification
Authors: Samiah Alammari, Nassim Ammour
Abstract:
When providing a massive number of tasks successively to a deep learning process, a good performance of the model requires preserving the previous tasks data to retrain the model for each upcoming classification. Otherwise, the model performs poorly due to the catastrophic forgetting phenomenon. To overcome this shortcoming, we developed a successful continual learning deep model for remote sensing hyperspectral image regions classification. The proposed neural network architecture encapsulates two trainable subnetworks. The first module adapts its weights by minimizing the discrimination error between the land-cover classes during the new task learning, and the second module tries to learn how to replicate the data of the previous tasks by discovering the latent data structure of the new task dataset. We conduct experiments on HSI dataset Indian Pines. The results confirm the capability of the proposed method.Keywords: continual learning, data reconstruction, remote sensing, hyperspectral image segmentation
Procedia PDF Downloads 2712690 Convergence of Sinc Methods Applied to Kuramoto-Sivashinsky Equation
Authors: Kamel Al-Khaled
Abstract:
A comparative study of the Sinc-Galerkin and Sinc-Collocation methods for solving the Kuramoto-Sivashinsky equation is given. Both approaches depend on using Sinc basis functions. Firstly, a numerical scheme using Sinc-Galerkin method is developed to approximate the solution of Kuramoto-Sivashinsky equation. Sinc approximations to both derivatives and indefinite integrals reduces the solution to an explicit system of algebraic equations. The error in the solution is shown to converge to the exact solution at an exponential. The convergence proof of the solution for the discrete system is given using fixed-point iteration. Secondly, a combination of a Crank-Nicolson formula in the time direction, with the Sinc-collocation in the space direction is presented, where the derivatives in the space variable are replaced by the necessary matrices to produce a system of algebraic equations. The methods are tested on two examples. The demonstrated results show that both of the presented methods more or less have the same accuracy.Keywords: Sinc-Collocation, nonlinear PDEs, numerical methods, fixed-point
Procedia PDF Downloads 4742689 Numerical Computation of Generalized Rosenau Regularized Long-Wave Equation via B-Spline Over Butcher’s Fifth Order Runge-Kutta Approach
Authors: Guesh Simretab Gebremedhin, Saumya Rajan Jena
Abstract:
In this work, a septic B-spline scheme has been used to simplify the process of solving an approximate solution of the generalized Rosenau-regularized long-wave equation (GR-RLWE) with initial boundary conditions. The resulting system of first-order ODEs has dealt with Butcher’s fifth order Runge-Kutta (BFRK) approach without using finite difference techniques for discretizing the time-dependent variables at each time level. Here, no transformation or any kind of linearization technique is employed to tackle the nonlinearity of the equation. Two test problems have been selected for numerical justifications and comparisons with other researchers on the basis of efficiency, accuracy, and results of the two invariants Mᵢ (mass) and Eᵢ (energy) of some motion that has been used to test the conservative properties of the proposed scheme.Keywords: septic B-spline scheme, Butcher's fifth order Runge-Kutta approach, error norms, generalized Rosenau-RLW equation
Procedia PDF Downloads 702688 Examining Macroeconomics Determinants of Inflation Rate in Somalia
Authors: Farhia Hassan Mohamed
Abstract:
This study examined the macroeconomic factors that affect the inflation Rate in Somalia using quarterly time series data from 1991q1 to 2017q4 retired from World Development Indicators and SESRIC. It employed the vector error correction model (VECM) and Granger Causality method to measure the long-run and short-run causality of the GDP, inflation exchange rate, and unemployment. The study confirmed that there is one cointegration equation between GDP, exchange rate, inflation, and unemployment in Somalia. However, the VECM model's result indicates a long-run relationship among variables. The VEC Granger causality/Block Exogeneity Wald test result confirmed that all covariates are statistically significant at 5% and are Granger's cause of inflation in the short term. Finally, the impulse response result showed that inflation responds negatively to the shocks from the exchange rate and unemployment rate and positively to GDP and itself. Drawing from the empirical findings, the study makes several policy recommendations for both the monetary and Government sides.Keywords: CPI, OP, exchange rate, inflation ADF, Johansen, PP, VECM, impulse, ECT
Procedia PDF Downloads 532687 Developments in Performance of Autistic Students in the Egyptian School System
Authors: Magy Atef Awad Attia
Abstract:
The objective of this study was to study the effect of social stories on social interaction of students with autism. The sample was at level 5 student with autism, Another University Demonstration School student, who was diagnosed by the Physician as High Functioning Autism since he was able to read, write, calculate and was studying in inclusive classroom. However, he still had disability in social interaction to participate in social activity group and communication. He could not learn how to develop friendship or create relationship. He had inappropriate behavior in social context. He did not understand complex social situations. In addition, he did seemed to not know time and place. He was not able to understand feeling of oneself as well as the others. Consequently, he could not express his emotion appropriately. He did not understand or express his non-verbal language for communicating with friends. He lacked of common interest or emotion with nearby persons. He greeted inappropriately or was not interested in greeting. In addition, he did not have eye contact. He used inadequate language etc. He was elected by Purposive Sampling. His parents were willing to allow them to participate in this study. The research instruments were the lesson plan of social stories, and the picture book of social stories. The instruments used for data collection, were the social interaction evaluation of autistic students. This research was Experimental Research as One Group Pre-test, Post-test Design. For the Pre-test, the experiment was conducted by social stories. Then, the Post-test was implemented. The statistic used for data analysis. The research results were shown by scale. The results revealed that the autistic students taught by social stories indicated better social reaction after being taught by social stories.Keywords: autism, autistic behavior, stability, harsh environments, techniques, thermal, properties, materials, applications, brittleness, fragility, disadvantages, bank, branches, profitability, setting prediction, effective target, measurement, evaluation, performance, commercial, business, sustainability, financial, system.
Procedia PDF Downloads 432686 The Potential of Renewable Energy in Tunisia and Its Impact on Economic Growth
Authors: Assaad Ghazouani
Abstract:
Tunisia is ranked among the countries with low energy diversification, but this configuration makes the country too dependent on fossil fuel exporting countries and therefore extremely sensitive to any oil crises, many measures to diversify electricity production must be taken in making use of other forms of renewable and nuclear energy. One of the solutions required to escape this dependence is the liberalization of the electricity industry which can lead to an improvement of supply, energy diversification, and reducing some of the negative effects of the trade balance. This paper examines the issue of renewable electricity and economic growth in Tunisia consumption. The main objective is to study and analyze the causal link between renewable energy consumption and economic growth in Tunisia over the period 1980-2010. To examine the relationship in the short and in the long terms, we used a multidimensional approach to cointegration based on recent advances in time series econometrics (test Zivot - Andrews, Test of Cointegration Johannsen, Granger causality test, error correction model (ECM)).Keywords: renewable electricity, economic growth, VECM, cointegration, Tunisia
Procedia PDF Downloads 5462685 Time Series Regression with Meta-Clusters
Authors: Monika Chuchro
Abstract:
This paper presents a preliminary attempt to apply classification of time series using meta-clusters in order to improve the quality of regression models. In this case, clustering was performed as a method to obtain a subgroups of time series data with normal distribution from inflow into waste water treatment plant data which Composed of several groups differing by mean value. Two simple algorithms: K-mean and EM were chosen as a clustering method. The rand index was used to measure the similarity. After simple meta-clustering, regression model was performed for each subgroups. The final model was a sum of subgroups models. The quality of obtained model was compared with the regression model made using the same explanatory variables but with no clustering of data. Results were compared by determination coefficient (R2), measure of prediction accuracy mean absolute percentage error (MAPE) and comparison on linear chart. Preliminary results allows to foresee the potential of the presented technique.Keywords: clustering, data analysis, data mining, predictive models
Procedia PDF Downloads 4692684 Semantic Processing in Chinese: Category Effects, Task Effects and Age Effects
Authors: Yi-Hsiu Lai
Abstract:
The present study aimed to elucidate the nature of semantic processing in Chinese. Language and cognition related to the issue of aging are examined from the perspective of picture naming and category fluency tasks. Twenty Chinese-speaking adults (ranging from 25 to 45 years old) and twenty Chinese-speaking seniors (ranging from 65 to 75 years old) in Taiwan participated in this study. Each of them individually completed two tasks: a picture naming task and a category fluency task. Instruments for the naming task were sixty black-and-white pictures: thirty-five object and twenty-five action pictures. Category fluency task also consisted of two semantic categories – objects (or nouns) and actions (or verbs). Participants were asked to report as many items within a category as possible in one minute. Scores of action fluency and of object fluency were a summation of correct responses in these two categories. Category effects (actions vs. objects) and age effects were examined in these tasks. Objects were further divided into two major types: living objects and non-living objects. Actions were also categorized into two major types: action verbs and process verbs. Reaction time to each picture/question was additionally calculated and analyzed. Results of the category fluency task indicated that the content of information in Chinese seniors was comparatively deteriorated, thus producing smaller number of semantic-lexical items. Significant group difference was also found in the results of reaction time. Category Effect was significant for both Chinese adults and seniors in the semantic fluency task. Findings in the present study helped characterize the nature of semantic processing in Chinese-speaking adults and seniors and contributed to the issue of language and aging.Keywords: semantic processing, aging, Chinese, category effects
Procedia PDF Downloads 3632683 The Quality of Health Services and Patient Satisfaction in Hospital
Authors: Malki Nadia Fatima Zahra, Kellal Chaimaa, Brahimi Houria
Abstract:
Quality is one of the most important modern management patterns that organizations seek to achieve in all areas and sectors in order to meet the needs and desires of customers and to remain continuity, as they constitute a competitive advantage for the organization, and among the most prominent organizations that must be available on the quality factor are health organizations as they relate to the most valuable component of production It is a person and his health, and that any error in it threatens his life and may lead to death, so she must provide health services of high quality to achieve the highest degree of satisfaction for the patient. This research aims to study the quality of health services and the extent of their impact on patient satisfaction, and this is through an applied study that relied on measuring the level of quality of health services in the university hospital center of Algeria and the extent of their impact on patient satisfaction according to the dimensions of the quality of health services, and we reached a conclusion that the determinants of the quality of health services. It affects patient satisfaction, which necessitates developing health services according to patients' requirements and improving their quality to obtain patient satisfaction.Keywords: health service, health quality, quality determinants, patient satisfaction
Procedia PDF Downloads 752682 Management of Femoral Neck Stress Fractures at a Specialist Centre and Predictive Factors to Return to Activity Time: An Audit
Authors: Charlotte K. Lee, Henrique R. N. Aguiar, Ralph Smith, James Baldock, Sam Botchey
Abstract:
Background: Femoral neck stress fractures (FNSF) are uncommon, making up 1 to 7.2% of stress fractures in healthy subjects. FNSFs are prevalent in young women, military recruits, endurance athletes, and individuals with energy deficiency syndrome or female athlete triad. Presentation is often non-specific and is often misdiagnosed following the initial examination. There is limited research addressing the return–to–activity time after FNSF. Previous studies have demonstrated prognostic time predictions based on various imaging techniques. Here, (1) OxSport clinic FNSF practice standards are retrospectively reviewed, (2) FNSF cohort demographics are examined, (3) Regression models were used to predict return–to–activity prognosis and consequently determine bone stress risk factors. Methods: Patients with a diagnosis of FNSF attending Oxsport clinic between 01/06/2020 and 01/01/2020 were selected from the Rheumatology Assessment Database Innovation in Oxford (RhADiOn) and OxSport Stress Fracture Database (n = 14). (1) Clinical practice was audited against five criteria based on local and National Institute for Health Care Excellence guidance, with a 100% standard. (2) Demographics of the FNSF cohort were examined with Student’s T-Test. (3) Lastly, linear regression and Random Forest regression models were used on this patient cohort to predict return–to–activity time. Consequently, an analysis of feature importance was conducted after fitting each model. Results: OxSport clinical practice met standard (100%) in 3/5 criteria. The criteria not met were patient waiting times and documentation of all bone stress risk factors. Importantly, analysis of patient demographics showed that of the population with complete bone stress risk factor assessments, 53% were positive for modifiable bone stress risk factors. Lastly, linear regression analysis was utilized to identify demographic factors that predicted return–to–activity time [R2 = 79.172%; average error 0.226]. This analysis identified four key variables that predicted return-to-activity time: vitamin D level, total hip DEXA T value, femoral neck DEXA T value, and history of an eating disorder/disordered eating. Furthermore, random forest regression models were employed for this task [R2 = 97.805%; average error 0.024]. Analysis of the importance of each feature again identified a set of 4 variables, 3 of which matched with the linear regression analysis (vitamin D level, total hip DEXA T value, and femoral neck DEXA T value) and the fourth: age. Conclusion: OxSport clinical practice could be improved by more comprehensively evaluating bone stress risk factors. The importance of this evaluation is demonstrated by the population found positive for these risk factors. Using this cohort, potential bone stress risk factors that significantly impacted return-to-activity prognosis were predicted using regression models.Keywords: eating disorder, bone stress risk factor, femoral neck stress fracture, vitamin D
Procedia PDF Downloads 1862681 Contribution of Word Decoding and Reading Fluency on Reading Comprehension in Young Typical Readers of Kannada Language
Authors: Vangmayee V. Subban, Suzan Deelan. Pinto, Somashekara Haralakatta Shivananjappa, Shwetha Prabhu, Jayashree S. Bhat
Abstract:
Introduction and Need: During early years of schooling, the instruction in the schools mainly focus on children’s word decoding abilities. However, the skilled readers should master all the components of reading such as word decoding, reading fluency and comprehension. Nevertheless, the relationship between each component during the process of learning to read is less clear. The studies conducted in alphabetical languages have mixed opinion on relative contribution of word decoding and reading fluency on reading comprehension. However, the scenarios in alphasyllabary languages are unexplored. Aim and Objectives: The aim of the study was to explore the role of word decoding, reading fluency on reading comprehension abilities in children learning to read Kannada between the age ranges of 5.6 to 8.6 years. Method: In this cross sectional study, a total of 60 typically developing children, 20 each from Grade I, Grade II, Grade III maintaining equal gender ratio between the age range of 5.6 to 6.6 years, 6.7 to 7.6 years and 7.7 to 8.6 years respectively were selected from Kannada medium schools. The reading fluency and reading comprehension abilities of the children were assessed using Grade level passages selected from the Kannada text book of children core curriculum. All the passages consist of five questions to assess reading comprehension. The pseudoword decoding skills were assessed using 40 pseudowords with varying syllable length and their Akshara composition. Pseudowords are formed by interchanging the syllables within the meaningful word while maintaining the phonotactic constraints of Kannada language. The assessment material was subjected to content validation and reliability measures before collecting the data on the study samples. The data were collected individually, and reading fluency was assessed for words correctly read per minute. Pseudoword decoding was scored for the accuracy of reading. Results: The descriptive statistics indicated that the mean pseudoword reading, reading comprehension, words accurately read per minute increased with the Grades. The performance of Grade III children found to be higher, Grade I lower and Grade II remained intermediate of Grade III and Grade I. The trend indicated that reading skills gradually improve with the Grades. Pearson’s correlation co-efficient showed moderate and highly significant (p=0.00) positive co-relation between the variables, indicating the interdependency of all the three components required for reading. The hierarchical regression analysis revealed 37% variance in reading comprehension was explained by pseudoword decoding and was highly significant. Subsequent entry of reading fluency measure, there was no significant change in R-square and was only change 3%. Therefore, pseudoword-decoding evolved as a single most significant predictor of reading comprehension during early Grades of reading acquisition. Conclusion: The present study concludes that the pseudoword decoding skills contribute significantly to reading comprehension than reading fluency during initial years of schooling in children learning to read Kannada language.Keywords: alphasyllabary, pseudo-word decoding, reading comprehension, reading fluency
Procedia PDF Downloads 2662680 Analysis of Joint Source Channel LDPC Coding for Correlated Sources Transmission over Noisy Channels
Authors: Marwa Ben Abdessalem, Amin Zribi, Ammar Bouallègue
Abstract:
In this paper, a Joint Source Channel coding scheme based on LDPC codes is investigated. We consider two concatenated LDPC codes, one allows to compress a correlated source and the second to protect it against channel degradations. The original information can be reconstructed at the receiver by a joint decoder, where the source decoder and the channel decoder run in parallel by transferring extrinsic information. We investigate the performance of the JSC LDPC code in terms of Bit-Error Rate (BER) in the case of transmission over an Additive White Gaussian Noise (AWGN) channel, and for different source and channel rate parameters. We emphasize how JSC LDPC presents a performance tradeoff depending on the channel state and on the source correlation. We show that, the JSC LDPC is an efficient solution for a relatively low Signal-to-Noise Ratio (SNR) channel, especially with highly correlated sources. Finally, a source-channel rate optimization has to be applied to guarantee the best JSC LDPC system performance for a given channel.Keywords: AWGN channel, belief propagation, joint source channel coding, LDPC codes
Procedia PDF Downloads 3622679 Speech Intelligibility Improvement Using Variable Level Decomposition DWT
Authors: Samba Raju, Chiluveru, Manoj Tripathy
Abstract:
Intelligibility is an essential characteristic of a speech signal, which is used to help in the understanding of information in speech signal. Background noise in the environment can deteriorate the intelligibility of a recorded speech. In this paper, we presented a simple variance subtracted - variable level discrete wavelet transform, which improve the intelligibility of speech. The proposed algorithm does not require an explicit estimation of noise, i.e., prior knowledge of the noise; hence, it is easy to implement, and it reduces the computational burden. The proposed algorithm decides a separate decomposition level for each frame based on signal dominant and dominant noise criteria. The performance of the proposed algorithm is evaluated with speech intelligibility measure (STOI), and results obtained are compared with Universal Discrete Wavelet Transform (DWT) thresholding and Minimum Mean Square Error (MMSE) methods. The experimental results revealed that the proposed scheme outperformed competing methodsKeywords: discrete wavelet transform, speech intelligibility, STOI, standard deviation
Procedia PDF Downloads 151