Search results for: computer aided classifier.
981 Combination of Different Classifiers for Cardiac Arrhythmia Recognition
Authors: M. R. Homaeinezhad, E. Tavakkoli, M. Habibi, S. A. Atyabi, A. Ghaffari
Abstract:
This paper describes a new supervised fusion (hybrid) electrocardiogram (ECG) classification solution consisting of a new QRS complex geometrical feature extraction as well as a new version of the learning vector quantization (LVQ) classification algorithm aimed for overcoming the stability-plasticity dilemma. Toward this objective, after detection and delineation of the major events of ECG signal via an appropriate algorithm, each QRS region and also its corresponding discrete wavelet transform (DWT) are supposed as virtual images and each of them is divided into eight polar sectors. Then, the curve length of each excerpted segment is calculated and is used as the element of the feature space. To increase the robustness of the proposed classification algorithm versus noise, artifacts and arrhythmic outliers, a fusion structure consisting of five different classifiers namely as Support Vector Machine (SVM), Modified Learning Vector Quantization (MLVQ) and three Multi Layer Perceptron-Back Propagation (MLP–BP) neural networks with different topologies were designed and implemented. The new proposed algorithm was applied to all 48 MIT–BIH Arrhythmia Database records (within–record analysis) and the discrimination power of the classifier in isolation of different beat types of each record was assessed and as the result, the average accuracy value Acc=98.51% was obtained. Also, the proposed method was applied to 6 number of arrhythmias (Normal, LBBB, RBBB, PVC, APB, PB) belonging to 20 different records of the aforementioned database (between– record analysis) and the average value of Acc=95.6% was achieved. To evaluate performance quality of the new proposed hybrid learning machine, the obtained results were compared with similar peer– reviewed studies in this area.Keywords: Feature Extraction, Curve Length Method, SupportVector Machine, Learning Vector Quantization, Multi Layer Perceptron, Fusion (Hybrid) Classification, Arrhythmia Classification, Supervised Learning Machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2227980 Generalized π-Armendariz Authentication Cryptosystem
Authors: Areej M. Abduldaim, Nadia M. G. Al-Saidi
Abstract:
Algebra is one of the important fields of mathematics. It concerns with the study and manipulation of mathematical symbols. It also concerns with the study of abstractions such as groups, rings, and fields. Due to the development of these abstractions, it is extended to consider other structures, such as vectors, matrices, and polynomials, which are non-numerical objects. Computer algebra is the implementation of algebraic methods as algorithms and computer programs. Recently, many algebraic cryptosystem protocols are based on non-commutative algebraic structures, such as authentication, key exchange, and encryption-decryption processes are adopted. Cryptography is the science that aimed at sending the information through public channels in such a way that only an authorized recipient can read it. Ring theory is the most attractive category of algebra in the area of cryptography. In this paper, we employ the algebraic structure called skew -Armendariz rings to design a neoteric algorithm for zero knowledge proof. The proposed protocol is established and illustrated through numerical example, and its soundness and completeness are proved.
Keywords: Cryptosystem, identification, skew π-Armendariz rings, skew polynomial rings, zero knowledge protocol.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 796979 Impact of Computer-Mediated Communication on Virtual Teams- Performance: An Empirical Study
Authors: Nadeem Ehsan, Ebtisam Mirza, Muhammad Ahmad
Abstract:
In a complex project environment, project teams face multi-dimensional communication problems that can ultimately lead to project breakdown. Team Performance varies in Face-to-Face (FTF) environment versus groups working remotely in a computermediated communication (CMC) environment. A brief review of the Input_Process_Output model suggested by James E. Driskell, Paul H. Radtke and Eduardo Salas in “Virtual Teams: Effects of Technological Mediation on Team Performance (2003)", has been done to develop the basis of this research. This model theoretically analyzes the effects of technological mediation on team processes, such as, cohesiveness, status and authority relations, counternormative behavior and communication. An empirical study described in this paper has been undertaken to test the “cohesiveness" of diverse project teams in a multi-national organization. This study uses both quantitative and qualitative techniques for data gathering and analysis. These techniques include interviews, questionnaires for data collection and graphical data representation for analyzing the collected data. Computer-mediated technology may impact team performance because of difference in cohesiveness among teams and this difference may be moderated by factors, such as, the type of communication environment, the type of task and the temporal context of the team. Based on the reviewed model, sets of hypotheses are devised and tested. This research, reports on a study that compared team cohesiveness among virtual teams using CMC and non-CMC communication mediums. The findings suggest that CMC can help virtual teams increase team cohesiveness among their members, making CMC an effective medium for increasing productivity and team performance.Keywords: Computer-mediated Communication, Virtual Teams, Team Performance, Team Cohesiveness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2332978 Numerical Simulation of Investment Casting of Gold Jewelry: Experiments and Validations
Authors: Marco Actis Grande, Somlak Wannarumon
Abstract:
This paper proposes the numerical simulation of the investment casting of gold jewelry. It aims to study the behavior of fluid flow during mould filling and solidification and to optimize the process parameters, which lead to predict and control casting defects such as gas porosity and shrinkage porosity. A finite difference method, computer simulation software FLOW-3D was used to simulate the jewelry casting process. The simplified model was designed for both numerical simulation and real casting production. A set of sensor acquisitions were allocated on the different positions of the wax tree of the model to detect filling times, while a set of thermocouples were allocated to detect the temperature during casting and cooling. Those detected data were applied to validate the results of the numerical simulation to the results of the real casting. The resulting comparisons signify that the numerical simulation can be used as an effective tool in investment-casting-process optimization and casting-defect prediction.Keywords: Computer fluid dynamic, Investment casting, Jewelry, Mould filling, Simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2737977 Lung Cancer Detection and Multi Level Classification Using Discrete Wavelet Transform Approach
Authors: V. Veeraprathap, G. S. Harish, G. Narendra Kumar
Abstract:
Uncontrolled growth of abnormal cells in the lung in the form of tumor can be either benign (non-cancerous) or malignant (cancerous). Patients with Lung Cancer (LC) have an average of five years life span expectancy provided diagnosis, detection and prediction, which reduces many treatment options to risk of invasive surgery increasing survival rate. Computed Tomography (CT), Positron Emission Tomography (PET), and Magnetic Resonance Imaging (MRI) for earlier detection of cancer are common. Gaussian filter along with median filter used for smoothing and noise removal, Histogram Equalization (HE) for image enhancement gives the best results without inviting further opinions. Lung cavities are extracted and the background portion other than two lung cavities is completely removed with right and left lungs segmented separately. Region properties measurements area, perimeter, diameter, centroid and eccentricity measured for the tumor segmented image, while texture is characterized by Gray-Level Co-occurrence Matrix (GLCM) functions, feature extraction provides Region of Interest (ROI) given as input to classifier. Two levels of classifications, K-Nearest Neighbor (KNN) is used for determining patient condition as normal or abnormal, while Artificial Neural Networks (ANN) is used for identifying the cancer stage is employed. Discrete Wavelet Transform (DWT) algorithm is used for the main feature extraction leading to best efficiency. The developed technology finds encouraging results for real time information and on line detection for future research.
Keywords: ANN, DWT, GLCM, KNN, ROI, artificial neural networks, discrete wavelet transform, gray-level co-occurrence matrix, k-nearest neighbor, region of interest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 960976 Magnetic Field Analysis for a Distribution Transformer with Unbalanced Load Conditions by using 3-D Finite Element Method
Authors: P. Meesuk, T. Kulworawanichpong, P. Pao-la-or
Abstract:
This paper proposes a set of quasi-static mathematical model of magnetic fields caused by high voltage conductors of distribution transformer by using a set of second-order partial differential equation. The modification for complex magnetic field analysis and time-harmonic simulation are also utilized. In this research, transformers were study in both balanced and unbalanced loading conditions. Computer-based simulation utilizing the threedimensional finite element method (3-D FEM) is exploited as a tool for visualizing magnetic fields distribution volume a distribution transformer. Finite Element Method (FEM) is one among popular numerical methods that is able to handle problem complexity in various forms. At present, the FEM has been widely applied in most engineering fields. Even for problems of magnetic field distribution, the FEM is able to estimate solutions of Maxwell-s equations governing the power transmission systems. The computer simulation based on the use of the FEM has been developed in MATLAB programming environment.Keywords: Distribution Transformer, Magnetic Field, Load Unbalance, 3-D Finite Element Method (3-D FEM)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2692975 Noise Depressed in a Micro Stepping Motor
Authors: Bo-Wun Huang, Jao-Hwa Kuang, J.-G. Tseng, Yan-De Wu
Abstract:
An investigation of noise in a micro stepping motor is considered to study in this article. Because of the trend towards higher precision and more and more small 3C (including Computer, Communication and Consumer Electronics) products, the micro stepping motor is frequently used to drive the micro system or the other 3C products. Unfortunately, noise in a micro stepped motor is too large to accept by the customs. To depress the noise of a micro stepped motor, the dynamic characteristics in this system must be studied. In this article, a Visual Basic (VB) computer program speed controlled micro stepped motor in a digital camera is investigated. Karman KD2300-2S non-contract eddy current displacement sensor, probe microphone, and HP 35670A analyzer are employed to analyze the dynamic characteristics of vibration and noise in a motor. The vibration and noise measurement of different type of bearings and different treatment of coils are compared. The rotating components, bearings, coil, etc. of the motor play the important roles in producing vibration and noise. It is found that the noise will be depressed about 3~4 dB and 6~7 dB, when substitutes the copper bearing with plastic one and coats the motor coil with paraffin wax, respectively.Keywords: micro motor, noise, vibration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1745974 Cybersecurity Protection Structures: The Case of Lesotho
Authors: N. N. Mosola, K. F. Moeketsi, R. Sehobai, N. Pule
Abstract:
The Internet brings increasing use of Information and Communications Technology (ICT) services and facilities. Consequently, new computing paradigms emerge to provide services over the Internet. Although there are several benefits stemming from these services, they pose several risks inherited from the Internet. For example, cybercrime, identity theft, malware etc. To thwart these risks, this paper proposes a holistic approach. This approach involves multidisciplinary interactions. The paper proposes a top-down and bottom-up approach to deal with cyber security concerns in developing countries. These concerns range from regulatory and legislative areas, cyber awareness, research and development, technical dimensions etc. The main focus areas are highlighted and a cybersecurity model solution is proposed. The paper concludes by combining all relevant solutions into a proposed cybersecurity model to assist developing countries in enhancing a cyber-safe environment to instill and promote a culture of cybersecurity.Keywords: Cybercrime, cybersecurity, computer emergency response team, computer security incident response team.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2092973 Spectral Mixture Model Applied to Cannabis Parcel Determination
Authors: Levent Basayigit, Sinan Demir, Yusuf Ucar, Burhan Kara
Abstract:
Many research projects require accurate delineation of the different land cover type of the agricultural area. Especially it is critically important for the definition of specific plants like cannabis. However, the complexity of vegetation stands structure, abundant vegetation species, and the smooth transition between different seconder section stages make vegetation classification difficult when using traditional approaches such as the maximum likelihood classifier. Most of the time, classification distinguishes only between trees/annual or grain. It has been difficult to accurately determine the cannabis mixed with other plants. In this paper, a mixed distribution models approach is applied to classify pure and mix cannabis parcels using Worldview-2 imagery in the Lakes region of Turkey. Five different land use types (i.e. sunflower, maize, bare soil, and cannabis) were identified in the image. A constrained Gaussian mixture discriminant analysis (GMDA) was used to unmix the image. In the study, 255 reflectance ratios derived from spectral signatures of seven bands (Blue-Green-Yellow-Red-Rededge-NIR1-NIR2) were randomly arranged as 80% for training and 20% for test data. Gaussian mixed distribution model approach is proved to be an effective and convenient way to combine very high spatial resolution imagery for distinguishing cannabis vegetation. Based on the overall accuracies of the classification, the Gaussian mixed distribution model was found to be very successful to achieve image classification tasks. This approach is sensitive to capture the illegal cannabis planting areas in the large plain. This approach can also be used for monitoring and determination with spectral reflections in illegal cannabis planting areas.
Keywords: Gaussian mixture discriminant analysis, spectral mixture model, World View-2, land parcels.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 800972 Evaluating per-user Fairness of Goal-Oriented Parallel Computer Job Scheduling Policies
Authors: Sangsuree Vasupongayya
Abstract:
Fair share objective has been included into the goaloriented parallel computer job scheduling policy recently. However, the previous work only presented the overall scheduling performance. Thus, the per-user performance of the policy is still lacking. In this work, the details of per-user fair share performance under the Tradeoff-fs(Tx:avgX) policy will be further evaluated. A basic fair share priority backfill policy namely RelShare(1d) is also studied. The performance of all policies is collected using an event-driven simulator with three real job traces as input. The experimental results show that the high demand users are usually benefited under most policies because their jobs are large or they have a lot of jobs. In the large job case, one job executed may result in over-share during that period. In the other case, the jobs may be backfilled for performances. However, the users with a mixture of jobs may suffer because if the smaller jobs are executing the priority of the remaining jobs from the same user will be lower. Further analysis does not show any significant impact of users with a lot of jobs or users with a large runtime approximation error.Keywords: deviation, fair share, discrepancy search, priority scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1352971 FITTER - A Framework for Integrating Activity Tracking Technologies into Electric Recreation for Children and Adolescents
Authors: R. Altamimi, G. Skinner, K. Nesbitt
Abstract:
Encouraging physical activity amongst children and adolescents is becoming an increasingly relevant issue in modern society. Studies have shown that involving children and adolescents in physical activity is essential for their physical, mental and social development. However, with technology playing an increasingly important role in reducing physical work it is becoming more critical to incorporate adequate physical activities into our lives. One way to overcome this problem is to harness technology so that it promotes physical activities, for example, by motivating children and adolescents to exercise more. This paper describes a promising solution to the question of how to increase levels of physical activity in children and adolescents by combining gaming technologies with exercise tracking goals. This research describes a framework called FITTER (Framework for Integrating activity Tracking Technologies for Electronic Recreation) that combines video game play with more traditional, non-computer physical activities.
Keywords: Exergames, Home-based eHealth, Human-computer Interaction, Natural User Interfaces, Wearable Health Informatics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2447970 Study of the Late Phase of Core Degradation during Reflooding by Safety Injection System for VVER1000 with ASTECv2 Computer Code
Authors: Antoaneta Stefanova, Rositsa Gencheva, Pavlin Groudev
Abstract:
This paper presents the modeling approach in SBO sequence for VVER 1000 reactors and describes the reactor core behavior at late in-vessel phase in case of late reflooding by HPIS and gives preliminary results for the ASTECv2 validation. The work is focused on investigation of plant behavior during total loss of power and the operator actions. The main goal of these analyses is to assess the phenomena arising during the Station blackout (SBO) followed by primary side high pressure injection system (HPIS) reflooding of already damaged reactor core at very late “in-vessel” phase. The purpose of the analyses is to define how the later HPIS switching on can delay the time of vessel failure or possibly avoid vessel failure. The times for HPP injection were chosen based on previously performed investigations.Keywords: VVER, operator action validation, reflooding of overheated reactor core, ASTEC computer code.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1442969 A P2P File Sharing Technique by Indexed-Priority Metric
Authors: Toshinori Takabatake, Yoshikazu Komano
Abstract:
Recently, the improvements in processing performance of a computer and in high speed communication of an optical fiber have been achieved, so that the amount of data which are processed by a computer and flowed on a network has been increasing greatly. However, in a client-server system, since the server receives and processes the amount of data from the clients through the network, a load on the server is increasing. Thus, there are needed to introduce a server with high processing ability and to have a line with high bandwidth. In this paper, concerning to P2P networks to resolve the load on a specific server, a criterion called an Indexed-Priority Metric is proposed and its performance is evaluated. The proposed metric is to allocate some files to each node. As a result, the load on a specific server can distribute them to each node equally well. A P2P file sharing system using the proposed metric is implemented. Simulation results show that the proposed metric can make it distribute files on the specific server.Keywords: peer-to-peer, file-sharing system, load-balancing, dependability
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1387968 A Study on the Modeling and Analysis of an Electro-Hydraulic Power Steering System
Authors: Ji-Hye Kim, Sung-Gaun Kim
Abstract:
Electro-hydraulic power steering (EHPS) system for the fuel rate reduction and steering feel improvement is comprised of ECU including the logic which controls the steering system and BL DC motor and produces the best suited cornering force, BLDC motor, high pressure pump integrated module and basic oil-hydraulic circuit of the commercial HPS system. Electro-hydraulic system can be studied in two ways such as experimental and computer simulation. To get accurate results in experimental study of EHPS system, the real boundary management is necessary which is difficult task. And the accuracy of the experimental results depends on the preparation of the experimental setup and accuracy of the data collection. The computer simulation gives accurate and reliable results if the simulation is carried out considering proper boundary conditions. So, in this paper, each component of EHPS was modeled, and the model-based analysis and control logic was designed by using AMESimKeywords: Power steering system, Electro-Hydraulic power steering (EHPS) system, Modeling of EHPS system, Analysis modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2707967 Measuring Cognitive Load - A Solution to Ease Learning of Programming
Authors: Muhammed Yousoof, Mohd Sapiyan, Khaja Kamaluddin
Abstract:
Learning programming is difficult for many learners. Some researches have found that the main difficulty relates to cognitive load. Cognitive overload happens in programming due to the nature of the subject which is intrinisicly over-bearing on the working memory. It happens due to the complexity of the subject itself. The problem is made worse by the poor instructional design methodology used in the teaching and learning process. Various efforts have been proposed to reduce the cognitive load, e.g. visualization softwares, part-program method etc. Use of many computer based systems have also been tried to tackle the problem. However, little success has been made to alleviate the problem. More has to be done to overcome this hurdle. This research attempts at understanding how cognitive load can be managed so as to reduce the problem of overloading. We propose a mechanism to measure the cognitive load during pre instruction, post instruction and in instructional stages of learning. This mechanism is used to help the instruction. As the load changes the instruction is made to adapt itself to ensure cognitive viability. This mechanism could be incorporated as a sub domain in the student model of various computer based instructional systems to facilitate the learning of programming.
Keywords: Cognitive load, Working memory, Cognitive Loadmeasurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2562966 Adjustment and Compensation Techniques for the Rotary Axes of Five-axis CNC Machine Tools
Authors: Tung-Hui Hsu, Wen-Yuh Jywe
Abstract:
Five-axis computer numerical control (CNC) machine tools (three linear and two rotary axes) are ideally suited to the fabrication of complex work pieces, such as dies, turbo blades, and cams. The locations of the axis average line and centerline of the rotary axes strongly influence the performance of these machines; however, techniques to compensate for eccentric error in the rotary axes remain weak. This paper proposes optical (Non-Bar) techniques capable of calibrating five-axis CNC machine tools and compensating for eccentric error in the rotary axes. This approach employs the measurement path in ISO/CD 10791-6 to determine the eccentric error in two rotary axes, for which compensatory measures can be implemented. Experimental results demonstrate that the proposed techniques can improve the performance of various five-axis CNC machine tools by more than 90%. Finally, a result of the cutting test using a B-type five-axis CNC machine tool confirmed to the usefulness of this proposed compensation technique.
Keywords: Calibration, compensation, rotary axis, five-axis computer numerical control (CNC) machine tools, eccentric error, optical calibration system, ISO/CD 10791-6
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4168965 A Learner-Centred or Artefact-Centred Classroom? Impact of Technology, Artefacts, and Environment on Task Processes in an English as a Foreign Language Classroom
Authors: Nobue T. Ellis
Abstract:
This preliminary study attempts to see if a learning environment influences instructor’s teaching strategies and learners’ in-class activities in a foreign language class at a university in Japan. The class under study was conducted in a computer room, while the majority of classes of the same course were offered in traditional classrooms without computers. The study also sees if the unplanned blended learning environment, enhanced, or worked against, in achieving course goals, by paying close attention to in-class artefacts, such as computers. In the macro-level analysis, the course syllabus and weekly itinerary of the course were looked at; and in the microlevel analysis, nonhuman actors in their environments were named and analyzed to see how they influenced the learners’ task processes. The result indicated that students were heavily influenced by the presence of computers, which lead them to disregard some aspects of intended learning objectives.
Keywords: Computer-assisted language learning, actor-network theory, English as a foreign language, task-based teaching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610964 Web Data Scraping Technology Using Term Frequency Inverse Document Frequency to Enhance the Big Data Quality on Sentiment Analysis
Authors: Sangita Pokhrel, Nalinda Somasiri, Rebecca Jeyavadhanam, Swathi Ganesan
Abstract:
Tourism is a booming industry with huge future potential for global wealth and employment. There are countless data generated over social media sites every day, creating numerous opportunities to bring more insights to decision-makers. The integration of big data technology into the tourism industry will allow companies to conclude where their customers have been and what they like. This information can then be used by businesses, such as those in charge of managing visitor centres or hotels, etc., and the tourist can get a clear idea of places before visiting. The technical perspective of natural language is processed by analysing the sentiment features of online reviews from tourists, and we then supply an enhanced long short-term memory (LSTM) framework for sentiment feature extraction of travel reviews. We have constructed a web review database using a crawler and web scraping technique for experimental validation to evaluate the effectiveness of our methodology. The text form of sentences was first classified through VADER and RoBERTa model to get the polarity of the reviews. In this paper, we have conducted study methods for feature extraction, such as Count Vectorization and Term Frequency – Inverse Document Frequency (TFIDF) Vectorization and implemented Convolutional Neural Network (CNN) classifier algorithm for the sentiment analysis to decide if the tourist’s attitude towards the destinations is positive, negative, or simply neutral based on the review text that they posted online. The results demonstrated that from the CNN algorithm, after pre-processing and cleaning the dataset, we received an accuracy of 96.12% for the positive and negative sentiment analysis.
Keywords: Counter vectorization, Convolutional Neural Network, Crawler, data technology, Long Short-Term Memory, LSTM, Web Scraping, sentiment analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 175963 Promoting Complex Systems Learning through the use of Computer Modeling
Authors: Kamel Hashem, David Mioduser
Abstract:
This paper describes part of a project about Learningby- Modeling (LbM). Studying complex systems is increasingly important in teaching and learning many science domains. Many features of complex systems make it difficult for students to develop deep understanding. Previous research indicates that involvement with modeling scientific phenomena and complex systems can play a powerful role in science learning. Some researchers argue with this view indicating that models and modeling do not contribute to understanding complexity concepts, since these increases the cognitive load on students. This study will investigate the effect of different modes of involvement in exploring scientific phenomena using computer simulation tools, on students- mental model from the perspective of structure, behavior and function. Quantitative and qualitative methods are used to report about 121 freshmen students that engaged in participatory simulations about complex phenomena, showing emergent, self-organized and decentralized patterns. Results show that LbM plays a major role in students' concept formation about complexity concepts.Keywords: Complexity, Educational technology, Learning by modeling, Mental models
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1572962 Machine Scoring Model Using Data Mining Techniques
Authors: Wimalin S. Laosiritaworn, Pongsak Holimchayachotikul
Abstract:
this article proposed a methodology for computer numerical control (CNC) machine scoring. The case study company is a manufacturer of hard disk drive parts in Thailand. In this company, sample of parts manufactured from CNC machine are usually taken randomly for quality inspection. These inspection data were used to make a decision to shut down the machine if it has tendency to produce parts that are out of specification. Large amount of data are produced in this process and data mining could be very useful technique in analyzing them. In this research, data mining techniques were used to construct a machine scoring model called 'machine priority assessment model (MPAM)'. This model helps to ensure that the machine with higher risk of producing defective parts be inspected before those with lower risk. If the defective prone machine is identified sooner, defective part and rework could be reduced hence improving the overall productivity. The results showed that the proposed method can be successfully implemented and approximately 351,000 baht of opportunity cost could have saved in the case study company.Keywords: Computer Numerical Control, Data Mining, HardDisk Drive.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1395961 A Motion Dictionary to Real-Time Recognition of Sign Language Alphabet Using Dynamic Time Warping and Artificial Neural Network
Authors: Marcio Leal, Marta Villamil
Abstract:
Computacional recognition of sign languages aims to allow a greater social and digital inclusion of deaf people through interpretation of their language by computer. This article presents a model of recognition of two of global parameters from sign languages; hand configurations and hand movements. Hand motion is captured through an infrared technology and its joints are built into a virtual three-dimensional space. A Multilayer Perceptron Neural Network (MLP) was used to classify hand configurations and Dynamic Time Warping (DWT) recognizes hand motion. Beyond of the method of sign recognition, we provide a dataset of hand configurations and motion capture built with help of fluent professionals in sign languages. Despite this technology can be used to translate any sign from any signs dictionary, Brazilian Sign Language (Libras) was used as case study. Finally, the model presented in this paper achieved a recognition rate of 80.4%.Keywords: Sign language recognition, computer vision, infrared, artificial neural network, dynamic time warping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 878960 Heart Rate Variability Analysis for Early Stage Prediction of Sudden Cardiac Death
Authors: Reeta Devi, Hitender Kumar Tyagi, Dinesh Kumar
Abstract:
In present scenario, cardiovascular problems are growing challenge for researchers and physiologists. As heart disease have no geographic, gender or socioeconomic specific reasons; detecting cardiac irregularities at early stage followed by quick and correct treatment is very important. Electrocardiogram is the finest tool for continuous monitoring of heart activity. Heart rate variability (HRV) is used to measure naturally occurring oscillations between consecutive cardiac cycles. Analysis of this variability is carried out using time domain, frequency domain and non-linear parameters. This paper presents HRV analysis of the online dataset for normal sinus rhythm (taken as healthy subject) and sudden cardiac death (SCD subject) using all three methods computing values for parameters like standard deviation of node to node intervals (SDNN), square root of mean of the sequences of difference between adjacent RR intervals (RMSSD), mean of R to R intervals (mean RR) in time domain, very low-frequency (VLF), low-frequency (LF), high frequency (HF) and ratio of low to high frequency (LF/HF ratio) in frequency domain and Poincare plot for non linear analysis. To differentiate HRV of healthy subject from subject died with SCD, k –nearest neighbor (k-NN) classifier has been used because of its high accuracy. Results show highly reduced values for all stated parameters for SCD subjects as compared to healthy ones. As the dataset used for SCD patients is recording of their ECG signal one hour prior to their death, it is therefore, verified with an accuracy of 95% that proposed algorithm can identify mortality risk of a patient one hour before its death. The identification of a patient’s mortality risk at such an early stage may prevent him/her meeting sudden death if in-time and right treatment is given by the doctor.Keywords: Early stage prediction, heart rate variability, linear and non linear analysis, sudden cardiac death.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1807959 Development of Online Islamic Medication Expert System (OIMES)
Authors: Hanita Daud, Noorhana Yahya, Low Tan Jung, Azizuddin Abd Aziz, Rabibah Ahmad Samawe
Abstract:
This paper presents an overview of the design and implementation of an online rule-based Expert Systems for Islamic medication. T his Online Islamic Medication Expert System (OIMES) focuses on physical illnesses only. Knowledge base of this Expert System contains exhaustively the types of illness together with their related cures or treatments/therapies, obtained exclusively from the Quran and Hadith. Extensive research and study are conducted to ensure that the Expert System is able to provide the most suitable treatment with reference to the relevant verses cited in Quran or Hadith. These verses come together with their related 'actions' (bodily actions/gestures or some acts) to be performed by the patient to treat a particular illness/sickness. These verses and the instructions for the 'actions' are to be displayed unambiguously on the computer screen. The online platform provides the advantage for patient getting treatment practically anytime and anywhere as long as the computer and Internet facility exist. Patient does not need to make appointment to see an expert for a therapy.Keywords: Expert System, Quran and Hadith, Islamic Medication, Rule-Based.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1839958 Effects of Computer–Based Instructional Designs among Pupils of Different Music Intelligence Levels
Authors: Aldalalah, M. Osamah, Soon Fook Fong
Abstract:
The purpose of this study was to investigate the effects of computer–based instructional designs, namely modality and redundancy principles on the attitude and learning of music theory among primary pupils of different Music Intelligence levels. The lesson of music theory was developed in three different modes, audio and image (AI), text with image (TI) and audio with image and text (AIT). The independent variables were the three modes of courseware. The moderator variable was music intelligence. The dependent variables were the post test score. ANOVA was used to determine the significant differences of the pretest scores among the three groups. Analyses of covariance (ANCOVA) and Post hoc were carried out to examine the main effects as well as the interaction effects of the independent variables on the dependent variables. High music intelligence pupils performed significantly better than low music intelligence pupils in all the three treatment modes. The AI mode was found to help pupils with low music intelligence significantly more than the TI and AIT modes.
Keywords: Modality, Redundancy, Music theory, Cognitivetheory of multimedia learning, Cognitive load theory, Musicintelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1669957 An Intelligent Combined Method Based on Power Spectral Density, Decision Trees and Fuzzy Logic for Hydraulic Pumps Fault Diagnosis
Authors: Kaveh Mollazade, Hojat Ahmadi, Mahmoud Omid, Reza Alimardani
Abstract:
Recently, the issue of machine condition monitoring and fault diagnosis as a part of maintenance system became global due to the potential advantages to be gained from reduced maintenance costs, improved productivity and increased machine availability. The aim of this work is to investigate the effectiveness of a new fault diagnosis method based on power spectral density (PSD) of vibration signals in combination with decision trees and fuzzy inference system (FIS). To this end, a series of studies was conducted on an external gear hydraulic pump. After a test under normal condition, a number of different machine defect conditions were introduced for three working levels of pump speed (1000, 1500, and 2000 rpm), corresponding to (i) Journal-bearing with inner face wear (BIFW), (ii) Gear with tooth face wear (GTFW), and (iii) Journal-bearing with inner face wear plus Gear with tooth face wear (B&GW). The features of PSD values of vibration signal were extracted using descriptive statistical parameters. J48 algorithm is used as a feature selection procedure to select pertinent features from data set. The output of J48 algorithm was employed to produce the crisp if-then rule and membership function sets. The structure of FIS classifier was then defined based on the crisp sets. In order to evaluate the proposed PSD-J48-FIS model, the data sets obtained from vibration signals of the pump were used. Results showed that the total classification accuracy for 1000, 1500, and 2000 rpm conditions were 96.42%, 100%, and 96.42% respectively. The results indicate that the combined PSD-J48-FIS model has the potential for fault diagnosis of hydraulic pumps.Keywords: Power Spectral Density, Machine ConditionMonitoring, Hydraulic Pump, Fuzzy Logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2712956 Individual Differences and Paired Learning in Virtual Environments
Authors: Patricia M. Boechler, Heather M. Gautreau
Abstract:
In this research study, postsecondary students completed an information learning task in an avatar-based 3D virtual learning environment. Three factors were of interest in relation to learning; 1) the influence of collaborative vs. independent conditions, 2) the influence of the spatial arrangement of the virtual environment (linear, random and clustered), and 3) the relationship of individual differences such as spatial skill, general computer experience and video game experience to learning. Students completed pretest measures of prior computer experience and prior spatial skill. Following the premeasure administration, students were given instruction to move through the virtual environment and study all the material within 10 information stations. In the collaborative condition, students proceeded in randomly assigned pairs, while in the independent condition they proceeded alone. After this learning phase, all students individually completed a multiple choice test to determine information retention. The overall results indicated that students in pairs did not perform any better or worse than independent students. As far as individual differences, only spatial ability predicted the performance of students. General computer experience and video game experience did not. Taking a closer look at the pairs and spatial ability, comparisons were made on pairs high/matched spatial ability, pairs low/matched spatial ability and pairs that were mismatched on spatial ability. The results showed that both high/matched pairs and mismatched pairs outperformed low/matched pairs. That is, if a pair had even one individual with strong spatial ability they would perform better than pairs with only low spatial ability individuals. This suggests that, in virtual environments, the specific individuals that are paired together are important for performance outcomes. The paper also includes a discussion of trends within the data that have implications for virtual environment education.
Keywords: Avatar-based, virtual environment, paired learning, individual differences.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 780955 Harris Extraction and SIFT Matching for Correlation of Two Tablets
Authors: Ali Alzaabi, Georges Alquié, Hussain Tassadaq, Ali Seba
Abstract:
This article presents the developments of efficient algorithms for tablet copies comparison. Image recognition has specialized use in digital systems such as medical imaging, computer vision, defense, communication etc. Comparison between two images that look indistinguishable is a formidable task. Two images taken from different sources might look identical but due to different digitizing properties they are not. Whereas small variation in image information such as cropping, rotation, and slight photometric alteration are unsuitable for based matching techniques. In this paper we introduce different matching algorithms designed to facilitate, for art centers, identifying real painting images from fake ones. Different vision algorithms for local image features are implemented using MATLAB. In this framework a Table Comparison Computer Tool “TCCT" is designed to facilitate our research. The TCCT is a Graphical Unit Interface (GUI) tool used to identify images by its shapes and objects. Parameter of vision system is fully accessible to user through this graphical unit interface. And then for matching, it applies different description technique that can identify exact figures of objects.Keywords: Harris Extraction and SIFT Matching
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1734954 The Socio-Technical Indicator Model: Socially-Sensitive CMC Technology, with an Implementation of Representative Moderation
Authors: Zach-Amaury Boufoy-Bastick, Lenandlar Singh
Abstract:
Computer-mediated communication technologies which provide for virtual communities have typically evolved in a cross-dichotomous manner, such that technical constructs of the technology have evolved independently from the social environment of the community. The present paper analyses some limitations of current implementations of computer-mediated communication technology that are implied by such a dichotomy, and discusses their inhibiting effects on possible developments of virtual communities. A Socio-Technical Indicator Model is introduced that utilizes integrated feedback to describe, simulate and operationalise increasing representativeness within a variety of structurally and parametrically diverse systems. In illustration, applications of the model are briefly described for financial markets and for eco-systems. A detailed application is then provided to resolve the aforementioned technical limitations of moderation on the evolution of virtual communities. The application parameterises virtual communities to function as self-transforming social-technical systems which are sensitive to emergent and shifting community values as products of on-going communications within the collective.
Keywords: Virtual community, e-democracy, feedback systems, moderation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1570953 Seismic Behaviour of Romanian Ortodox Churches, Modeling of Failure Modes by Rigid Blocks
Authors: Marius Mosoarca, Victor Gioncu, Ovidiu Cosma
Abstract:
Historic religious buildings located in seismic areas have developed different failure mechanisms. Simulation of failure modes is done with computer programs through a nonlinear dynamic analysis or simplified using the method of failure blocks. Currently there are simulation methodologies of failure modes based on the failure rigid blocks method only for Roman Catholic churches type. Due to differences of shape in plan, elevation and construction systems between Orthodox churches and Catholic churches, for the first time there were initiated researches in the development of this simulation methodology for Orthodox churches. In this article are presented the first results from the researches. The theoretical results were compared with real failure modes recorded at an Orthodox church from Banat region, severely damaged by earthquakes in 1991. Simulated seismic response, using a computer program based on finite element method was confirmed by cracks after earthquakes. The consolidation of the church was made according to these theoretical results, realizing a rigid floor connecting all the failure blocks.Keywords: Dinamic analysis, failure mechanism, rigid blocks seismic simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1637952 High Level Synthesis of Canny Edge Detection Algorithm on Zynq Platform
Authors: Hanaa M. Abdelgawad, Mona Safar, Ayman M. Wahba
Abstract:
Real time image and video processing is a demand in many computer vision applications, e.g. video surveillance, traffic management and medical imaging. The processing of those video applications requires high computational power. Thus, the optimal solution is the collaboration of CPU and hardware accelerators. In this paper, a Canny edge detection hardware accelerator is proposed. Edge detection is one of the basic building blocks of video and image processing applications. It is a common block in the pre-processing phase of image and video processing pipeline. Our presented approach targets offloading the Canny edge detection algorithm from processing system (PS) to programmable logic (PL) taking the advantage of High Level Synthesis (HLS) tool flow to accelerate the implementation on Zynq platform. The resulting implementation enables up to a 100x performance improvement through hardware acceleration. The CPU utilization drops down and the frame rate jumps to 60 fps of 1080p full HD input video stream.
Keywords: High Level Synthesis, Canny edge detection, Hardware accelerators, and Computer Vision.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5431