Search results for: feature for feature match
1266 Learners and Teachers Experiences in Collaborative Learning
Authors: Bengi Sonyel, Kheder Kasem
Abstract:
Nowadays technology is growing so fast. Everybody agrees that technology should be enhanced more in educational field in order to achieve maximum level of teaching and learning effectiveness. Collaborative learning is one of the most important subjects that have been discussed widely in the last 20 years. In this growing of technology and the widely spread of e-learning systems most of face-to-face processes are changing to be completely online base. Online collaborative learning considered one of the new feature that applied recently in some e-Learning systems but still there are much differences between face-to-face instance of collaborative learning and what really occur and happen in networked online environment.In this research we will compare face-to-face collaborative learning with online collaborative learning to define the key success for achieving course’s outcomes. We will also study the current teachers and students experience in today e-Learning systems, more specifically in online collaborative system and study them interaction to today’s technology that related to education. We will apply quantitative and qualitative research method in order to get accurate results. Finally we will gather all of our findings, analyze it and try to find the advantages and disadvantages as well as the current problems and possible solutions.Keywords: collaborative learning, learning by doing, technology, teachers, learners experiences
Procedia PDF Downloads 5261265 Emotiv EPOC BCI Matrix Speller Based on Single Emokey
Authors: S. M. Abdullah Al Mamun
Abstract:
Human Computer Interaction (HCI) is an excellent area for the researchers to make daily life more simple and fast. Necessary hardware equipments for any BCI are generally expensive and not affordable for most of the people. Emotiv is one of the solutions for this problem, which can provide electroencephalograph (EEG) signal and explain the brain activities. BCI virtual speller was one of the important applications for the people who have lost their hand or speaking ability because of diseases or unexpected accident. In this paper, a matrix speller has been designed for the first time for Bengali speaking people around the world. Bengali is one of the most commonly spoken languages. Among them, a lot of disabled person will be able to express their desire in their mother tongue. This application is also usable for the social networks and daily life communications. For this virtual keyboard, the well-known matrix speller method with column flashing is applied and controlled by single Emokey only. Emokey is a great feature which translates emotional state for application inputs. In this paper, it is presented that the ITR (Information Transfer Rate) were 29.4 bits/min and typing speed achieved up to 7.43 char/per min.Keywords: brain computer interface, Emotiv EPOC, EEG, virtual keyboard, matrix speller
Procedia PDF Downloads 3091264 Mobile Learning: Toward Better Understanding of Compression Techniques
Authors: Farouk Lawan Gambo
Abstract:
Data compression shrinks files into fewer bits then their original presentation. It has more advantage on internet because the smaller a file, the faster it can be transferred but learning most of the concepts in data compression are abstract in nature therefore making them difficult to digest by some students (Engineers in particular). To determine the best approach toward learning data compression technique, this paper first study the learning preference of engineering students who tend to have strong active, sensing, visual and sequential learning preferences, the paper also study the advantage that mobility of learning have experienced; Learning at the point of interest, efficiency, connection, and many more. A survey is carried out with some reasonable number of students, through random sampling to see whether considering the learning preference and advantages in mobility of learning will give a promising improvement over the traditional way of learning. Evidence from data analysis using Ms-Excel as a point of concern for error-free findings shows that there is significance different in the students after using learning content provided on smart phone, also the result of the findings presented in, bar charts and pie charts interpret that mobile learning has to be promising feature of learning.Keywords: data analysis, compression techniques, learning content, traditional learning approach
Procedia PDF Downloads 3471263 Transmission Design That Eliminates Gradual System Problems in Gearboxes
Authors: Ömer Ateş, Atilla Savaş
Abstract:
Reducers and transmission systems are power and speed transfer tools that have been used for many years in the technology world and in all engineering fields. Since today's transmissions have a threaded tap system, torque interruption occurs during tap change. besides, breakdown and manufacturing costs are high. Another problem is the limited torque and rpm setting in stepped gearbox systems. In this study, a new type of transmission system is designed to solve these problems. This new type of transmission system has been called the Continuously Variable Pulley. The most important feature of the transmission system in the study is that it can be adjusted Revolutions Per Minute-wise and torque-wise at the millimeter (precision) adjustment level. In order to make adjustments at this level, an adjustable pulley with the help of hydraulic piston is designed. The efficiency of the designed transmission system is 97 percent, the efficiency of today's transmissions is in the range of 85-95 percent. examined at the analysis and calculations, it is seen that the designed system gives realistic results and can be compared with today's transmissions and reducers. Therefore, this new type of transmission has been proven to be usable in production areas and the world of technology.Keywords: gearbox, reducer, transmission, torque
Procedia PDF Downloads 1211262 Hybrid Anomaly Detection Using Decision Tree and Support Vector Machine
Authors: Elham Serkani, Hossein Gharaee Garakani, Naser Mohammadzadeh, Elaheh Vaezpour
Abstract:
Intrusion detection systems (IDS) are the main components of network security. These systems analyze the network events for intrusion detection. The design of an IDS is through the training of normal traffic data or attack. The methods of machine learning are the best ways to design IDSs. In the method presented in this article, the pruning algorithm of C5.0 decision tree is being used to reduce the features of traffic data used and training IDS by the least square vector algorithm (LS-SVM). Then, the remaining features are arranged according to the predictor importance criterion. The least important features are eliminated in the order. The remaining features of this stage, which have created the highest level of accuracy in LS-SVM, are selected as the final features. The features obtained, compared to other similar articles which have examined the selected features in the least squared support vector machine model, are better in the accuracy, true positive rate, and false positive. The results are tested by the UNSW-NB15 dataset.Keywords: decision tree, feature selection, intrusion detection system, support vector machine
Procedia PDF Downloads 2661261 Identifying Degradation Patterns of LI-Ion Batteries from Impedance Spectroscopy Using Machine Learning
Authors: Yunwei Zhang, Qiaochu Tang, Yao Zhang, Jiabin Wang, Ulrich Stimming, Alpha Lee
Abstract:
Forecasting the state of health and remaining useful life of Li-ion batteries is an unsolved challenge that limits technologies such as consumer electronics and electric vehicles. Here we build an accurate battery forecasting system by combining electrochemical impedance spectroscopy (EIS) -- a real-time, non-invasive and information-rich measurement that is hitherto underused in battery diagnosis -- with Gaussian process machine learning. We collect over 20,000 EIS spectra of commercial Li-ion batteries at different states of health, states of charge and temperatures -- the largest dataset to our knowledge of its kind. Our Gaussian process model takes the entire spectrum as input, without further feature engineering, and automatically determines which spectral features predict degradation. Our model accurately predicts the remaining useful life, even without complete knowledge of past operating conditions of the battery. Our results demonstrate the value of EIS signals in battery management systems.Keywords: battery degradation, machine learning method, electrochemical impedance spectroscopy, battery diagnosis
Procedia PDF Downloads 1501260 Natural Preservatives: An Alternative for Chemical Preservative Used in Foods
Authors: Zerrin Erginkaya, Gözde Konuray
Abstract:
Microbial degradation of foods is defined as a decrease of food safety due to microorganism activity. Organic acids, sulfur dioxide, sulfide, nitrate, nitrite, dimethyl dicarbonate and several preservative gases have been used as chemical preservatives in foods as well as natural preservatives which are indigenous in foods. It is determined that usage of herbal preservatives such as blueberry, dried grape, prune, garlic, mustard, spices inhibited several microorganisms. Moreover, it is determined that animal origin preservatives such as whey, honey, lysosomes of duck egg and chicken egg, chitosan have antimicrobial effect. Other than indigenous antimicrobials in foods, antimicrobial agents produced by microorganisms could be used as natural preservatives. The antimicrobial feature of preservatives depends on the antimicrobial spectrum, chemical and physical features of material, concentration, mode of action, components of food, process conditions, and pH and storage temperature. In this review, studies about antimicrobial components which are indigenous in food (such as herbal and animal origin antimicrobial agents), antimicrobial materials synthesized by microorganisms, and their usage as an antimicrobial agent to preserve foods are discussed.Keywords: animal origin preservatives, antimicrobial, chemical preservatives, herbal preservatives
Procedia PDF Downloads 3771259 Artificial Neural Network-Based Short-Term Load Forecasting for Mymensingh Area of Bangladesh
Authors: S. M. Anowarul Haque, Md. Asiful Islam
Abstract:
Electrical load forecasting is considered to be one of the most indispensable parts of a modern-day electrical power system. To ensure a reliable and efficient supply of electric energy, special emphasis should have been put on the predictive feature of electricity supply. Artificial Neural Network-based approaches have emerged to be a significant area of interest for electric load forecasting research. This paper proposed an Artificial Neural Network model based on the particle swarm optimization algorithm for improved electric load forecasting for Mymensingh, Bangladesh. The forecasting model is developed and simulated on the MATLAB environment with a large number of training datasets. The model is trained based on eight input parameters including historical load and weather data. The predicted load data are then compared with an available dataset for validation. The proposed neural network model is proved to be more reliable in terms of day-wise load forecasting for Mymensingh, Bangladesh.Keywords: load forecasting, artificial neural network, particle swarm optimization
Procedia PDF Downloads 1721258 Performance Analysis of Bluetooth Low Energy Mesh Routing Algorithm in Case of Disaster Prediction
Authors: Asmir Gogic, Aljo Mujcic, Sandra Ibric, Nermin Suljanovic
Abstract:
Ubiquity of natural disasters during last few decades have risen serious questions towards the prediction of such events and human safety. Every disaster regardless its proportion has a precursor which is manifested as a disruption of some environmental parameter such as temperature, humidity, pressure, vibrations and etc. In order to anticipate and monitor those changes, in this paper we propose an overall system for disaster prediction and monitoring, based on wireless sensor network (WSN). Furthermore, we introduce a modified and simplified WSN routing protocol built on the top of the trickle routing algorithm. Routing algorithm was deployed using the bluetooth low energy protocol in order to achieve low power consumption. Performance of the WSN network was analyzed using a real life system implementation. Estimates of the WSN parameters such as battery life time, network size and packet delay are determined. Based on the performance of the WSN network, proposed system can be utilized for disaster monitoring and prediction due to its low power profile and mesh routing feature.Keywords: bluetooth low energy, disaster prediction, mesh routing protocols, wireless sensor networks
Procedia PDF Downloads 3881257 Vfx-Creativity or Cost Cutting Study of the Use of Vfx in Hindi Cinema
Authors: Nidhi Patel, Amol Shinde, Amrin Moger
Abstract:
Mainstream Hindi cinema also known as Bollywood, is the largest film producing industry in India. The Indian film industry underwent a sea change since last few years. The industry adapted to the latest technologies and creative manpower to improve visual and cinematic effects. The changes helped the industry to improve its creative looks and ease on production budget. The research focuses on this very change, i.e. the use of VFX. There has been growing use of VFX in feature films. The primary focus is on how VFX can make a difference in the experience of watching a movie. The research examines the use of CGI/VFX in the narrative, which delivers a visually fulfilling film. It also focuses on the use of CGI/ VFX as a cost cutting tool. The research was exploratory in nature. It studies the industry’s evolvement, increment in its use by filmmakers and their intention to use it in their films. The researcher used qualitative method for data collection as an in-depth interview of 10 artists from VFX studios in Mumbai was conducted. The finding reveals the way VFX is used in Hindi cinema by the directors. The researcher learnt that VFX is majorly used as a tool to enhance creativity and provide the audience with creative viewing experience.Keywords: Bollywood, Hindi cinema, VFX, CGI, technology, creativity, cost cutting
Procedia PDF Downloads 3601256 Vehicle Detection and Tracking Using Deep Learning Techniques in Surveillance Image
Authors: Abe D. Desta
Abstract:
This study suggests a deep learning-based method for identifying and following moving objects in surveillance video. The proposed method uses a fast regional convolution neural network (F-RCNN) trained on a substantial dataset of vehicle images to first detect vehicles. A Kalman filter and a data association technique based on a Hungarian algorithm are then used to monitor the observed vehicles throughout time. However, in general, F-RCNN algorithms have been shown to be effective in achieving high detection accuracy and robustness in this research study. For example, in one study The study has shown that the vehicle detection and tracking, the system was able to achieve an accuracy of 97.4%. In this study, the F-RCNN algorithm was compared to other popular object detection algorithms and was found to outperform them in terms of both detection accuracy and speed. The presented system, which has application potential in actual surveillance systems, shows the usefulness of deep learning approaches in vehicle detection and tracking.Keywords: artificial intelligence, computer vision, deep learning, fast-regional convolutional neural networks, feature extraction, vehicle tracking
Procedia PDF Downloads 1291255 Investigating (Im)Politeness Strategies in Email Communication: The Case Algerian PhD Supervisees and Irish Supervisors
Authors: Zehor Ktitni
Abstract:
In pragmatics, politeness is regarded as a feature of paramount importance to successful interpersonal relationships. On the other hand, emails have recently become one of the indispensable means of communication in educational settings. This research puts email communication at the core of the study and analyses it from a politeness perspective. More specifically, it endeavours to look closely at how the concept of (im)politeness is reflected through students’ emails. To this end, a corpus of Algerian supervisees’ email threads, exchanged with their Irish supervisors, was compiled. Leech’s model of politeness (2014) was selected as the main theoretical framework of this study, in addition to making reference to Brown and Levinson’s model (1987) as it is one of the most influential models in the area of pragmatic politeness. Further, some follow-up interviews are to be conducted with Algerian students to reinforce the results derived from the corpus. Initial findings suggest that Algerian Ph.D. students’ emails tend to include more politeness markers than impoliteness ones, they heavily make use of academic titles when addressing their supervisors (Dr. or Prof.), and they rely on hedging devices in order to sound polite.Keywords: politeness, email communication, corpus pragmatics, Algerian PhD supervisees, Irish supervisors
Procedia PDF Downloads 701254 Bag of Words Representation Based on Fusing Two Color Local Descriptors and Building Multiple Dictionaries
Authors: Fatma Abdedayem
Abstract:
We propose an extension to the famous method called Bag of words (BOW) which proved a successful role in the field of image categorization. Practically, this method based on representing image with visual words. In this work, firstly, we extract features from images using Spatial Pyramid Representation (SPR) and two dissimilar color descriptors which are opponent-SIFT and transformed-color-SIFT. Secondly, we fuse color local features by joining the two histograms coming from these descriptors. Thirdly, after collecting of all features, we generate multi-dictionaries coming from n random feature subsets that obtained by dividing all features into n random groups. Then, by using these dictionaries separately each image can be represented by n histograms which are lately concatenated horizontally and form the final histogram, that allows to combine Multiple Dictionaries (MDBoW). In the final step, in order to classify image we have applied Support Vector Machine (SVM) on the generated histograms. Experimentally, we have used two dissimilar image datasets in order to test our proposition: Caltech 256 and PASCAL VOC 2007.Keywords: bag of words (BOW), color descriptors, multi-dictionaries, MDBoW
Procedia PDF Downloads 2971253 Rules in Policy Integration, Case Study: Victoria Catchment Management
Authors: Ratri Werdiningtyas, Yongping Wei, Andrew Western
Abstract:
This paper contributes to on-going attempts at bringing together land, water and environmental policy in catchment management. A tension remains in defining the boundaries of policy integration. Most of Integrated Water Resource Management is valued as rhetoric policy. It is far from being achieved on the ground because the socio-ecological system has not been understood and developed into complete and coherent problem representation. To clarify the feature of integration, this article draws on institutional fit for public policy integration and uses these insights in an empirical setting to identify the mechanism that can facilitate effective public integration for catchment management. This research is based on the journey of Victoria’s government from 1890-2016. A total of 274 Victorian Acts related to land, water, environment management published in those periods has been investigated. Four conditions of integration have been identified in their co-evolution: (1) the integration policy based on reserves, (2) the integration policy based on authority interest, (3) policy based on integrated information and, (4) policy based coordinated resource, authority and information. Results suggest that policy coordination among their policy instrument is superior rather than policy integration in the case of catchment management.Keywords: catchment management, co-evolution, policy integration, phase
Procedia PDF Downloads 2491252 MindFlow: A Collective Intelligence-Based System for Helping Stress Pattern Diagnosis
Authors: Andres Frederic
Abstract:
We present the MindFlow system supporting the detection and the diagnosis of stresses. The heart of the system is a knowledge synthesis engine allowing occupational health stakeholders (psychologists, occupational therapists and human resource managers) to formulate queries related to stress and responding to users requests by recommending a pattern of stress if one exists. The stress pattern diagnosis is based on expert knowledge stored in the MindFlow stress ontology including stress feature vector. The query processing may involve direct access to the MindFlow system by occupational health stakeholders, online communication between the MindFlow system and the MindFlow domain experts, or direct dialog between a occupational health stakeholder and a MindFlow domain expert. The MindFlow knowledge model is generic in the sense that it supports the needs of psychologists, occupational therapists and human resource managers. The system presented in this paper is currently under development as part of a Dutch-Japanese project and aims to assist organisation in the quick diagnosis of stress patterns.Keywords: occupational stress, stress management, physiological measurement, accident prevention
Procedia PDF Downloads 4341251 Adaptive Kaman Filter for Fault Diagnosis of Linear Parameter-Varying Systems
Authors: Rajamani Doraiswami, Lahouari Cheded
Abstract:
Fault diagnosis of Linear Parameter-Varying (LPV) system using an adaptive Kalman filter is proposed. The LPV model is comprised of scheduling parameters, and the emulator parameters. The scheduling parameters are chosen such that they are capable of tracking variations in the system model as a result of changes in the operating regimes. The emulator parameters, on the other hand, simulate variations in the subsystems during the identification phase and have negligible effect during the operational phase. The nominal model and the influence vectors, which are the gradient of the feature vector respect to the emulator parameters, are identified off-line from a number of emulator parameter perturbed experiments. A Kalman filter is designed using the identified nominal model. As the system varies, the Kalman filter model is adapted using the scheduling variables. The residual is employed for fault diagnosis. The proposed scheme is successfully evaluated on simulated system as well as on a physical process control system.Keywords: identification, linear parameter-varying systems, least-squares estimation, fault diagnosis, Kalman filter, emulators
Procedia PDF Downloads 5001250 Operating System Based Virtualization Models in Cloud Computing
Authors: Dev Ras Pandey, Bharat Mishra, S. K. Tripathi
Abstract:
Cloud computing is ready to transform the structure of businesses and learning through supplying the real-time applications and provide an immediate help for small to medium sized businesses. The ability to run a hypervisor inside a virtual machine is important feature of virtualization and it is called nested virtualization. In today’s growing field of information technology, many of the virtualization models are available, that provide a convenient approach to implement, but decision for a single model selection is difficult. This paper explains the applications of operating system based virtualization in cloud computing with an appropriate/suitable model with their different specifications and user’s requirements. In the present paper, most popular models are selected, and the selection was based on container and hypervisor based virtualization. Selected models were compared with a wide range of user’s requirements as number of CPUs, memory size, nested virtualization supports, live migration and commercial supports, etc. and we identified a most suitable model of virtualization.Keywords: virtualization, OS based virtualization, container based virtualization, hypervisor based virtualization
Procedia PDF Downloads 3311249 Collaborative and Context-Aware Learning Approach Using Mobile Technology
Authors: Sameh Baccari, Mahmoud Neji
Abstract:
In recent years, the rapid developments on mobile devices and wireless technologies enable new dimension capabilities for the learning domain. This dimension facilitates people daily activities and shortens the distances between individuals. When these technologies have been used in learning, a new paradigm has been emerged giving birth to mobile learning. Because of the mobility feature, m-learning courses have to be adapted dynamically to the learner’s context. The main challenge in context-aware mobile learning is to develop an approach building the best learning resources according to dynamic learning situations. In this paper, we propose a context-aware mobile learning system called Collaborative and Context-aware Mobile Learning System (CCMLS). It takes into account the requirements of Mobility, Collaboration and Context-Awareness. This system is based on the semantic modeling of the learning context and the learning content. The adaptation part of this approach is made up of adaptation rules to propose and select relevant resources, learning partners and learning activities based not only on the user’s needs, but also on its current context.Keywords: mobile learning, mobile technologies, context-awareness, collaboration, semantic web, adaptation engine, adaptation strategy, learning object, learning context
Procedia PDF Downloads 3081248 Enhancement of X-Rays Images Intensity Using Pixel Values Adjustments Technique
Authors: Yousif Mohamed Y. Abdallah, Razan Manofely, Rajab M. Ben Yousef
Abstract:
X-Ray images are very popular as a first tool for diagnosis. Automating the process of analysis of such images is important in order to help physician procedures. In this practice, teeth segmentation from the radiographic images and feature extraction are essential steps. The main objective of this study was to study correction preprocessing of x-rays images using local adaptive filters in order to evaluate contrast enhancement pattern in different x-rays images such as grey color and to evaluate the usage of new nonlinear approach for contrast enhancement of soft tissues in x-rays images. The data analyzed by using MatLab program to enhance the contrast within the soft tissues, the gray levels in both enhanced and unenhanced images and noise variance. The main techniques of enhancement used in this study were contrast enhancement filtering and deblurring images using the blind deconvolution algorithm. In this paper, prominent constraints are firstly preservation of image's overall look; secondly, preservation of the diagnostic content in the image and thirdly detection of small low contrast details in diagnostic content of the image.Keywords: enhancement, x-rays, pixel intensity values, MatLab
Procedia PDF Downloads 4881247 A Comparative Analysis on QRS Peak Detection Using BIOPAC and MATLAB Software
Authors: Chandra Mukherjee
Abstract:
The present paper is a representation of the work done in the field of ECG signal analysis using MATLAB 7.1 Platform. An accurate and simple ECG feature extraction algorithm is presented in this paper and developed algorithm is validated using BIOPAC software. To detect the QRS peak, ECG signal is processed by following mentioned stages- First Derivative, Second Derivative and then squaring of that second derivative. Efficiency of developed algorithm is tested on ECG samples from different database and real time ECG signals acquired using BIOPAC system. Firstly we have lead wise specified threshold value the samples above that value is marked and in the original signal, where these marked samples face change of slope are spotted as R-peak. On the left and right side of the R-peak, faces change of slope identified as Q and S peak, respectively. Now the inbuilt Detection algorithm of BIOPAC software is performed on same output sample and both outputs are compared. ECG baseline modulation correction is done after detecting characteristics points. The efficiency of the algorithm is tested using some validation parameters like Sensitivity, Positive Predictivity and we got satisfied value of these parameters.Keywords: first derivative, variable threshold, slope reversal, baseline modulation correction
Procedia PDF Downloads 4111246 KSVD-SVM Approach for Spontaneous Facial Expression Recognition
Authors: Dawood Al Chanti, Alice Caplier
Abstract:
Sparse representations of signals have received a great deal of attention in recent years. In this paper, the interest of using sparse representation as a mean for performing sparse discriminative analysis between spontaneous facial expressions is demonstrated. An automatic facial expressions recognition system is presented. It uses a KSVD-SVM approach which is made of three main stages: A pre-processing and feature extraction stage, which solves the problem of shared subspace distribution based on the random projection theory, to obtain low dimensional discriminative and reconstructive features; A dictionary learning and sparse coding stage, which uses the KSVD model to learn discriminative under or over dictionaries for sparse coding; Finally a classification stage, which uses a SVM classifier for facial expressions recognition. Our main concern is to be able to recognize non-basic affective states and non-acted expressions. Extensive experiments on the JAFFE static acted facial expressions database but also on the DynEmo dynamic spontaneous facial expressions database exhibit very good recognition rates.Keywords: dictionary learning, random projection, pose and spontaneous facial expression, sparse representation
Procedia PDF Downloads 3081245 Intrusion Detection System Using Linear Discriminant Analysis
Authors: Zyad Elkhadir, Khalid Chougdali, Mohammed Benattou
Abstract:
Most of the existing intrusion detection systems works on quantitative network traffic data with many irrelevant and redundant features, which makes detection process more time’s consuming and inaccurate. A several feature extraction methods, such as linear discriminant analysis (LDA), have been proposed. However, LDA suffers from the small sample size (SSS) problem which occurs when the number of the training samples is small compared with the samples dimension. Hence, classical LDA cannot be applied directly for high dimensional data such as network traffic data. In this paper, we propose two solutions to solve SSS problem for LDA and apply them to a network IDS. The first method, reduce the original dimension data using principal component analysis (PCA) and then apply LDA. In the second solution, we propose to use the pseudo inverse to avoid singularity of within-class scatter matrix due to SSS problem. After that, the KNN algorithm is used for classification process. We have chosen two known datasets KDDcup99 and NSLKDD for testing the proposed approaches. Results showed that the classification accuracy of (PCA+LDA) method outperforms clearly the pseudo inverse LDA method when we have large training data.Keywords: LDA, Pseudoinverse, PCA, IDS, NSL-KDD, KDDcup99
Procedia PDF Downloads 2281244 Tapered Double Cantilever Beam: Evaluation of the Test Set-up for Self-Healing Polymers
Authors: Eleni Tsangouri, Xander Hillewaere, David Garoz Gómez, Dimitrios Aggelis, Filip Du Prez, Danny Van Hemelrijck
Abstract:
Tapered Double Cantilever Beam (TDCB) is the most commonly used test set-up to evaluate the self-healing feature of thermoset polymers autonomously activated in the presence of crack. TDCB is a modification of the established fracture mechanics set-up of Double Cantilever Beam and is designed to provide constant strain energy release rate with crack length under stable load evolution (mode-I). In this study, the damage of virgin and autonomously healed TDCB polymer samples is evaluated considering the load-crack opening diagram, the strain maps provided by Digital Image Correlation technique and the fractography maps given by optical microscopy. It is shown that the pre-crack introduced prior to testing (razor blade tapping), the loading rate and the length of the side groove are the features that dominate the crack propagation and lead to inconstant fracture energy release rate.Keywords: polymers, autonomous healing, fracture, tapered double cantilever beam
Procedia PDF Downloads 3521243 Contrast Enhancement of Color Images with Color Morphing Approach
Authors: Javed Khan, Aamir Saeed Malik, Nidal Kamel, Sarat Chandra Dass, Azura Mohd Affandi
Abstract:
Low contrast images can result from the wrong setting of image acquisition or poor illumination conditions. Such images may not be visually appealing and can be difficult for feature extraction. Contrast enhancement of color images can be useful in medical area for visual inspection. In this paper, a new technique is proposed to improve the contrast of color images. The RGB (red, green, blue) color image is transformed into normalized RGB color space. Adaptive histogram equalization technique is applied to each of the three channels of normalized RGB color space. The corresponding channels in the original image (low contrast) and that of contrast enhanced image with adaptive histogram equalization (AHE) are morphed together in proper proportions. The proposed technique is tested on seventy color images of acne patients. The results of the proposed technique are analyzed using cumulative variance and contrast improvement factor measures. The results are also compared with decorrelation stretch. Both subjective and quantitative analysis demonstrates that the proposed techniques outperform the other techniques.Keywords: contrast enhacement, normalized RGB, adaptive histogram equalization, cumulative variance.
Procedia PDF Downloads 3791242 Characteristics and Item Parameters Fitness on Chemistry Teacher-Made Test Instrument
Authors: Rizki Nor Amelia, Farida A. Setiawati
Abstract:
This study aimed to: (1) describe the characteristics of teacher-made test instrument used to measure the ability of students’chemistry, and (2) identify the presence of the compability difficulty level set by teachers to difficulty level by empirical results. Based on these objectives, this study was a descriptive research. The analysis in this study used the Rasch model and Chi-square statistics. Analysis using Rasch Model was based on the response patterns of high school students to the teacher-made test instrument on chemistry subject Academic Year 2015/2016 in the Yogyakarta. The sample of this research were 358 students taken by cluster random sampling technique. The analysis showed that: (1) a teacher-made tests instrument has a medium on the mean difficulty level. This instrument is capable to measure the ability on the interval of -0,259 ≤ θ ≤ 0,659 logit. Maximum Test Information Function obtained at 18.187 on the ability +0,2 logit; (2) 100% items categorized either as easy or difficult by rasch model is match with the teachers’ judgment; while 37 items are categorized according to rasch model which 8.10% and 10.81% categorized as easy and difficult items respectively according to the teachers, the others are medium categorized. Overall, the distribution of the level of difficulty formulated by the teachers has the distinction (not match) to the level of difficulty based on the empirical results.Keywords: chemistry, items parameter fitness, Rasch model, teacher-made test
Procedia PDF Downloads 2391241 Multimedia Data Fusion for Event Detection in Twitter by Using Dempster-Shafer Evidence Theory
Authors: Samar M. Alqhtani, Suhuai Luo, Brian Regan
Abstract:
Data fusion technology can be the best way to extract useful information from multiple sources of data. It has been widely applied in various applications. This paper presents a data fusion approach in multimedia data for event detection in twitter by using Dempster-Shafer evidence theory. The methodology applies a mining algorithm to detect the event. There are two types of data in the fusion. The first is features extracted from text by using the bag-ofwords method which is calculated using the term frequency-inverse document frequency (TF-IDF). The second is the visual features extracted by applying scale-invariant feature transform (SIFT). The Dempster - Shafer theory of evidence is applied in order to fuse the information from these two sources. Our experiments have indicated that comparing to the approaches using individual data source, the proposed data fusion approach can increase the prediction accuracy for event detection. The experimental result showed that the proposed method achieved a high accuracy of 0.97, comparing with 0.93 with texts only, and 0.86 with images only.Keywords: data fusion, Dempster-Shafer theory, data mining, event detection
Procedia PDF Downloads 4111240 “Ethical Porn” and the Right to Withdraw Consent
Authors: Nathan Elvidge
Abstract:
This paper offers a philosophical argument against the possibility of so-called “ethical porn,” that is, pornographic material produced in a way attempting to remain consistent with feminist principles and female empowerment. One key feature of such material is the requirement for the material to be consensual on the part of the actors or those involved in the material. However, in the contemporary pornography industry, this typically amounts to a single historic act of consent given in exchange for a lump-sum payment which grants the producer lifetime property rights over the explicit material. This paper argues that, by the lights of feminist principles, this situation is inherently unjust and that, as a consequence, the pornography industry requires a radical systematic upheaval before any material produced within it can be considered genuinely ethical. These feminist principles require that for the consumption of pornography to be genuinely ethical, the actors must consent not only to the acts recorded in the material but also to the consumption of that material. This paper argues that this consent to consumption should be treated as on par with other matters of sexual consent and, therefore, that actors should have the right to withdraw consent to the consumption of their material. From this, it is argued to follow that the system of third-party ownership of property rights over someone else’s sexually explicit material legally nullifies this right and therefore is inherently unjust.Keywords: consent, feminism, pornography, sex work
Procedia PDF Downloads 1161239 A Local Invariant Generalized Hough Transform Method for Integrated Circuit Visual Positioning
Authors: Wei Feilong
Abstract:
In this study, an local invariant generalized Houghtransform (LI-GHT) method is proposed for integrated circuit (IC) visual positioning. The original generalized Hough transform (GHT) is robust to external noise; however, it is not suitable for visual positioning of IC chips due to the four-dimensionality (4D) of parameter space which leads to the substantial storage requirement and high computational complexity. The proposed LI-GHT method can reduce the dimensionality of parameter space to 2D thanks to the rotational invariance of local invariant geometric feature and it can estimate the accuracy position and rotation angle of IC chips in real-time under noise and blur influence. The experiment results show that the proposed LI-GHT can estimate position and rotation angle of IC chips with high accuracy and fast speed. The proposed LI-GHT algorithm was implemented in IC visual positioning system of radio frequency identification (RFID) packaging equipment.Keywords: Integrated Circuit Visual Positioning, Generalized Hough Transform, Local invariant Generalized Hough Transform, ICpacking equipment
Procedia PDF Downloads 2661238 High Purity Germanium Detector Characterization by Means of Monte Carlo Simulation through Application of Geant4 Toolkit
Authors: Milos Travar, Jovana Nikolov, Andrej Vranicar, Natasa Todorovic
Abstract:
Over the years, High Purity Germanium (HPGe) detectors proved to be an excellent practical tool and, as such, have established their today's wide use in low background γ-spectrometry. One of the advantages of gamma-ray spectrometry is its easy sample preparation as chemical processing and separation of the studied subject are not required. Thus, with a single measurement, one can simultaneously perform both qualitative and quantitative analysis. One of the most prominent features of HPGe detectors, besides their excellent efficiency, is their superior resolution. This feature virtually allows a researcher to perform a thorough analysis by discriminating photons of similar energies in the studied spectra where otherwise they would superimpose within a single-energy peak and, as such, could potentially scathe analysis and produce wrongly assessed results. Naturally, this feature is of great importance when the identification of radionuclides, as well as their activity concentrations, is being practiced where high precision comes as a necessity. In measurements of this nature, in order to be able to reproduce good and trustworthy results, one has to have initially performed an adequate full-energy peak (FEP) efficiency calibration of the used equipment. However, experimental determination of the response, i.e., efficiency curves for a given detector-sample configuration and its geometry, is not always easy and requires a certain set of reference calibration sources in order to account for and cover broader energy ranges of interest. With the goal of overcoming these difficulties, a lot of researches turned towards the application of different software toolkits that implement the Monte Carlo method (e.g., MCNP, FLUKA, PENELOPE, Geant4, etc.), as it has proven time and time again to be a very powerful tool. In the process of creating a reliable model, one has to have well-established and described specifications of the detector. Unfortunately, the documentation that manufacturers provide alongside the equipment is rarely sufficient enough for this purpose. Furthermore, certain parameters tend to evolve and change over time, especially with older equipment. Deterioration of these parameters consequently decreases the active volume of the crystal and can thus affect the efficiencies by a large margin if they are not properly taken into account. In this study, the optimisation method of two HPGe detectors through the implementation of the Geant4 toolkit developed by CERN is described, with the goal of further improving simulation accuracy in calculations of FEP efficiencies by investigating the influence of certain detector variables (e.g., crystal-to-window distance, dead layer thicknesses, inner crystal’s void dimensions, etc.). Detectors on which the optimisation procedures were carried out were a standard traditional co-axial extended range detector (XtRa HPGe, CANBERRA) and a broad energy range planar detector (BEGe, CANBERRA). Optimised models were verified through comparison with experimentally obtained data from measurements of a set of point-like radioactive sources. Acquired results of both detectors displayed good agreement with experimental data that falls under an average statistical uncertainty of ∼ 4.6% for XtRa and ∼ 1.8% for BEGe detector within the energy range of 59.4−1836.1 [keV] and 59.4−1212.9 [keV], respectively.Keywords: HPGe detector, γ spectrometry, efficiency, Geant4 simulation, Monte Carlo method
Procedia PDF Downloads 1211237 Improved Performance in Content-Based Image Retrieval Using Machine Learning Approach
Authors: B. Ramesh Naik, T. Venugopal
Abstract:
This paper presents a novel approach which improves the high-level semantics of images based on machine learning approach. The contemporary approaches for image retrieval and object recognition includes Fourier transforms, Wavelets, SIFT and HoG. Though these descriptors helpful in a wide range of applications, they exploit zero order statistics, and this lacks high descriptiveness of image features. These descriptors usually take benefit of primitive visual features such as shape, color, texture and spatial locations to describe images. These features do not adequate to describe high-level semantics of the images. This leads to a gap in semantic content caused to unacceptable performance in image retrieval system. A novel method has been proposed referred as discriminative learning which is derived from machine learning approach that efficiently discriminates image features. The analysis and results of proposed approach were validated thoroughly on WANG and Caltech-101 Databases. The results proved that this approach is very competitive in content-based image retrieval.Keywords: CBIR, discriminative learning, region weight learning, scale invariant feature transforms
Procedia PDF Downloads 183