Search results for: facial nerve recovery.
556 Semi-Automated Tracking of Vibrissal Movements in Free-Moving Rodents Captured by High-Speed Videos
Authors: Hyun June Kim, Tailong Shi, Seden Akdagli, Sam Most, Yuling Yan
Abstract:
Quantitative analyses of whisker movements provide a means to study functional recovery and regeneration of mouse facial nerve after an injury. However, accurate tracking of the mouse whisker movement is challenging. Most methods for whisker tracking require manual intervention, e.g. fixing the head of the mouse during a study. Here we describe a semi-automated image processing method, which is applied to high-speed video recordings of free-moving mice to track the whisker movements. We first track the head movement of a mouse by delineating the lower head contour frame-by-frame that allows for detection of the location and orientation of the head. Then, a region of interest is identified for each frame; the subsequent application of a mask and the Hough transform detects the selected whiskers on each side of the head. Our approach is used to examine the functional recovery of damaged facial nerves in mice over a course of 21 days.Keywords: Mystacial macrovibrissae, whisker tracking, head tracking, facial nerve recovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685555 Close Loop Controlled Current Nerve Locator
Authors: H. A. Alzomor, B. K. Ouda, A. M. Eldeib
Abstract:
Successful regional anesthesia depends upon precise location of the peripheral nerve or nerve plexus. Locating peripheral nerves is preferred to be done using nerve stimulation. In order to generate a nerve impulse by electrical means, a minimum threshold stimulus of current “rheobase” must be applied to the nerve. The technique depends on stimulating muscular twitching at a close distance to the nerve without actually touching it. Success rate of this operation depends on the accuracy of current intensity pulses used for stimulation .In this paper, we will discuss a circuit and algorithm for closed loop control for the current, theoretical analysis and test results is discussed and results is compared to previous techniques.
Keywords: Close Loop Control, Constant Current, Nerve Locator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1406554 A Survey on Facial Feature Points Detection Techniques and Approaches
Authors: Rachid Ahdid, Khaddouj Taifi, Said Safi, Bouzid Manaut
Abstract:
Automatic detection of facial feature points plays an important role in applications such as facial feature tracking, human-machine interaction and face recognition. The majority of facial feature points detection methods using two-dimensional or three-dimensional data are covered in existing survey papers. In this article chosen approaches to the facial features detection have been gathered and described. This overview focuses on the class of researches exploiting facial feature points detection to represent facial surface for two-dimensional or three-dimensional face. In the conclusion, we discusses advantages and disadvantages of the presented algorithms.Keywords: Facial feature points, face recognition, facial feature tracking, two-dimensional data, three-dimensional data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1680553 Facial Expressions Animation and Lip Tracking Using Facial Characteristic Points and Deformable Model
Authors: Hadi Seyedarabi, Ali Aghagolzadeh, Sohrab Khanmohammadi
Abstract:
Face and facial expressions play essential roles in interpersonal communication. Most of the current works on the facial expression recognition attempt to recognize a small set of the prototypic expressions such as happy, surprise, anger, sad, disgust and fear. However the most of the human emotions are communicated by changes in one or two of discrete features. In this paper, we develop a facial expressions synthesis system, based on the facial characteristic points (FCP's) tracking in the frontal image sequences. Selected FCP's are automatically tracked using a crosscorrelation based optical flow. The proposed synthesis system uses a simple deformable facial features model with a few set of control points that can be tracked in original facial image sequences.Keywords: Deformable face model, facial animation, facialcharacteristic points, optical flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1633552 Automatic Facial Skin Segmentation Using Possibilistic C-Means Algorithm for Evaluation of Facial Surgeries
Authors: Elham Alaee, Mousa Shamsi, Hossein Ahmadi, Soroosh Nazem, Mohammadhossein Sedaaghi
Abstract:
Human face has a fundamental role in the appearance of individuals. So the importance of facial surgeries is undeniable. Thus, there is a need for the appropriate and accurate facial skin segmentation in order to extract different features. Since Fuzzy CMeans (FCM) clustering algorithm doesn’t work appropriately for noisy images and outliers, in this paper we exploit Possibilistic CMeans (PCM) algorithm in order to segment the facial skin. For this purpose, first, we convert facial images from RGB to YCbCr color space. To evaluate performance of the proposed algorithm, the database of Sahand University of Technology, Tabriz, Iran was used. In order to have a better understanding from the proposed algorithm; FCM and Expectation-Maximization (EM) algorithms are also used for facial skin segmentation. The proposed method shows better results than the other segmentation methods. Results include misclassification error (0.032) and the region’s area error (0.045) for the proposed algorithm.
Keywords: Facial image, segmentation, PCM, FCM, skin error, facial surgery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1989551 Fast Facial Feature Extraction and Matching with Artificial Face Models
Authors: Y. H. Tsai, Y. W. Chen
Abstract:
Facial features are frequently used to represent local properties of a human face image in computer vision applications. In this paper, we present a fast algorithm that can extract the facial features online such that they can give a satisfying representation of a face image. It includes one step for a coarse detection of each facial feature by AdaBoost and another one to increase the accuracy of the found points by Active Shape Models (ASM) in the regions of interest. The resulted facial features are evaluated by matching with artificial face models in the applications of physiognomy. The distance measure between the features and those in the fate models from the database is carried out by means of the Hausdorff distance. In the experiment, the proposed method shows the efficient performance in facial feature extractions and online system of physiognomy.Keywords: Facial feature extraction, AdaBoost, Active shapemodel, Hausdorff distance
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1809550 Facial Expression Phoenix (FePh): An Annotated Sequenced Dataset for Facial and Emotion-Specified Expressions in Sign Language
Authors: Marie Alaghband, Niloofar Yousefi, Ivan Garibay
Abstract:
Facial expressions are important parts of both gesture and sign language recognition systems. Despite the recent advances in both fields, annotated facial expression datasets in the context of sign language are still scarce resources. In this manuscript, we introduce an annotated sequenced facial expression dataset in the context of sign language, comprising over 3000 facial images extracted from the daily news and weather forecast of the public tv-station PHOENIX. Unlike the majority of currently existing facial expression datasets, FePh provides sequenced semi-blurry facial images with different head poses, orientations, and movements. In addition, in the majority of images, identities are mouthing the words, which makes the data more challenging. To annotate this dataset we consider primary, secondary, and tertiary dyads of seven basic emotions of "sad", "surprise", "fear", "angry", "neutral", "disgust", and "happy". We also considered the "None" class if the image’s facial expression could not be described by any of the aforementioned emotions. Although we provide FePh as a facial expression dataset of signers in sign language, it has a wider application in gesture recognition and Human Computer Interaction (HCI) systems.Keywords: Annotated Facial Expression Dataset, Sign Language Recognition, Gesture Recognition, Sequenced Facial Expression Dataset.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 717549 An Efficient Algorithm for Motion Detection Based Facial Expression Recognition using Optical Flow
Authors: Ahmad R. Naghsh-Nilchi, Mohammad Roshanzamir
Abstract:
One of the popular methods for recognition of facial expressions such as happiness, sadness and surprise is based on deformation of facial features. Motion vectors which show these deformations can be specified by the optical flow. In this method, for detecting emotions, the resulted set of motion vectors are compared with standard deformation template that caused by facial expressions. In this paper, a new method is introduced to compute the quantity of likeness in order to make decision based on the importance of obtained vectors from an optical flow approach. For finding the vectors, one of the efficient optical flow method developed by Gautama and VanHulle[17] is used. The suggested method has been examined over Cohn-Kanade AU-Coded Facial Expression Database, one of the most comprehensive collections of test images available. The experimental results show that our method could correctly recognize the facial expressions in 94% of case studies. The results also show that only a few number of image frames (three frames) are sufficient to detect facial expressions with rate of success of about 83.3%. This is a significant improvement over the available methods.Keywords: Facial expression, Facial features, Optical flow, Motion vectors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2375548 A Polyimide Based Split-Ring Neural Interface Electrode for Neural Signal Recording
Authors: Ning Xue, Srinivas Merugu, Ignacio Delgado Martinez, Tao Sun, John Tsang, Shih-Cheng Yen
Abstract:
We have developed a polyimide based neural interface electrode to record nerve signals from the sciatic nerve of a rat. The neural interface electrode has a split-ring shape, with four protruding gold electrodes for recording, and two reference gold electrodes around the split-ring. The split-ring electrode can be opened up to encircle the sciatic nerve. The four electrodes can be bent to sit on top of the nerve and hold the device in position, while the split-ring frame remains flat. In comparison, while traditional cuff electrodes can only fit certain sizes of the nerve, the developed device can fit a variety of rat sciatic nerve dimensions from 0.6 mm to 1.0 mm, and adapt to the chronic changes in the nerve as the electrode tips are bendable. The electrochemical impedance spectroscopy measurement was conducted. The gold electrode impedance is on the order of 10 kΩ, showing excellent charge injection capacity to record neural signals.
Keywords: Impedance, neural interface, split-ring electrode.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2597547 Human Facial Expression Recognition using MANFIS Model
Authors: V. Gomathi, Dr. K. Ramar, A. Santhiyaku Jeevakumar
Abstract:
Facial expression analysis plays a significant role for human computer interaction. Automatic analysis of human facial expression is still a challenging problem with many applications. In this paper, we propose neuro-fuzzy based automatic facial expression recognition system to recognize the human facial expressions like happy, fear, sad, angry, disgust and surprise. Initially facial image is segmented into three regions from which the uniform Local Binary Pattern (LBP) texture features distributions are extracted and represented as a histogram descriptor. The facial expressions are recognized using Multiple Adaptive Neuro Fuzzy Inference System (MANFIS). The proposed system designed and tested with JAFFE face database. The proposed model reports 94.29% of classification accuracy.Keywords: Adaptive neuro-fuzzy inference system, Facialexpression, Local binary pattern, Uniform Histogram
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2102546 The Effect of Facial Expressions on Students in Virtual Educational Environments
Authors: G. Theonas, D. Hobbs, D. Rigas
Abstract:
The scope of this research was to study the relation between the facial expressions of three lecturers in a real academic lecture theatre and the reactions of the students to those expressions. The first experiment aimed to investigate the effectiveness of a virtual lecturer-s expressions on the students- learning outcome in a virtual pedagogical environment. The second experiment studied the effectiveness of a single facial expression, i.e. the smile, on the students- performance. Both experiments involved virtual lectures, with virtual lecturers teaching real students. The results suggest that the students performed better by 86%, in the lectures where the lecturer performed facial expressions compared to the results of the lectures that did not use facial expressions. However, when simple or basic information was used, the facial expressions of the virtual lecturer had no substantial effect on the students- learning outcome. Finally, the appropriate use of smiles increased the interest of the students and consequently their performance.
Keywords: emotion, facial expression, smile, virtual educational environment, virtual learning, virtual lecturer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1985545 Comparing Emotion Recognition from Voice and Facial Data Using Time Invariant Features
Authors: Vesna Kirandziska, Nevena Ackovska, Ana Madevska Bogdanova
Abstract:
The problem of emotion recognition is a challenging problem. It is still an open problem from the aspect of both intelligent systems and psychology. In this paper, both voice features and facial features are used for building an emotion recognition system. A Support Vector Machine classifiers are built by using raw data from video recordings. In this paper, the results obtained for the emotion recognition are given, and a discussion about the validity and the expressiveness of different emotions is presented. A comparison between the classifiers build from facial data only, voice data only and from the combination of both data is made here. The need for a better combination of the information from facial expression and voice data is argued.
Keywords: Emotion recognition, facial recognition, signal processing, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2017544 Improving the Performance of Deep Learning in Facial Emotion Recognition with Image Sharpening
Authors: Ksheeraj Sai Vepuri, Nada Attar
Abstract:
We as humans use words with accompanying visual and facial cues to communicate effectively. Classifying facial emotion using computer vision methodologies has been an active research area in the computer vision field. In this paper, we propose a simple method for facial expression recognition that enhances accuracy. We tested our method on the FER-2013 dataset that contains static images. Instead of using Histogram equalization to preprocess the dataset, we used Unsharp Mask to emphasize texture and details and sharpened the edges. We also used ImageDataGenerator from Keras library for data augmentation. Then we used Convolutional Neural Networks (CNN) model to classify the images into 7 different facial expressions, yielding an accuracy of 69.46% on the test set. Our results show that using image preprocessing such as the sharpening technique for a CNN model can improve the performance, even when the CNN model is relatively simple.Keywords: Facial expression recognition, image pre-processing, deep learning, CNN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 543543 Modeling of Radiofrequency Nerve Lesioning in Inhomogeneous Media
Authors: Nour Ismail, Sahar El Kardawy, Bassant Badwy
Abstract:
Radiofrequency (RF) lesioning of nerves have been commonly used to alleviate chronic pain, where RF current preventing transmission of pain signals through the nerve by heating the nerve causing the pain. There are some factors that affect the temperature distribution and the nerve lesion size, one of these factors is the inhomogeneities in the tissue medium. Our objective is to calculate the temperature distribution and the nerve lesion size in an inhomogeneous medium surrounding the RF electrode. A two 3-D finite element models are used to compare the temperature distribution in the homogeneous and inhomogeneous medium. Also the effect of temperature-dependent electric conductivity on maximum temperature and lesion size is observed. Results show that the presence of an inhomogeneous medium around the RF electrode has a valuable effect on the temperature distribution and lesion size. The dependency of electric conductivity on tissue temperature increased lesion size.
Keywords: Finite element model, nerve lesioning, pain relief, radiofrequency lesion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1971542 RBF Based Face Recognition and Expression Analysis
Authors: Praseeda Lekshmi.V, Dr.M.Sasikumar
Abstract:
Facial recognition and expression analysis is rapidly becoming an area of intense interest in computer science and humancomputer interaction design communities. The most expressive way humans display emotions is through facial expressions. In this paper skin and non-skin pixels were separated. Face regions were extracted from the detected skin regions. Facial expressions are analyzed from facial images by applying Gabor wavelet transform (GWT) and Discrete Cosine Transform (DCT) on face images. Radial Basis Function (RBF) Network is used to identify the person and to classify the facial expressions. Our method reliably works even with faces, which carry heavy expressions.Keywords: Face Recognition, Radial Basis Function, Gabor Wavelet Transform, Discrete Cosine Transform
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1594541 A Neural Network Based Facial Expression Analysis using Gabor Wavelets
Authors: Praseeda Lekshmi.V, Dr.M.Sasikumar
Abstract:
Facial expression analysis is rapidly becoming an area of intense interest in computer science and human-computer interaction design communities. The most expressive way humans display emotions is through facial expressions. In this paper we present a method to analyze facial expression from images by applying Gabor wavelet transform (GWT) and Discrete Cosine Transform (DCT) on face images. Radial Basis Function (RBF) Network is used to classify the facial expressions. As a second stage, the images are preprocessed to enhance the edge details and non uniform down sampling is done to reduce the computational complexity and processing time. Our method reliably works even with faces, which carry heavy expressions.Keywords: Face Expression, Radial Basis Function, GaborWavelet Transform, Human Computer Interaction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2105540 Optimized Facial Features-based Age Classification
Authors: Md. Zahangir Alom, Mei-Lan Piao, Md. Shariful Islam, Nam Kim, Jae-Hyeung Park
Abstract:
The evaluation and measurement of human body dimensions are achieved by physical anthropometry. This research was conducted in view of the importance of anthropometric indices of the face in forensic medicine, surgery, and medical imaging. The main goal of this research is to optimization of facial feature point by establishing a mathematical relationship among facial features and used optimize feature points for age classification. Since selected facial feature points are located to the area of mouth, nose, eyes and eyebrow on facial images, all desire facial feature points are extracted accurately. According this proposes method; sixteen Euclidean distances are calculated from the eighteen selected facial feature points vertically as well as horizontally. The mathematical relationships among horizontal and vertical distances are established. Moreover, it is also discovered that distances of the facial feature follows a constant ratio due to age progression. The distances between the specified features points increase with respect the age progression of a human from his or her childhood but the ratio of the distances does not change (d = 1 .618 ) . Finally, according to the proposed mathematical relationship four independent feature distances related to eight feature points are selected from sixteen distances and eighteen feature point-s respectively. These four feature distances are used for classification of age using Support Vector Machine (SVM)-Sequential Minimal Optimization (SMO) algorithm and shown around 96 % accuracy. Experiment result shows the proposed system is effective and accurate for age classification.Keywords: 3D Face Model, Face Anthropometrics, Facial Features Extraction, Feature distances, SVM-SMO
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2046539 Deep-Learning Based Approach to Facial Emotion Recognition Through Convolutional Neural Network
Authors: Nouha Khediri, Mohammed Ben Ammar, Monji Kherallah
Abstract:
Recently, facial emotion recognition (FER) has become increasingly essential to understand the state of the human mind. However, accurately classifying emotion from the face is a challenging task. In this paper, we present a facial emotion recognition approach named CV-FER benefiting from deep learning, especially CNN and VGG16. First, the data are pre-processed with data cleaning and data rotation. Then, we augment the data and proceed to our FER model, which contains five convolutions layers and five pooling layers. Finally, a softmax classifier is used in the output layer to recognize emotions. Based on the above contents, this paper reviews the works of facial emotion recognition based on deep learning. Experiments show that our model outperforms the other methods using the same FER2013 database and yields a recognition rate of 92%. We also put forward some suggestions for future work.
Keywords: CNN, deep-learning, facial emotion recognition, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 708538 Emotion Classification using Adaptive SVMs
Authors: P. Visutsak
Abstract:
The study of the interaction between humans and computers has been emerging during the last few years. This interaction will be more powerful if computers are able to perceive and respond to human nonverbal communication such as emotions. In this study, we present the image-based approach to emotion classification through lower facial expression. We employ a set of feature points in the lower face image according to the particular face model used and consider their motion across each emotive expression of images. The vector of displacements of all feature points input to the Adaptive Support Vector Machines (A-SVMs) classifier that classify it into seven basic emotions scheme, namely neutral, angry, disgust, fear, happy, sad and surprise. The system was tested on the Japanese Female Facial Expression (JAFFE) dataset of frontal view facial expressions [7]. Our experiments on emotion classification through lower facial expressions demonstrate the robustness of Adaptive SVM classifier and verify the high efficiency of our approach.Keywords: emotion classification, facial expression, adaptive support vector machines, facial expression classifier.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2223537 Robust Face Recognition using AAM and Gabor Features
Authors: Sanghoon Kim, Sun-Tae Chung, Souhwan Jung, Seoungseon Jeon, Jaemin Kim, Seongwon Cho
Abstract:
In this paper, we propose a face recognition algorithm using AAM and Gabor features. Gabor feature vectors which are well known to be robust with respect to small variations of shape, scaling, rotation, distortion, illumination and poses in images are popularly employed for feature vectors for many object detection and recognition algorithms. EBGM, which is prominent among face recognition algorithms employing Gabor feature vectors, requires localization of facial feature points where Gabor feature vectors are extracted. However, localization method employed in EBGM is based on Gabor jet similarity and is sensitive to initial values. Wrong localization of facial feature points affects face recognition rate. AAM is known to be successfully applied to localization of facial feature points. In this paper, we devise a facial feature point localization method which first roughly estimate facial feature points using AAM and refine facial feature points using Gabor jet similarity-based facial feature localization method with initial points set by the rough facial feature points obtained from AAM, and propose a face recognition algorithm using the devised localization method for facial feature localization and Gabor feature vectors. It is observed through experiments that such a cascaded localization method based on both AAM and Gabor jet similarity is more robust than the localization method based on only Gabor jet similarity. Also, it is shown that the proposed face recognition algorithm using this devised localization method and Gabor feature vectors performs better than the conventional face recognition algorithm using Gabor jet similarity-based localization method and Gabor feature vectors like EBGM.Keywords: Face Recognition, AAM, Gabor features, EBGM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2205536 Spiral Cuff for Fiber-Diameter Selective VNS
Abstract:
In this paper we present the modeling, design, and experimental testing of a nerve cuff multi-electrode system for diameter-selective vagus nerve stimulation. The multi-electrode system contained ninety-nine platinum electrodes embedded within a self-curling spiral silicone sheet. The electrodes were organized in a matrix having nine parallel groups, each containing eleven electrodes. Preliminary testing of the nerve cuff was performed in an isolated segment of a swinish left cervical vagus nerve. For selective vagus nerve stimulation, precisely defined current quasitrapezoidal, asymmetric and biphasic stimulating pulses were applied to preselected locations along the left vagus segment via appointed group of three electrodes within the cuff. Selective stimulation was obtained by anodal block. However, these pulses may not be safe for a long-term application because of a frequently used high imbalance between the cathodic and anodic part of the stimulating pulse. Preliminary results show that the cuff was capable of exciting A and B-fibres, and, that for a certain range of parameters used in stimulating pulses, the contribution of A-fibres to the CAP was slightly reduced and the contribution of B-fibres was slightly larger. Results also showed that measured CAPs are not greatly influenced by the imbalance between a charge Qc injected in cathodic and Qa in anodic phase of quasitrapezoidal, asymmetric and biphasic pulses.Keywords: Vagus nerve stimulation, multi-electrode nerve cuff.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678535 Curvelet Features with Mouth and Face Edge Ratios for Facial Expression Identification
Authors: S. Kherchaoui, A. Houacine
Abstract:
This paper presents a facial expression recognition system. It performs identification and classification of the seven basic expressions; happy, surprise, fear, disgust, sadness, anger, and neutral states. It consists of three main parts. The first one is the detection of a face and the corresponding facial features to extract the most expressive portion of the face, followed by a normalization of the region of interest. Then calculus of curvelet coefficients is performed with dimensionality reduction through principal component analysis. The resulting coefficients are combined with two ratios; mouth ratio and face edge ratio to constitute the whole feature vector. The third step is the classification of the emotional state using the SVM method in the feature space.
Keywords: Facial expression identification, curvelet coefficients, support vector machine (SVM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1841534 Development of Face Surrogate for Impact Protection Design for Cyclist
Authors: Sanga Monthatipkul, Pio Iovenitti, Igor Sbarski
Abstract:
Bicycle usage for exercise, recreation, and commuting to work in Australia shows that pedal cycling is the fourth most popular activity with 10.6% increase in participants between 2001 and 2007. As with other means of transport, accident and injury becomes common although mandatory bicycle helmet wearing has been introduced. The research aims to develop a face surrogate made of sandwich of rigid foam and rubber sheets to represent human facial bone under blunt impact. The facial surrogate will serve as an important test device for further development of facial-impact protection for cyclist. A test procedure was developed to simulate the energy of impact and record data to evaluate the effect of impact on facial bones. Drop tests were performed to establish a suitable combination of materials. It was found that the sandwich structure of rigid extruded-polystyrene foam (density of 40 kg/m3 with a pattern of 6-mm-holes), Neoprene rubber sponge, and Abrasaflex rubber backing, had impact characteristics comparable to that of human facial bone. In particular, the foam thickness of 30 mm and 25 mm was found suitable to represent human zygoma (cheekbone) and maxilla (upper-jaw bone), respectively.Keywords: Facial impact protection, face surrogate, cyclist, accident prevention
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1525533 Matching Facial Images using Age Related Morphing Changes
Authors: Udeni Jayasinghe, Anuja Dharmaratne
Abstract:
Each year many people are reported missing in most of the countries in the world owing to various reasons. Arrangements have to be made to find these people after some time. So the investigating agencies are compelled to make out these people by using manpower. But in many cases, the investigations carried out to find out an absconding for a long time may not be successful. At a time like that it may be difficult to identify these people by examining their old photographs, because their facial appearance might have changed mainly due to the natural aging process. On some occasions in forensic medicine if a dead body is found, investigations should be held to make sure that this corpse belongs to the same person disappeared some time ago. With the passage of time the face of the person might have changed and there should be a mechanism to reveal the person-s identity. In order to make this process easy, we must guess and decide as to how he will look like by now. To address this problem this paper presents a way of synthesizing a facial image with the aging effects.
Keywords: Cranio-facial growth model, eigenfaces, eigenvectors, Face Anthropometry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1739532 The Service Failure and Recovery in the Information Technology Services
Authors: Jun Luo, Weiguo Zhang., Dabin Qin
Abstract:
It is important to retain customer satisfaction in information technology services. When a service failure occurs, companies need to take service recovery action to recover their customer satisfaction. Although companies cannot avoid all problems and complaints, they should try to make up. Therefore, service failure and service recovery have become an important and challenging issue for companies. In this paper, the literature and the problems in the information technology services were reviewed. An integrated model of profit driven for the service failure and service recovery was established in view of the benefit of customer and enterprise. Moreover, the interaction between service failure and service recovery strategy was studied, the result of which verified the matching principles of the service recovery strategy and the type of service failure. In addition, the relationship between the cost of service recovery and customer-s cumulative value of service after recovery was analyzed with the model. The result attributes to managers in deciding on appropriate resource allocations for recovery strategies.Keywords: service failure, service recovery, informationtechnology services
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2104531 Reliable Face Alignment Using Two-Stage AAM
Authors: Sunho Ki, Daehwan Kim, Seongwon Cho, Sun-Tae Chung, Jaemin Kim, Yun-Kwang Hong, Chang Joon Park, Dongmin Kwon, Minhee Kang, Yusung Kim, Younghan Yoon
Abstract:
AAM (active appearance model) has been successfully applied to face and facial feature localization. However, its performance is sensitive to initial parameter values. In this paper, we propose a two-stage AAM for robust face alignment, which first fits an inner face-AAM model to the inner facial feature points of the face and then localizes the whole face and facial features by optimizing the whole face-AAM model parameters. Experiments show that the proposed face alignment method using two-stage AAM is more reliable to the background and the head pose than the standard AAM-based face alignment method.Keywords: AAM, Face Alignment, Feature Extraction, PCA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1476530 Application of Sensory Thermography as Measuring Method to Study Median Nerve Temperatures
Authors: Javier Ordorica Villalvazo, Claudia Camargo Wilson, Jesus Everardo Olguin Tiznado
Abstract:
This paper presents an experimental case using sensory thermography to describe temperatures behavior on median nerve once an activity of repetitive motion was done. Thermography is a noninvasive technique without biological hazard and not harm at all times and has been applied in many experiments to seek for temperature patterns that help to understand diseases like cancer and cumulative trauma disorders (CTD’s). An infrared sensory thermography technology was developed to execute this study. Three women in good shape were selected for the repetitive motion tests for 4 days, two right-handed women and 1 left handed woman, two sensory thermographers were put on both median nerve wrists to get measures. The evaluation time was of 3 hours 30 minutes in a controlled temperature, 20 minutes of stabilization time at the beginning and end of the operation. Temperatures distributions are statistically evaluated and showed similar temperature patterns behavior.
Keywords: Median nerve, temperature, sensory thermography, wrists, CTD’s.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1488529 Bioarm, a Prostheses without Surgery
Authors: J. Sagouis, A. Chamel, E. Carre, C. Casasreales, G. Rudnik, M. Cerdan
Abstract:
Robotics provides answers to amputees. The most expensive solutions surgically connect the prosthesis to nerve endings. There are also several types of non-invasive technologies that recover nerve messages passing through the muscles. After analyzing these messages, myoelectric prostheses perform the desired movement. The main goal is to avoid all surgeries, which can be heavy and offer cheaper alternatives. For an amputee, we use valid muscles to recover the electrical signal involved in a muscle movement. EMG sensors placed on the muscle allows us to measure a potential difference, which our program transforms into control for a robotic arm with two degrees of freedom. We have shown the feasibility of non-invasive prostheses with two degrees of freedom. Signal analysis and an increase in degrees of freedom is still being improved.
Keywords: Prosthesis, electromyography (EMG), robotic arm, nerve message.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1873528 Development System for Emotion Detection Based on Brain Signals and Facial Images
Authors: Suprijanto, Linda Sari, Vebi Nadhira , IGN. Merthayasa. Farida I.M
Abstract:
Detection of human emotions has many potential applications. One of application is to quantify attentiveness audience in order evaluate acoustic quality in concern hall. The subjective audio preference that based on from audience is used. To obtain fairness evaluation of acoustic quality, the research proposed system for multimodal emotion detection; one modality based on brain signals that measured using electroencephalogram (EEG) and the second modality is sequences of facial images. In the experiment, an audio signal was customized which consist of normal and disorder sounds. Furthermore, an audio signal was played in order to stimulate positive/negative emotion feedback of volunteers. EEG signal from temporal lobes, i.e. T3 and T4 was used to measured brain response and sequence of facial image was used to monitoring facial expression during volunteer hearing audio signal. On EEG signal, feature was extracted from change information in brain wave, particularly in alpha and beta wave. Feature of facial expression was extracted based on analysis of motion images. We implement an advance optical flow method to detect the most active facial muscle form normal to other emotion expression that represented in vector flow maps. The reduce problem on detection of emotion state, vector flow maps are transformed into compass mapping that represents major directions and velocities of facial movement. The results showed that the power of beta wave is increasing when disorder sound stimulation was given, however for each volunteer was giving different emotion feedback. Based on features derived from facial face images, an optical flow compass mapping was promising to use as additional information to make decision about emotion feedback.
Keywords: Multimodal Emotion Detection, EEG, Facial Image, Optical Flow, compass mapping, Brain Wave
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2291527 Face Texture Reconstruction for Illumination Variant Face Recognition
Authors: Pengfei Xiong, Lei Huang, Changping Liu
Abstract:
In illumination variant face recognition, existing methods extracting face albedo as light normalized image may lead to loss of extensive facial details, with light template discarded. To improve that, a novel approach for realistic facial texture reconstruction by combining original image and albedo image is proposed. First, light subspaces of different identities are established from the given reference face images; then by projecting the original and albedo image into each light subspace respectively, texture reference images with corresponding lighting are reconstructed and two texture subspaces are formed. According to the projections in texture subspaces, facial texture with normal light can be synthesized. Due to the combination of original image, facial details can be preserved with face albedo. In addition, image partition is applied to improve the synthesization performance. Experiments on Yale B and CMUPIE databases demonstrate that this algorithm outperforms the others both in image representation and in face recognition.Keywords: texture reconstruction, illumination, face recognition, subspaces
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1481