Search results for: automatic facial expression analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 28692

Search results for: automatic facial expression analysis

28632 Automatic Moment-Based Texture Segmentation

Authors: Tudor Barbu

Abstract:

An automatic moment-based texture segmentation approach is proposed in this paper. First, we describe the related work in this computer vision domain. Our texture feature extraction, the first part of the texture recognition process, produces a set of moment-based feature vectors. For each image pixel, a texture feature vector is computed as a sequence of area moments. Second, an automatic pixel classification approach is proposed. The feature vectors are clustered using some unsupervised classification algorithm, the optimal number of clusters being determined using a measure based on validation indexes. From the resulted pixel classes one determines easily the desired texture regions of the image.

Keywords: image segmentation, moment-based, texture analysis, automatic classification, validation indexes

Procedia PDF Downloads 386
28631 Multimodal Database of Emotional Speech, Video and Gestures

Authors: Tomasz Sapiński, Dorota Kamińska, Adam Pelikant, Egils Avots, Cagri Ozcinar, Gholamreza Anbarjafari

Abstract:

People express emotions through different modalities. Integration of verbal and non-verbal communication channels creates a system in which the message is easier to understand. Expanding the focus to several expression forms can facilitate research on emotion recognition as well as human-machine interaction. In this article, the authors present a Polish emotional database composed of three modalities: facial expressions, body movement and gestures, and speech. The corpora contains recordings registered in studio conditions, acted out by 16 professional actors (8 male and 8 female). The data is labeled with six basic emotions categories, according to Ekman’s emotion categories. To check the quality of performance, all recordings are evaluated by experts and volunteers. The database is available to academic community and might be useful in the study on audio-visual emotion recognition.

Keywords: body movement, emotion recognition, emotional corpus, facial expressions, gestures, multimodal database, speech

Procedia PDF Downloads 322
28630 In vivo Mechanical Characterization of Facial Skin Combining Digital Image Correlation and Finite Element

Authors: Huixin Wei, Shibin Wang, Linan Li, Lei Zhou, Xinhao Tu

Abstract:

Facial skin is a biomedical material with complex mechanical properties of anisotropy, viscoelasticity, and hyperelasticity. The mechanical properties of facial skin are crucial for a number of applications including facial plastic surgery, animation, dermatology, cosmetic industry, and impact biomechanics. Skin is a complex multi-layered material which can be broadly divided into three main layers, the epidermis, the dermis, and the hypodermis. Collagen fibers account for 75% of the dry weight of dermal tissue, and it is these fibers which are responsible for the mechanical properties of skin. Many research on the anisotropic mechanical properties are mainly concentrated on in vitro, but there is a great difference between in vivo and in vitro for mechanical properties of the skin. In this study, we presented a method to measure the mechanical properties of facial skin in vivo. Digital image correlation (DIC) and indentation tests were used to obtain the experiment data, including the deformation of facial surface and indentation force-displacement curve. Then, the experiment was simulated using a finite element (FE) model. Application of Computed Tomography (CT) and reconstruction techniques obtained the real tissue geometry. A three-dimensional FE model of facial skin, including a bi-layer system, was obtained. As the epidermis is relatively thin, the epidermis and dermis were regarded as one layer and below it was hypodermis in this study. The upper layer was modeled as a Gasser-Ogden-Holzapfel (GOH) model to describe hyperelastic and anisotropic behaviors of the dermis. The under layer was modeled as a linear elastic model. In conclusion, the material properties of two-layer were determined by minimizing the error between the FE data and experimental data.

Keywords: facial skin, indentation test, finite element, digital image correlation, computed tomography

Procedia PDF Downloads 89
28629 Correlation between Cephalometric Measurements and Visual Perception of Facial Profile in Skeletal Type II Patients

Authors: Choki, Supatchai Boonpratham, Suwannee Luppanapornlarp

Abstract:

The objective of this study was to find a correlation between cephalometric measurements and visual perception of facial profile in skeletal type II patients. In this study, 250 lateral cephalograms of female patients from age, 20 to 22 years were analyzed. The profile outlines of all the samples were hand traced and transformed into silhouettes by the principal investigator. Profile ratings were done by 9 orthodontists on Visual Analogue Scale from score one to ten (increasing level of convexity). 37 hard issue and soft tissue cephalometric measurements were analyzed by the principal investigator. All the measurements were repeated after 2 weeks interval for error assessment. At last, the rankings of visual perceptions were correlated with cephalometric measurements using Spearman correlation coefficient (P < 0.05). The results show that the increase in facial convexity was correlated with higher values of ANB (A point, nasion and B point), AF-BF (distance from A point to B point in mm), L1-NB (distance from lower incisor to NB line in mm), anterior maxillary alveolar height, posterior maxillary alveolar height, overjet, H angle hard tissue, H angle soft tissue and lower lip to E plane (absolute correlation values from 0.277 to 0.711). In contrast, the increase in facial convexity was correlated with lower values of Pg. to N perpendicular and Pg. to NB (mm) (absolute correlation value -0.302 and -0.294 respectively). From the soft tissue measurements, H angles had a higher correlation with visual perception than facial contour angle, nasolabial angle, and lower lip to E plane. In conclusion, the findings of this study indicated that the correlation of cephalometric measurements with visual perception was less than expected. Only 29% of cephalometric measurements had a significant correlation with visual perception. Therefore, diagnosis based solely on cephalometric analysis can result in failure to meet the patient’s esthetic expectation.

Keywords: cephalometric measurements, facial profile, skeletal type II, visual perception

Procedia PDF Downloads 113
28628 Analysis of The Effect about Different Automatic Sprinkler System Extinguishing The Scooter Fire in Underground Parking Space

Authors: Yu-Hsiu Li, Chun-Hsun Chen

Abstract:

Analysis of automatic sprinkler system protects the scooter in underground parking space, the current of general buildings is mainly equipped with foam fire-extinguishing equipment in Taiwan, the automatic sprinkling system has economic and environmental benefits, even high stability, China and the United States allow the parking space to set the automatic sprinkler system under certain conditions. The literature about scooter full-scale fire indicates that the average fire growth coefficient is 0.19 KW/sec2, it represents the scooter fire is classified as ultra-fast time square fire growth model, automatic sprinkler system can suppress the flame height and prevent extending burning. According to the computer simulation (FDS) literature, no matter computer simulation or full-scale experiments, the active order and trend about sprinkler heads are the same. This study uses the computer simulation program (FDS), the simulation scenario designed includes using a different system (enclosed wet type and open type), and different configurations. The simulation result demonstrates that the open type requires less time to extinguish the fire than the enclosed wet type if the horizontal distance between the sprinkler and the scooter ignition source is short, the sprinkler can act quickly, the heat release rate of fire can be suppressed in advance.

Keywords: automatic sprinkler system, underground parking Spac, FDS, scooter fire extinguishing

Procedia PDF Downloads 114
28627 When and Why Unhappy People Avoid Enjoyable Experiences

Authors: Hao Shen, Aparna Labroo

Abstract:

Across four studies, we show people in a negative mood avoid anticipated enjoyable experiences because of the subjective difficulty in simulating those experiences, and they misattribute these feelings of difficulty to reduced pleasantness of the anticipated experience. We observe the avoidance of enjoyable experiences only for anticipated experiences that involve smile-like facial-muscular simulation. When the need for facial-muscular simulation is attenuated, or when the anticipated experience relies on facial-muscular simulation to a lesser extent, people in a negative mood no longer avoid enjoyable experiences, but rather seek such experiences because they fit better with their ongoing mood-repair goals.

Keywords: emotion regulation, mood repair, embodiment, anticipated experiences

Procedia PDF Downloads 391
28626 Characterising the Processes Underlying Emotion Recognition Deficits in Adolescents with Conduct Disorder

Authors: Nayra Martin-Key, Erich Graf, Wendy Adams, Graeme Fairchild

Abstract:

Children and adolescents with Conduct Disorder (CD) have been shown to demonstrate impairments in emotion recognition, but it is currently unclear whether this deficit is related to specific emotions or whether it represents a global deficit in emotion recognition. An emotion recognition task with concurrent eye-tracking was employed to further explore this relationship in a sample of male and female adolescents with CD. Participants made emotion categorization judgements for presented dynamic and morphed static facial expressions. The results demonstrated that males with CD, and to a lesser extent, females with CD, displayed impaired facial expression recognition in general, whereas callous-unemotional (CU) traits were linked to specific problems in sadness recognition in females with CD. A region-of-interest analysis of the eye-tracking data indicated that males with CD exhibited reduced fixation times for the eye-region of the face compared to typically-developing (TD) females, but not TD males. Females with CD did not show reduced fixation to the eye-region of the face relative to TD females. In addition, CU traits did not influence CD subjects’ attention to the eye-region of the face. These findings suggest that the emotion recognition deficits found in CD males, the worst performing group in the behavioural tasks, are partly driven by reduced attention to the eyes.

Keywords: attention, callous-unemotional traits, conduct disorder, emotion recognition, eye-region, eye-tracking, sex differences

Procedia PDF Downloads 278
28625 Monocular 3D Person Tracking AIA Demographic Classification and Projective Image Processing

Authors: McClain Thiel

Abstract:

Object detection and localization has historically required two or more sensors due to the loss of information from 3D to 2D space, however, most surveillance systems currently in use in the real world only have one sensor per location. Generally, this consists of a single low-resolution camera positioned above the area under observation (mall, jewelry store, traffic camera). This is not sufficient for robust 3D tracking for applications such as security or more recent relevance, contract tracing. This paper proposes a lightweight system for 3D person tracking that requires no additional hardware, based on compressed object detection convolutional-nets, facial landmark detection, and projective geometry. This approach involves classifying the target into a demographic category and then making assumptions about the relative locations of facial landmarks from the demographic information, and from there using simple projective geometry and known constants to find the target's location in 3D space. Preliminary testing, although severely lacking, suggests reasonable success in 3D tracking under ideal conditions.

Keywords: monocular distancing, computer vision, facial analysis, 3D localization

Procedia PDF Downloads 106
28624 Exploring Multimodal Communication: Intersections of Language, Gesture, and Technology

Authors: Rasha Ali Dheyab

Abstract:

In today's increasingly interconnected and technologically-driven world, communication has evolved beyond traditional verbal exchanges. This paper delves into the fascinating realm of multimodal communication, a dynamic field at the intersection of linguistics, gesture studies, and technology. The study of how humans convey meaning through a combination of spoken language, gestures, facial expressions, and digital platforms has gained prominence as our modes of interaction continue to diversify. This exploration begins by examining the foundational theories in linguistics and gesture studies, tracing their historical development and mutual influences. It further investigates the role of nonverbal cues, such as gestures and facial expressions, in augmenting and sometimes even altering the meanings conveyed by spoken language. Additionally, the paper delves into the modern technological landscape, where emojis, GIFs, and other digital symbols have emerged as new linguistic tools, reshaping the ways in which we communicate and express emotions. The interaction between traditional and digital modes of communication is a central focus of this study. The paper investigates how technology has not only introduced new modes of expression but has also influenced the adaptation of existing linguistic and gestural patterns in online discourse. The emergence of virtual reality and augmented reality environments introduces yet another layer of complexity to multimodal communication, offering new avenues for studying how humans navigate and negotiate meaning in immersive digital spaces. Through a combination of literature review, case studies, and theoretical analysis, this paper seeks to shed light on the intricate interplay between language, gesture, and technology in the realm of multimodal communication. By understanding how these diverse modes of expression intersect and interact, we gain valuable insights into the ever-evolving nature of human communication and its implications for fields ranging from linguistics and psychology to human-computer interaction and digital anthropology.

Keywords: multimodal communication, linguistics ., gesture studies., emojis., verbal communication., digital

Procedia PDF Downloads 49
28623 Dominant Ideology among Filipino Women as Dictated by Cosmopolitan Magazine

Authors: Yvonne Christelle M. de Guzman, Charity Faye T. Cabie

Abstract:

This study analyzed the contents of ten issues of Cosmopolitan Magazine from 2011 to 2015. The researcher found out the hegemony among Filipino women as dictated by Cosmopolitan magazine through the use of Semiotic Analysis, Laura Mulvey’s Male Gaze and Gramsci’s concept of hegemony. The researcher also looked at the themes of cover stories, words used to describe women, meanings behind the color of magazine’s front cover, clothing, physique and pose such as gesture and facial expression used by the cover girl. However, the entire content of the magazine was not taken into account.

Keywords: dominant ideology, male gaze, semiotics, women

Procedia PDF Downloads 360
28622 Digi-Buddy: A Smart Cane with Artificial Intelligence and Real-Time Assistance

Authors: Amaladhithyan Krishnamoorthy, Ruvaitha Banu

Abstract:

Vision is considered as the most important sense in humans, without which leading a normal can be often difficult. There are many existing smart canes for visually impaired with obstacle detection using ultrasonic transducer to help them navigate. Though the basic smart cane increases the safety of the users, it does not help in filling the void of visual loss. This paper introduces the concept of Digi-Buddy which is an evolved smart cane for visually impaired. The cane consists for several modules, apart from the basic obstacle detection features; the Digi-Buddy assists the user by capturing video/images and streams them to the server using a wide-angled camera, which then detects the objects using Deep Convolutional Neural Network. In addition to determining what the particular image/object is, the distance of the object is assessed by the ultrasonic transducer. The sound generation application, modelled with the help of Natural Language Processing is used to convert the processed images/object into audio. The object detected is signified by its name which is transmitted to the user with the help of Bluetooth hear phones. The object detection is extended to facial recognition which maps the faces of the person the user meets in the database of face images and alerts the user about the person. One of other crucial function consists of an automatic-intimation-alarm which is triggered when the user is in an emergency. If the user recovers within a set time, a button is provisioned in the cane to stop the alarm. Else an automatic intimation is sent to friends and family about the whereabouts of the user using GPS. In addition to safety and security by the existing smart canes, the proposed concept devices to be implemented as a prototype helping visually-impaired visualize their surroundings through audio more in an amicable way.

Keywords: artificial intelligence, facial recognition, natural language processing, internet of things

Procedia PDF Downloads 320
28621 Automatic Staging and Subtype Determination for Non-Small Cell Lung Carcinoma Using PET Image Texture Analysis

Authors: Seyhan Karaçavuş, Bülent Yılmaz, Ömer Kayaaltı, Semra İçer, Arzu Taşdemir, Oğuzhan Ayyıldız, Kübra Eset, Eser Kaya

Abstract:

In this study, our goal was to perform tumor staging and subtype determination automatically using different texture analysis approaches for a very common cancer type, i.e., non-small cell lung carcinoma (NSCLC). Especially, we introduced a texture analysis approach, called Law’s texture filter, to be used in this context for the first time. The 18F-FDG PET images of 42 patients with NSCLC were evaluated. The number of patients for each tumor stage, i.e., I-II, III or IV, was 14. The patients had ~45% adenocarcinoma (ADC) and ~55% squamous cell carcinoma (SqCCs). MATLAB technical computing language was employed in the extraction of 51 features by using first order statistics (FOS), gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and Laws’ texture filters. The feature selection method employed was the sequential forward selection (SFS). Selected textural features were used in the automatic classification by k-nearest neighbors (k-NN) and support vector machines (SVM). In the automatic classification of tumor stage, the accuracy was approximately 59.5% with k-NN classifier (k=3) and 69% with SVM (with one versus one paradigm), using 5 features. In the automatic classification of tumor subtype, the accuracy was around 92.7% with SVM one vs. one. Texture analysis of FDG-PET images might be used, in addition to metabolic parameters as an objective tool to assess tumor histopathological characteristics and in automatic classification of tumor stage and subtype.

Keywords: cancer stage, cancer cell type, non-small cell lung carcinoma, PET, texture analysis

Procedia PDF Downloads 298
28620 Autonomy and Other Variables Related to the Expression of Love among Saudi Couples

Authors: Reshaa Alruwaili

Abstract:

The primary aim of this study was to examine the hypothesis presented by Self Determination theory which suggests that autonomy impacts positively the expression of love. Other hypotheses were also examined which suggest that other variables explain the expression of love, including: dyadic adjustment (dyadic consensus, dyadic satisfaction and dyadic cohesion), couple satisfaction, age, gender, the length of marriage, number of children and attachment styles. The participants were Saudi couples, which provided the opportunity to consider the influence of Saudi culture on the expression of love. A questionnaire was employed to obtain measures of all the relevant variables, including a measure of expression of love that was built from 27 items, constituting verbal, physical and caring features, and a measure of autonomy based on three features: authorship, interest-taking and susceptibility. Data were collected from both members of 34 Saudi couples. Descriptive analysis of both expression of love and autonomy was conducted. Correlation and regression were used to assess the relationships between expression of love and autonomy and other variables. Results indicated that Saudi couples who most often express their love tend to be more than somewhat autonomous. Not much difference was found between husbands and wives in expressing love, although wives were slightly more autonomous than husbands. Expression of love was enhanced by the autonomy of the participants to a greater extent when dyadic satisfaction was controlled, since the latter was negatively correlated with autonomy and had no effect on the expression of love. Basic psychological needs, dyadic consensus and dismissive-avoidant attachment improve the expression of love, while it is decreased by the number of children.

Keywords: autonomy, determination theory, expression of love, dyadic adjustment

Procedia PDF Downloads 202
28619 Combined Automatic Speech Recognition and Machine Translation in Business Correspondence Domain for English-Croatian

Authors: Sanja Seljan, Ivan Dunđer

Abstract:

The paper presents combined automatic speech recognition (ASR) for English and machine translation (MT) for English and Croatian in the domain of business correspondence. The first part presents results of training the ASR commercial system on two English data sets, enriched by error analysis. The second part presents results of machine translation performed by online tool Google Translate for English and Croatian and Croatian-English language pairs. Human evaluation in terms of usability is conducted and internal consistency calculated by Cronbach's alpha coefficient, enriched by error analysis. Automatic evaluation is performed by WER (Word Error Rate) and PER (Position-independent word Error Rate) metrics, followed by investigation of Pearson’s correlation with human evaluation.

Keywords: automatic machine translation, integrated language technologies, quality evaluation, speech recognition

Procedia PDF Downloads 452
28618 Macronutrients and the FTO Gene Expression in Hypothalamus: A Systematic Review of Experimental Studies

Authors: Saeid Doaei

Abstract:

The various studies have examined the relationship between FTO gene expression and macronutrients levels. In order to obtain better viewpoint from this interactions, all of the existing studies were reviewed systematically. All published papers have been obtained and reviewed using standard and sensitive keywords from databases such as CINAHL, Embase, PubMed, PsycInfo, and the Cochrane, from 1990 to 2016. The results indicated that all of 6 studies that met the inclusion criteria (from a total of 428 published article) found FTO gene expression changes at short-term follow-ups. Four of six studies found an increased FTO gene expression after calorie restriction, while two of them indicated decreased FTO gene expression. The effect of protein, carbohydrate and fat were separately assessed and suggested by all of six studies. In conclusion, the level of FTO gene expression in hypothalamus is related to macronutrients levels. Future research should evaluate the long-term impact of dietary interventions.

Keywords: obesity, gene expression, FTO, macronutrients

Procedia PDF Downloads 238
28617 Performance Evaluation of Contemporary Classifiers for Automatic Detection of Epileptic EEG

Authors: K. E. Ch. Vidyasagar, M. Moghavvemi, T. S. S. T. Prabhat

Abstract:

Epilepsy is a global problem, and with seizures eluding even the smartest of diagnoses a requirement for automatic detection of the same using electroencephalogram (EEG) would have a huge impact in diagnosis of the disorder. Among a multitude of methods for automatic epilepsy detection, one should find the best method out, based on accuracy, for classification. This paper reasons out, and rationalizes, the best methods for classification. Accuracy is based on the classifier, and thus this paper discusses classifiers like quadratic discriminant analysis (QDA), classification and regression tree (CART), support vector machine (SVM), naive Bayes classifier (NBC), linear discriminant analysis (LDA), K-nearest neighbor (KNN) and artificial neural networks (ANN). Results show that ANN is the most accurate of all the above stated classifiers with 97.7% accuracy, 97.25% specificity and 98.28% sensitivity in its merit. This is followed closely by SVM with 1% variation in result. These results would certainly help researchers choose the best classifier for detection of epilepsy.

Keywords: classification, seizure, KNN, SVM, LDA, ANN, epilepsy

Procedia PDF Downloads 489
28616 Effect of Extraction Methods on the Fatty Acids and Physicochemical Properties of Serendipity Berry Seed Oil

Authors: Olufunmilola A. Abiodun, Adegbola O. Dauda, Ayobami Ojo, Samson A. Oyeyinka

Abstract:

Serendipity berry (Dioscoreophyllum cumminsii diel) is a tropical dioecious rainforest vine and native to tropical Africa. The vine grows during the raining season and is used mainly as sweetener. The sweetener in the berry is known as monellin which is sweeter than sucrose. The sweetener is extracted from the fruits and the seed is discarded. The discarded seeds contain bitter principles but had high yield of oil. Serendipity oil was extracted using three methods (N-hexane, expression and expression/n-hexane). Fatty acids and physicochemical properties of the oil obtained were determined. The oil obtained was clear, liquid and have odour similar to hydrocarbon. The percentage oil yield was 38.59, 12.34 and 49.57% for hexane, expression and expression-hexane method respectively. The seed contained high percentage of oil especially using combination of expression and hexane. Low percentage of oil was obtained using expression method. The refractive index values obtained were 1.443, 1.442 and 1.478 for hexane, expression and expression-hexane methods respectively. Peroxide value obtained for expression-hexane was higher than those for hexane and expression. The viscosities of the oil were 125.8, 128.76 and 126.87 cm³/s for hexane, expression and expression-hexane methods respectively which showed that the oil from expression method was more viscous than the other oils. The major fatty acids in serendipity seed oil were oleic acid (62.81%), linoleic acid (22.65%), linolenic (6.11%), palmitic acid (5.67%), stearic acid (2.21%) in decreasing order. Oleic acid which is monounsaturated fatty acid had the highest value. Total unsaturated fatty acids were 91.574, 92.256 and 90.426% for hexane, expression, and expression-hexane respectively. Combination of expression and hexane for extraction of serendipity oil produced high yield of oil. The oil could be refined for food and non-food application.

Keywords: serendipity seed oil, expression method, fatty acid, hexane

Procedia PDF Downloads 230
28615 Real Time PCR Analysis of microRNA Expression in Oral Cancer

Authors: Karl Kingsley

Abstract:

Many mechanisms are involved in the control of cellular differentiation and growth, which are often dysregulated in many cancers. Many distinct pathways are involved in these mechanisms of control, including deoxyribonuclease (DNA) methyltransferase and histone deacetylase (HDAC) activation that controls both genetic and epigenetic modifications and micro ribonucleic acid (RNA) expression. Less is known about the expression of DNA methyltransferase (DNMT) and HDAC in oral cancers and the effect on microRNA expression. The primary objective of this study was to evaluate the expression of DNMT and HDAC family members in oral cancer and the concomitant expression of cancer-associated microRNAs. Using commercially available oral cancers, including squamous cell carcinoma (SCC)-4, SCC-9, SCC-15, and SCC-25, RNA was extracted and screened for DNMT, HDAC, and microRNA expression using highly-specific primers and quantitative polymerase chain reaction (qPCR). These data revealed low or absent expression of DNMT-1, which is associated with cellular differentiation but increased expression of DNMT-3a and DNMT-3b in all SCC cell lines compared with normal non-cancerous cell controls. In addition, no expression of HDAC1 and HDAC2 expression was found among the normal, non-cancerous cells but was highly expressed in each of the SCC cell lines examined. Differential expression of oncogenic and cancer-associated microRNAs was also observed among the SCC cell lines, including miR-21, miR-133, miR-149, miR-155, miR-365, and miR-720. These findings also appeared to vary according to observed growth rates among these cells. These data may be the first to demonstrate the expression and association between HDAC and DNMT3 family members among oral cancers. In addition, the differential expression of these epigenetic modifiers may be associated with the expression of specific microRNAs in these cancers, which have not previously been observed to the best of the author's knowledge. In addition, some associations and relationships may exist between the expression of these biomarkers and the rates of growth and proliferation, which may suggest that these expression patterns might represent potentially useful biomarkers to determine tumor aggressiveness and other phenotypic behaviors among oral cancers.

Keywords: oral cancer, DNA methyltransferase, histone deacetylase, microRNA

Procedia PDF Downloads 107
28614 Gender Recognition with Deep Belief Networks

Authors: Xiaoqi Jia, Qing Zhu, Hao Zhang, Su Yang

Abstract:

A gender recognition system is able to tell the gender of the given person through a few of frontal facial images. An effective gender recognition approach enables to improve the performance of many other applications, including security monitoring, human-computer interaction, image or video retrieval and so on. In this paper, we present an effective method for gender classification task in frontal facial images based on deep belief networks (DBNs), which can pre-train model and improve accuracy a little bit. Our experiments have shown that the pre-training method with DBNs for gender classification task is feasible and achieves a little improvement of accuracy on FERET and CAS-PEAL-R1 facial datasets.

Keywords: gender recognition, beep belief net-works, semi-supervised learning, greedy-layer wise RBMs

Procedia PDF Downloads 420
28613 Human-Machine Cooperation in Facial Comparison Based on Likelihood Scores

Authors: Lanchi Xie, Zhihui Li, Zhigang Li, Guiqiang Wang, Lei Xu, Yuwen Yan

Abstract:

Image-based facial features can be classified into category recognition features and individual recognition features. Current automated face recognition systems extract a specific feature vector of different dimensions from a facial image according to their pre-trained neural network. However, to improve the efficiency of parameter calculation, an algorithm generally reduces the image details by pooling. The operation will overlook the details concerned much by forensic experts. In our experiment, we adopted a variety of face recognition algorithms based on deep learning, compared a large number of naturally collected face images with the known data of the same person's frontal ID photos. Downscaling and manual handling were performed on the testing images. The results supported that the facial recognition algorithms based on deep learning detected structural and morphological information and rarely focused on specific markers such as stains and moles. Overall performance, distribution of genuine scores and impostor scores, and likelihood ratios were tested to evaluate the accuracy of biometric systems and forensic experts. Experiments showed that the biometric systems were skilled in distinguishing category features, and forensic experts were better at discovering the individual features of human faces. In the proposed approach, a fusion was performed at the score level. At the specified false accept rate, the framework achieved a lower false reject rate. This paper contributes to improving the interpretability of the objective method of facial comparison and provides a novel method for human-machine collaboration in this field.

Keywords: likelihood ratio, automated facial recognition, facial comparison, biometrics

Procedia PDF Downloads 99
28612 The Effect of Experimentally Induced Stress on Facial Recognition Ability of Security Personnel’s

Authors: Zunjarrao Kadam, Vikas Minchekar

Abstract:

The facial recognition is an important task in criminal investigation procedure. The security guards-constantly watching the persons-can help to identify the suspected accused. The forensic psychologists are tackled such cases in the criminal justice system. The security personnel may loss their ability to correctly identify the persons due to constant stress while performing the duty. The present study aimed at to identify the effect of experimentally induced stress on facial recognition ability of security personnel’s. For this study 50, security guards from Sangli, Miraj & Jaysingpur city of the Maharashtra States of India were recruited in the experimental study. The randomized two group design was employed to carry out the research. In the initial condition twenty identity card size photographs were shown to both groups. Afterward, artificial stress was induced in the experimental group through the difficultpuzzle-solvingtask in a limited period. In the second condition, both groups were presented earlier photographs with another additional thirty new photographs. The subjects were asked to recognize the photographs which are shown earliest. The analyzed data revealed that control group has ahighest mean score of facial recognition than experimental group. The results were discussed in the present research.

Keywords: experimentally induced stress, facial recognition, cognition, security personnel

Procedia PDF Downloads 232
28611 Tensor Deep Stacking Neural Networks and Bilinear Mapping Based Speech Emotion Classification Using Facial Electromyography

Authors: P. S. Jagadeesh Kumar, Yang Yung, Wenli Hu

Abstract:

Speech emotion classification is a dominant research field in finding a sturdy and profligate classifier appropriate for different real-life applications. This effort accentuates on classifying different emotions from speech signal quarried from the features related to pitch, formants, energy contours, jitter, shimmer, spectral, perceptual and temporal features. Tensor deep stacking neural networks were supported to examine the factors that influence the classification success rate. Facial electromyography signals were composed of several forms of focuses in a controlled atmosphere by means of audio-visual stimuli. Proficient facial electromyography signals were pre-processed using moving average filter, and a set of arithmetical features were excavated. Extracted features were mapped into consistent emotions using bilinear mapping. With facial electromyography signals, a database comprising diverse emotions will be exposed with a suitable fine-tuning of features and training data. A success rate of 92% can be attained deprived of increasing the system connivance and the computation time for sorting diverse emotional states.

Keywords: speech emotion classification, tensor deep stacking neural networks, facial electromyography, bilinear mapping, audio-visual stimuli

Procedia PDF Downloads 221
28610 Semi-Automated Tracking of Vibrissal Movements in Free-Moving Rodents Captured by High-Speed Videos

Authors: Hyun June Kim, Tailong Shi, Seden Akdagli, Sam Most, Yuling Yan

Abstract:

Quantitative analysis of mouse whisker movement can be used to study functional recovery and regeneration of facial nerve after an injury. However, it is challenging to accurately track mouse whisker movements, and most whisker tracking methods require manual intervention, e.g. fixing the head of the mouse during a study. Here we describe a semi-automated image processing method that is applied to high-speed video recordings of free-moving mice to track whisker movements. We first track the head movement of a mouse by delineating the lower head contour frame-by-frame to locate and determine the orientation of its head. Then, a region of interest is identified for each frame, with subsequent application of the Hough transform to track individual whisker movements on each side of the head. Our approach is used to examine the functional recovery of damaged facial nerves in mice over a course of 21 days.

Keywords: mystacial macrovibrissae, whisker tracking, head tracking, facial nerve recovery

Procedia PDF Downloads 560
28609 High-Resolution Facial Electromyography in Freely Behaving Humans

Authors: Lilah Inzelberg, David Rand, Stanislav Steinberg, Moshe David Pur, Yael Hanein

Abstract:

Human facial expressions carry important psychological and neurological information. Facial expressions involve the co-activation of diverse muscles. They depend strongly on personal affective interpretation and on social context and vary between spontaneous and voluntary activations. Smiling, as a special case, is among the most complex facial emotional expressions, involving no fewer than 7 different unilateral muscles. Despite their ubiquitous nature, smiles remain an elusive and debated topic. Smiles are associated with happiness and greeting on one hand and anger or disgust-masking on the other. Accordingly, while high-resolution recording of muscle activation patterns, in a non-interfering setting, offers exciting opportunities, it remains an unmet challenge, as contemporary surface facial electromyography (EMG) methodologies are cumbersome, restricted to the laboratory settings, and are limited in time and resolution. Here we present a wearable and non-invasive method for objective mapping of facial muscle activation and demonstrate its application in a natural setting. The technology is based on a recently developed dry and soft electrode array, specially designed for surface facial EMG technique. Eighteen healthy volunteers (31.58 ± 3.41 years, 13 females), participated in the study. Surface EMG arrays were adhered to participant left and right cheeks. Participants were instructed to imitate three facial expressions: closing the eyes, wrinkling the nose and smiling voluntary and to watch a funny video while their EMG signal is recorded. We focused on muscles associated with 'enjoyment', 'social' and 'masked' smiles; three categories with distinct social meanings. We developed a customized independent component analysis algorithm to construct the desired facial musculature mapping. First, identification of the Orbicularis oculi and the Levator labii superioris muscles was demonstrated from voluntary expressions. Second, recordings of voluntary and spontaneous smiles were used to locate the Zygomaticus major muscle activated in Duchenne and non-Duchenne smiles. Finally, recording with a wireless device in an unmodified natural work setting revealed expressions of neutral, positive and negative emotions in face-to-face interaction. The algorithm outlined here identifies the activation sources in a subject-specific manner, insensitive to electrode placement and anatomical diversity. Our high-resolution and cross-talk free mapping performances, along with excellent user convenience, open new opportunities for affective processing and objective evaluation of facial expressivity, objective psychological and neurological assessment as well as gaming, virtual reality, bio-feedback and brain-machine interface applications.

Keywords: affective expressions, affective processing, facial EMG, high-resolution electromyography, independent component analysis, wireless electrodes

Procedia PDF Downloads 218
28608 A Novel PfkB Gene Cloning and Characterization for Expression in Potato Plants

Authors: Arfan Ali, Idrees Ahmad Nasir

Abstract:

Potato (Solanum tuberosum) is an important cash crop and popular vegetable in Pakistan and throughout the world. Cold storage of potatoes accelerates the conversion of starch into reduced sugars (glucose and fructose). This process causes dry mass and bitter taste in the potatoes that are not acceptable to end consumers. In the current study, the phosphofructokinase B gene was cloned into the pET-30 vector for protein expression and the pCambia-1301 vector for plant expression. Amplification of a 930bp product from an E. coli strain determined the successful isolation of the phosphofructokinase B gene. Restriction digestion using NcoI and BglII along with the amplification of the 930bp product using gene specific primers confirmed the successful cloning of the PfkB gene in both vectors. The protein was expressed as a His-PfkB fusion protein. Western blot analysis confirmed the presence of the 35 Kda PfkB protein when hybridized with anti-His antibodies. The construct Fani-01 was evaluated transiently using a histochemical gus assay. The appearance of blue color in the agroinfiltrated area of potato leaves confirmed the successful expression of construct Fani-01. Further, the area displaying gus expression was evaluated for PfkB expression using ELISA. Moreover, PfkB gene expression evaluated through transient expression determined successful gene expression and highlighted its potential utilization for stable expression in potato to reduce sweetening due to long-term storage.

Keywords: potato, Solanum tuberosum, transformation, PfkB, anti-sweetening

Procedia PDF Downloads 437
28607 Vicarious Cues in Portraying Emotion: Musicians' Self-Appraisal

Authors: W. Linthicum-Blackhorse, P. Martens

Abstract:

This present study seeks to discover attitudinal commonalities and differences within a musician population relative to the communication of emotion via music. We hypothesized that instrument type, as well as age and gender, would bear significantly on musicians’ opinions. A survey was administered to 178 participants; 152 were current music majors (mean age 20.3 years, 62 female) and 26 were adult participants in a community choir (mean age 54.0 years, 12 female). The adult participants were all vocalists, while student participants represented the full range of orchestral instruments. The students were grouped by degree program, (performance, music education, or other) and instrument type (voice, brass, woodwinds, strings, percussion). The survey asked 'How important are each of the following areas to you for portraying emotion in music?' Participants were asked to rate each of 15 items on a scale of 1 (not at all important) to 10 (very important). Participants were also instructed to leave blank any item that they did not understand. The 15 items were: dynamic contrast, overall volume, phrasing, facial expression, staging (placement), pitch accuracy, tempo changes, bodily movement, your mood, your attitude, vibrato, rubato, stage/room lighting, clothing type, and clothing color. Contrary to our hypothesis, there was no overall effect of gender or age, and neither did any single response item show a significant difference due to these subject parameters. Among the student participants, however, one-way ANOVA revealed a significant effect of degree program on the rated importance of four items: dynamic contrast, tempo changes, vibrato, and rubato. Significant effects of instrument type were found in the responses to eight items: facial expression, staging, body movement, vibrato, rubato, lighting, clothing type, and clothing color. Post hoc comparisons (Tukey) show that some variation follows from obvious differences between instrument types (e.g. string players are more concerned with vibrato than everyone but woodwind players; vocalists are significantly more concerned with facial expression than everyone but string players), but other differences could point to communal mindsets toward vicarious cues within instrument type. These mindsets could be global (e.g. brass players deeming body movement significantly less important than string players, being less often featured as soloists and appearing less often at the front of the stage) or local (e.g. string players being significantly more concerned than all other groups about both clothing color and type, perhaps due to the strongly-expressed opinions of specific teachers). Future work will attempt to identify the source of these self-appraisals, whether enculturated via explicit pedagogy, or whether absorbed from individuals' observations and performance experience.

Keywords: performance, vicarious cues, communication, emotion

Procedia PDF Downloads 81
28606 A Neuron Model of Facial Recognition and Detection of an Authorized Entity Using Machine Learning System

Authors: J. K. Adedeji, M. O. Oyekanmi

Abstract:

This paper has critically examined the use of Machine Learning procedures in curbing unauthorized access into valuable areas of an organization. The use of passwords, pin codes, user’s identification in recent times has been partially successful in curbing crimes involving identities, hence the need for the design of a system which incorporates biometric characteristics such as DNA and pattern recognition of variations in facial expressions. The facial model used is the OpenCV library which is based on the use of certain physiological features, the Raspberry Pi 3 module is used to compile the OpenCV library, which extracts and stores the detected faces into the datasets directory through the use of camera. The model is trained with 50 epoch run in the database and recognized by the Local Binary Pattern Histogram (LBPH) recognizer contained in the OpenCV. The training algorithm used by the neural network is back propagation coded using python algorithmic language with 200 epoch runs to identify specific resemblance in the exclusive OR (XOR) output neurons. The research however confirmed that physiological parameters are better effective measures to curb crimes relating to identities.

Keywords: biometric characters, facial recognition, neural network, OpenCV

Procedia PDF Downloads 228
28605 Systems Approach on Thermal Analysis of an Automatic Transmission

Authors: Sinsze Koo, Benjin Luo, Matthew Henry

Abstract:

In order to increase the performance of an automatic transmission, the automatic transmission fluid is required to be warm up to an optimal operating temperature. In a conventional vehicle, cold starts result in friction loss occurring in the gear box and engine. The stop and go nature of city driving dramatically affect the warm-up of engine oil and automatic transmission fluid and delay the time frame needed to reach an optimal operating temperature. This temperature phenomenon impacts both engine and transmission performance but also increases fuel consumption and CO2 emission. The aim of this study is to develop know-how of the thermal behavior in order to identify thermal impacts and functional principles in automatic transmissions. Thermal behavior was studied using models and simulations, developed using GT-Suit, on a one-dimensional thermal and flow transport. A power train of a conventional vehicle was modeled in order to emphasis the thermal phenomena occurring in the various components and how they impact the automatic transmission performance. The simulation demonstrates the thermal model of a transmission fluid cooling system and its component parts in warm-up after a cold start. The result of these analyses will support the future designs of transmission systems and components in an attempt to obtain better fuel efficiency and transmission performance. Therefore, these thermal analyses could possibly identify ways that improve existing thermal management techniques with prioritization on fuel efficiency.

Keywords: thermal management, automatic transmission, hybrid, and systematic approach

Procedia PDF Downloads 351
28604 Optimized Deep Learning-Based Facial Emotion Recognition System

Authors: Erick C. Valverde, Wansu Lim

Abstract:

Facial emotion recognition (FER) system has been recently developed for more advanced computer vision applications. The ability to identify human emotions would enable smart healthcare facility to diagnose mental health illnesses (e.g., depression and stress) as well as better human social interactions with smart technologies. The FER system involves two steps: 1) face detection task and 2) facial emotion recognition task. It classifies the human expression in various categories such as angry, disgust, fear, happy, sad, surprise, and neutral. This system requires intensive research to address issues with human diversity, various unique human expressions, and variety of human facial features due to age differences. These issues generally affect the ability of the FER system to detect human emotions with high accuracy. Early stage of FER systems used simple supervised classification task algorithms like K-nearest neighbors (KNN) and artificial neural networks (ANN). These conventional FER systems have issues with low accuracy due to its inefficiency to extract significant features of several human emotions. To increase the accuracy of FER systems, deep learning (DL)-based methods, like convolutional neural networks (CNN), are proposed. These methods can find more complex features in the human face by means of the deeper connections within its architectures. However, the inference speed and computational costs of a DL-based FER system is often disregarded in exchange for higher accuracy results. To cope with this drawback, an optimized DL-based FER system is proposed in this study.An extreme version of Inception V3, known as Xception model, is leveraged by applying different network optimization methods. Specifically, network pruning and quantization are used to enable lower computational costs and reduce memory usage, respectively. To support low resource requirements, a 68-landmark face detector from Dlib is used in the early step of the FER system.Furthermore, a DL compiler is utilized to incorporate advanced optimization techniques to the Xception model to improve the inference speed of the FER system. In comparison to VGG-Net and ResNet50, the proposed optimized DL-based FER system experimentally demonstrates the objectives of the network optimization methods used. As a result, the proposed approach can be used to create an efficient and real-time FER system.

Keywords: deep learning, face detection, facial emotion recognition, network optimization methods

Procedia PDF Downloads 78
28603 Facial Recognition on the Basis of Facial Fragments

Authors: Tetyana Baydyk, Ernst Kussul, Sandra Bonilla Meza

Abstract:

There are many articles that attempt to establish the role of different facial fragments in face recognition. Various approaches are used to estimate this role. Frequently, authors calculate the entropy corresponding to the fragment. This approach can only give approximate estimation. In this paper, we propose to use a more direct measure of the importance of different fragments for face recognition. We propose to select a recognition method and a face database and experimentally investigate the recognition rate using different fragments of faces. We present two such experiments in the paper. We selected the PCNC neural classifier as a method for face recognition and parts of the LFW (Labeled Faces in the Wild) face database as training and testing sets. The recognition rate of the best experiment is comparable with the recognition rate obtained using the whole face.

Keywords: face recognition, labeled faces in the wild (LFW) database, random local descriptor (RLD), random features

Procedia PDF Downloads 329