Search results for: voice response recognition
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7059

Search results for: voice response recognition

7029 Voice Signal Processing and Coding in MATLAB Generating a Plasma Signal in a Tesla Coil for a Security System

Authors: Juan Jimenez, Erika Yambay, Dayana Pilco, Brayan Parra

Abstract:

This paper presents an investigation of voice signal processing and coding using MATLAB, with the objective of generating a plasma signal on a Tesla coil within a security system. The approach focuses on using advanced voice signal processing techniques to encode and modulate the audio signal, which is then amplified and applied to a Tesla coil. The result is the creation of a striking visual effect of voice-controlled plasma with specific applications in security systems. The article explores the technical aspects of voice signal processing, the generation of the plasma signal, and its relationship to security. The implications and creative potential of this technology are discussed, highlighting its relevance at the forefront of research in signal processing and visual effect generation in the field of security systems.

Keywords: voice signal processing, voice signal coding, MATLAB, plasma signal, Tesla coil, security system, visual effects, audiovisual interaction

Procedia PDF Downloads 60
7028 Phone Number Spoofing Attack in VoLTE

Authors: Joo-Hyung Oh, Sekwon Kim, Myoungsun Noh, Chaetae Im

Abstract:

The number of service users of 4G VoLTE (voice over LTE) using LTE data networks is rapidly growing. VoLTE based on All-IP network enables clearer and higher-quality voice calls than 3G. It does, however, pose new challenges; a voice call through IP networks makes it vulnerable to security threats such as wiretapping and forged or falsified information. Moreover, in particular, stealing other users’ phone numbers and forging or falsifying call request messages from outgoing voice calls within VoLTE result in considerable losses that include user billing and voice phishing to acquaintances. This paper focuses on the threats of caller phone number spoofing in the VoLTE and countermeasure technology as safety measures for mobile communication networks.

Keywords: LTE, 4G, VoLTE, phone number spoofing

Procedia PDF Downloads 494
7027 Interactive Shadow Play Animation System

Authors: Bo Wan, Xiu Wen, Lingling An, Xiaoling Ding

Abstract:

The paper describes a Chinese shadow play animation system based on Kinect. Users, without any professional training, can personally manipulate the shadow characters to finish a shadow play performance by their body actions and get a shadow play video through giving the record command to our system if they want. In our system, Kinect is responsible for capturing human movement and voice commands data. Gesture recognition module is used to control the change of the shadow play scenes. After packaging the data from Kinect and the recognition result from gesture recognition module, VRPN transmits them to the server-side. At last, the server-side uses the information to control the motion of shadow characters and video recording. This system not only achieves human-computer interaction, but also realizes the interaction between people. It brings an entertaining experience to users and easy to operate for all ages. Even more important is that the application background of Chinese shadow play embodies the protection of the art of shadow play animation.

Keywords: hadow play animation, Kinect, gesture recognition, VRPN, HCI

Procedia PDF Downloads 378
7026 Using Augmented Reality to Enhance Doctor Patient Communication

Authors: Rutusha Bhutada, Gaurav Chavan, Sarvesh Kasat, Varsha Mujumdar

Abstract:

This software system will be an Augmented Reality application designed to maximize the doctor’s productivity by providing tools to assist in automating the patient recognition and updating patient’s records using face and voice recognition features, which would otherwise have to be performed manually. By maximizing the doctor’s work efficiency and production, the application will meet the doctor’s needs while remaining easy to understand and use. More specifically, this application is designed to allow a doctor to manage his productive time in handling the patient without losing eye-contact with him and communicate with a group of other doctors for consultation, for in-place treatments through video streaming, as a video study. The system also contains a relational database containing a list of doctor, patient and display techniques.

Keywords: augmented reality, hand-held devices, head-mounted devices, marker based systems, speech recognition, face detection

Procedia PDF Downloads 412
7025 Handwriting Recognition of Gurmukhi Script: A Survey of Online and Offline Techniques

Authors: Ravneet Kaur

Abstract:

Character recognition is a very interesting area of pattern recognition. From past few decades, an intensive research on character recognition for Roman, Chinese, and Japanese and Indian scripts have been reported. In this paper, a review of Handwritten Character Recognition work on Indian Script Gurmukhi is being highlighted. Most of the published papers were summarized, various methodologies were analysed and their results are reported.

Keywords: Gurmukhi character recognition, online, offline, HCR survey

Procedia PDF Downloads 404
7024 OCR/ICR Text Recognition Using ABBYY FineReader as an Example Text

Authors: A. R. Bagirzade, A. Sh. Najafova, S. M. Yessirkepova, E. S. Albert

Abstract:

This article describes a text recognition method based on Optical Character Recognition (OCR). The features of the OCR method were examined using the ABBYY FineReader program. It describes automatic text recognition in images. OCR is necessary because optical input devices can only transmit raster graphics as a result. Text recognition describes the task of recognizing letters shown as such, to identify and assign them an assigned numerical value in accordance with the usual text encoding (ASCII, Unicode). The peculiarity of this study conducted by the authors using the example of the ABBYY FineReader, was confirmed and shown in practice, the improvement of digital text recognition platforms developed by Electronic Publication.

Keywords: ABBYY FineReader system, algorithm symbol recognition, OCR/ICR techniques, recognition technologies

Procedia PDF Downloads 144
7023 NLRP3-Inflammassome Participates in the Inflammatory Response Induced by Paracoccidioides brasiliensis

Authors: Eduardo Kanagushiku Pereira, Frank Gregory Cavalcante da Silva, Barbara Soares Gonçalves, Ana Lúcia Bergamasco Galastri, Ronei Luciano Mamoni

Abstract:

The inflammatory response initiates after the recognition of pathogens by receptors expressed by innate immune cells. Among these receptors, the NLRP3 was associated with the recognition of pathogenic fungi in experimental models. NLRP3 operates forming a multiproteic complex called inflammasome, which actives caspase-1, responsible for the production of the inflammatory cytokines IL-1beta and IL-18. In this study, we aimed to investigate the involvement of NLRP3 in the inflammatory response elicited in macrophages against Paracoccidioides brasiliensis (Pb), the etiologic agent of PCM. Macrophages were differentiated from THP-1 cells by treatment with phorbol-myristate-acetate. Following differentiation, macrophages were stimulated by Pb yeast cells for 24 hours, after previous treatment with specific NLRP3 (3,4-methylenedioxy-beta-nitrostyrene) and/or caspase-1 (VX-765) inhibitors, or specific inhibitors of pathways involved in NLRP3 activation such as: Reactive Oxigen Species (ROS) production (N-Acetyl-L-cysteine), K+ efflux (Glibenclamide) or phagossome acidification (Bafilomycin). Quantification of IL-1beta and IL-18 in supernatants was performed by ELISA. Our results showed that the production of IL-1beta and IL-18 by THP-1-derived-macrophages stimulated with Pb yeast cells was dependent on NLRP3 and caspase-1 activation, once the presence of their specific inhibitors diminished the production of these cytokines. Furthermore, we found that the major pathways involved in NLRP3 activation, after Pb recognition, were dependent on ROS production and K+ efflux. In conclusion, our results showed that NLRP3 participates in the recognition of Pb yeast cells by macrophages, leading to the activation of the NLRP3-inflammasome and production of IL-1beta and IL-18. Together, these cytokines can induce an inflammatory response against P. brasiliensis, essential for the establishment of the initial inflammatory response and for the development of the subsequent acquired immune response.

Keywords: inflammation, IL-1beta, IL-18, NLRP3, Paracoccidioidomycosis

Procedia PDF Downloads 249
7022 Recognition by the Voice and Speech Features of the Emotional State of Children by Adults and Automatically

Authors: Elena E. Lyakso, Olga V. Frolova, Yuri N. Matveev, Aleksey S. Grigorev, Alexander S. Nikolaev, Viktor A. Gorodnyi

Abstract:

The study of the children’s emotional sphere depending on age and psychoneurological state is of great importance for the design of educational programs for children and their social adaptation. Atypical development may be accompanied by violations or specificities of the emotional sphere. To study characteristics of the emotional state reflection in the voice and speech features of children, the perceptual study with the participation of adults and the automatic recognition of speech were conducted. Speech of children with typical development (TD), with Down syndrome (DS), and with autism spectrum disorders (ASD) aged 6-12 years was recorded. To obtain emotional speech in children, model situations were created, including a dialogue between the child and the experimenter containing questions that can cause various emotional states in the child and playing with a standard set of toys. The questions and toys were selected, taking into account the child’s age, developmental characteristics, and speech skills. For the perceptual experiment by adults, test sequences containing speech material of 30 children: TD, DS, and ASD were created. The listeners were 100 adults (age 19.3 ± 2.3 years). The listeners were tasked with determining the children’s emotional state as “comfort – neutral – discomfort” while listening to the test material. Spectrographic analysis of speech signals was conducted. For automatic recognition of the emotional state, 6594 speech files containing speech material of children were prepared. Automatic recognition of three states, “comfort – neutral – discomfort,” was performed using automatically extracted from the set of acoustic features - the Geneva Minimalistic Acoustic Parameter Set (GeMAPS) and the extended Geneva Minimalistic Acoustic Parameter Set (eGeMAPS). The results showed that the emotional state is worse determined by the speech of TD children (comfort – 58% of correct answers, discomfort – 56%). Listeners better recognized discomfort in children with ASD and DS (78% of answers) than comfort (70% and 67%, respectively, for children with DS and ASD). The neutral state is better recognized by the speech of children with ASD (67%) than by the speech of children with DS (52%) and TD children (54%). According to the automatic recognition data using the acoustic feature set GeMAPSv01b, the accuracy of automatic recognition of emotional states for children with ASD is 0.687; children with DS – 0.725; TD children – 0.641. When using the acoustic feature set eGeMAPSv01b, the accuracy of automatic recognition of emotional states for children with ASD is 0.671; children with DS – 0.717; TD children – 0.631. The use of different models showed similar results, with better recognition of emotional states by the speech of children with DS than by the speech of children with ASD. The state of comfort is automatically determined better by the speech of TD children (precision – 0.546) and children with ASD (0.523), discomfort – children with DS (0.504). The data on the specificities of recognition by adults of the children’s emotional state by their speech may be used in recruitment for working with children with atypical development. Automatic recognition data can be used to create alternative communication systems and automatic human-computer interfaces for social-emotional learning. Acknowledgment: This work was financially supported by the Russian Science Foundation (project 18-18-00063).

Keywords: autism spectrum disorders, automatic recognition of speech, child’s emotional speech, Down syndrome, perceptual experiment

Procedia PDF Downloads 166
7021 Interaction between Breathiness and Nasality: An Acoustic Analysis

Authors: Pamir Gogoi, Ratree Wayland

Abstract:

This study investigates the acoustic measures of breathiness when coarticulated with nasality. The acoustic correlates of breathiness and nasality that has already been well established after years of empirical research. Some of these acoustic parameters - like low frequency peaks and wider bandwidths- are common for both nasal and breathy voice. Therefore, it is likely that these parameters interact when a sound is coarticulated with breathiness and nasality. This leads to the hypothesis that the acoustic parameters, which usually act as robust cues in differentiating between breathy and modal voice, might not be reliable cues for differentiating between breathy and modal voice when breathiness is coarticulated with nasality. The effect of nasality on the perception of breathiness has been explored in earlier studies using synthesized speech. The results showed that perceptually, nasality and breathiness do interact. The current study investigates if a similar pattern is observed in natural speech. The study is conducted on Marathi, an Indo-Aryan language which has a three-way contrast between nasality and breathiness. That is, there is a phonemic distinction between nasals, breathy voice and breathy-nasals. Voice quality parameters like – H1-H2 (Difference between the amplitude of first and second harmonic), H1-A3 (Difference between the amplitude of first harmonic and third formant, CPP (Cepstral Peak Prominence), HNR (Harmonics to Noise ratio) and B1 (Bandwidth of first formant) were extracted. Statistical models like linear mixed effects regression and Random Forest classifiers show that measures that capture the noise component in the signal- like CPP and HNR- can classify breathy voice from modal voice better than spectral measures when breathy voice is coarticulated with nasality.

Keywords: breathiness, marathi, nasality, voice quality

Procedia PDF Downloads 68
7020 An Improved OCR Algorithm on Appearance Recognition of Electronic Components Based on Self-adaptation of Multifont Template

Authors: Zhu-Qing Jia, Tao Lin, Tong Zhou

Abstract:

The recognition method of Optical Character Recognition has been expensively utilized, while it is rare to be employed specifically in recognition of electronic components. This paper suggests a high-effective algorithm on appearance identification of integrated circuit components based on the existing methods of character recognition, and analyze the pros and cons.

Keywords: optical character recognition, fuzzy page identification, mutual correlation matrix, confidence self-adaptation

Procedia PDF Downloads 514
7019 Electroencephalography-Based Intention Recognition and Consensus Assessment during Emergency Response

Authors: Siyao Zhu, Yifang Xu

Abstract:

After natural and man-made disasters, robots can bypass the danger, expedite the search, and acquire unprecedented situational awareness to design rescue plans. The hands-free requirement from the first responders excludes the use of tedious manual control and operation. In unknown, unstructured, and obstructed environments, natural-language-based supervision is not amenable for first responders to formulate, and is difficult for robots to understand. Brain-computer interface is a promising option to overcome the limitations. This study aims to test the feasibility of using electroencephalography (EEG) signals to decode human intentions and detect the level of consensus on robot-provided information. EEG signals were classified using machine-learning and deep-learning methods to discriminate search intentions and agreement perceptions. The results show that the average classification accuracy for intention recognition and consensus assessment is 67% and 72%, respectively, proving the potential of incorporating recognizable users’ bioelectrical responses into advanced robot-assisted systems for emergency response.

Keywords: consensus assessment, electroencephalogram, emergency response, human-robot collaboration, intention recognition, search and rescue

Procedia PDF Downloads 66
7018 Assessing the Preparedness of Teachers for Their Role in an Inclusive Classroom: Photo-Voice as a Reflexive Tool

Authors: Nan Stevens

Abstract:

Photo-voice is a participatory method through which participants identify and represent their lived experiences and contexts through the use of photo imagery. Photo-voice is a qualitative research method that explores individuals’ lived experiences. This method is known as a creative art form to help researchers listen to the 'voice' of a certain population. A teacher educator at Thompson Rivers University, responsible for preparing new teachers for the demands of the profession in an ever-changing demographic, utilized the Photo-voice method to enable a self-study of emerging teachers’ readiness for the inclusive classroom. Coding analysis was applied to 96 Photo-voice portfolios, which were created over two years with the Inclusive Education course work, in a Bachelor of Education program (Elementary). Coding utilized students’ written associations to their visual images, anecdotes attached to visual metaphors, and personal narratives that illustrated the professional development process in which they were engaged. Thematic findings include: 1) becoming an inclusive educator is a process; 2) one must be open to identifying and exploring their fear and biases, and 3) an attitudinal shift enables relevant skill acquisition and readiness for working with diverse student needs.

Keywords: teacher education, inclusive education, professional development, Photo-voice

Procedia PDF Downloads 116
7017 Voice Quality in Italian-Speaking Children with Autism

Authors: Patrizia Bonaventura, Magda Di Renzo

Abstract:

This project aims to measure and assess the voice quality in children with autism. Few previous studies exist which have analyzed the voice quality of individuals with autism: abnormal voice characteristics have been found, like a high pitch, great pitch range, and sing-song quality. Existing studies did not focus specifically on Italian-speaking children’s voices and provided analysis of a few acoustic parameters. The present study aimed to gather more data and to perform acoustic analysis of the voice of children with autism in order to identify patterns of abnormal voice features that might shed some light on the causes of the dysphonia and possibly be used to create a pediatric assessment tool for early identification of autism. The participants were five native Italian-speaking boys with autism between the age of 4 years and 10 years (mean 6.8 ± SD 1.4). The children had a diagnosis of autism, were verbal, and had no other comorbid conditions (like Down syndrome or ADHD). The voices of the autistic children were recorded in the production of sustained vowels [ah] and [ih] and of sentences from the Italian version of the CAPE-V voice assessment test. The following voice parameters, representative of normal quality, were analyzed by acoustic spectrography through Praat: Speaking Fundamental Frequency, F0 range, average intensity, and dynamic range. The results showed that the pitch parameters (Speaking Fundamental Frequency and F0 range), as well as the intensity parameters (average intensity and dynamic range), were significantly different from the relative normal reference thresholds. Also, variability among children was found, so confirming a tendency revealed in previous studies of individual variation in these aspects of voice quality. The results indicate a general pattern of abnormal voice quality characterized by a high pitch and large variations in pitch and intensity. These acoustic voice characteristics found in Italian-speaking autistic children match those found in children speaking other languages, indicating that autism symptoms affecting voice quality might be independent of the native language of the children.

Keywords: autism, voice disorders, speech science, acoustic analysis of voice

Procedia PDF Downloads 46
7016 Design and Development of Automatic Onion Harvester

Authors: P. Revathi, T. Mrunalini, K. Padma Priya, P. Ramya, R. Saranya

Abstract:

During the tough times of covid, those people who were hospitalized found it difficult to always convey what they wanted to or needed to the attendee. Sometimes the attendees might also not be there. In that case, the patients can use simple hand gestures to control electrical appliances (like its set it for a zero watts bulb)and three other gestures for voice note intimation. In this AI-based hand recognition project, NodeMCU is used for the control action of the relay, and it is connected to the firebase for storing the value in the cloud and is interfaced with the python code via raspberry pi. For three hand gestures, a voice clip is added for intimation to the attendee. This is done with the help of Google’s text to speech and the inbuilt audio file option in the raspberry pi 4. All the 5 gestures will be detected when shown with their hands via a webcam which is placed for gesture detection. A personal computer is used for displaying the gestures and for running the code in the raspberry pi imager.

Keywords: onion harvesting, automatic pluging, camera, raspberry pi

Procedia PDF Downloads 172
7015 Facial Recognition on the Basis of Facial Fragments

Authors: Tetyana Baydyk, Ernst Kussul, Sandra Bonilla Meza

Abstract:

There are many articles that attempt to establish the role of different facial fragments in face recognition. Various approaches are used to estimate this role. Frequently, authors calculate the entropy corresponding to the fragment. This approach can only give approximate estimation. In this paper, we propose to use a more direct measure of the importance of different fragments for face recognition. We propose to select a recognition method and a face database and experimentally investigate the recognition rate using different fragments of faces. We present two such experiments in the paper. We selected the PCNC neural classifier as a method for face recognition and parts of the LFW (Labeled Faces in the Wild) face database as training and testing sets. The recognition rate of the best experiment is comparable with the recognition rate obtained using the whole face.

Keywords: face recognition, labeled faces in the wild (LFW) database, random local descriptor (RLD), random features

Procedia PDF Downloads 333
7014 SLIITBOT: Design of a Socially Assistive Robot for SLIIT

Authors: Chandimal Jayawardena, Ridmal Mendis, Manoji Tennakoon, Theekshana Wijayathilaka, Randima Marasinghe

Abstract:

This research paper defines the research area of the implementation of the socially assistive robot (SLIITBOT). It consists of the overall process implemented within the robot’s system and limitations, along with a literature survey. This project considers developing a socially assistive robot called SLIITBOT that will interact using its voice outputs and graphical user interface with people within the university and benefit them with updates and tasks. The robot will be able to detect a person when he/she enters the room, navigate towards the position the human is standing, welcome and greet the particular person with a simple conversation using its voice, introduce the services through its voice, and provide the person with services through an electronic input via an app while guiding the person with voice outputs.

Keywords: application, detection, dialogue, navigation

Procedia PDF Downloads 147
7013 Prophylactic Replacement of Voice Prosthesis: A Study to Predict Prosthesis Lifetime

Authors: Anne Heirman, Vincent van der Noort, Rob van Son, Marije Petersen, Lisette van der Molen, Gyorgy Halmos, Richard Dirven, Michiel van den Brekel

Abstract:

Objective: Voice prosthesis leakage significantly impacts laryngectomies patients' quality of life, causing insecurity and frequent unplanned hospital visits and costs. In this study, the concept of prophylactic voice prosthesis replacement was explored to prevent leakages. Study Design: A retrospective cohort study. Setting: Tertiary hospital. Methods: Device lifetimes and voice prosthesis replacements of a retrospective cohort, including all patients with laryngectomies between 2000 and 2012 in the Netherlands Cancer Institute, were used to calculate the number of needed voice prostheses per patient per year when preventing 70% of the leakages by prophylactic replacement. Various strategies for the timing of prophylactic replacement were considered: Adaptive strategies based on the individual patient’s history of replacement and fixed strategies based on the results of patients with similar voice prosthesis or treatment characteristics. Results: Patients used a median of 3.4 voice prostheses per year (range 0.1-48.1). We found a high inter-and intrapatient variability in device lifetime. When applying prophylactic replacement, this would become a median of 9.4 voice prostheses per year, which means replacement every 38 days, implying more than six additional voice prostheses per patient per year. The individual adaptive model showed that preventing 70% of the leakages was impossible for most patients, and only a median of 25% can be prevented. Monte-Carlo simulations showed that prophylactic replacement is not feasible due to the high Coefficient of Variation (Standard Deviation/Mean) in device lifetime. Conclusion: Based on our simulations, prophylactic replacement of voice prostheses is not feasible due to high inter-and intrapatient variation in device lifetime.

Keywords: voice prosthesis, voice rehabilitation, total laryngectomy, prosthetic leakage, device lifetime

Procedia PDF Downloads 109
7012 Effect of Lullabies on Babies Growth and Development, Vital Signs and Hospitalization Times in the Neonatal Intensive Care Units

Authors: Işın Alkan, Meltem Kürtüncü

Abstract:

Objective: This study was carried out with an experimental design in order to determine whether the lullaby, which was listened from mother’s voice and a stranger’s voice to the babies born at term and hospitalized in neonatal intensive care unit, had an effect on saturation values (SpO2), peak heart rate (PHR), respiration, fever, growth and development and hospitalization times of the infants. Method: Data from the study were obtained from 90 newborn babies who were hospitalized in Neonatal Intensive Care Unit of Zonguldak Maternity And Children Hospital between September 2015-January 2016 and who met the eligibility criteria. Lullaby concert was performed by choosing one of the suitable care hours. SpO2, PHR, respiration, fever, growth and development and hospitalization times of the infants were recorded by the researcher on “Newborn response follow-up form” at pre-care and post-care. Vital signs of babies every day, weight, height and head circumference measurements at admission, weakly rated at an output. Results: In the experimental and control groups, like weight, height and head circumference anthropometric measurements were not found statistically significant difference intensive care units admission and output times. Hospitalization times on babies who listen to lullaby mother’s voice revealed statistically significant difference according to babies who listen to lullaby stranger’s voice. Before care and after care were examined, SpO2 rates of babies who listen to lullaby mother’s voice revealed statistically significant higher difference according to babies who listen to lullaby stranger’s voice and control group babies. Before care on PHR of babies in three groups were not found the statistical difference, but aftercare, it was found that statistically lower (normal range) on babies who listen to lullaby mother’s voice according to babies who listen to lullaby stranger’s voice. Before care in three groups were not found the statistical difference on respiration values of babies, but aftercare, it was found that statistically lower (normal range) on babies who listen to lullaby stranger’s voice according to babies who listen to mother’s voice and control groups. Before care and after care were examined, fever signs did not reveal statistically significant difference in three groups. Conclusion: Lullaby concerts as being normal ranges of vital signs of infants and also helping to shorten hospitalization times should be preferred in the neonatal intensive care units.

Keywords: growth and development, lullaby, mother voice, vital signs

Procedia PDF Downloads 194
7011 Sarcasm Recognition System Using Hybrid Tone-Word Spotting Audio Mining Technique

Authors: Sandhya Baskaran, Hari Kumar Nagabushanam

Abstract:

Sarcasm sentiment recognition is an area of natural language processing that is being probed into in the recent times. Even with the advancements in NLP, typical translations of words, sentences in its context fail to provide the exact information on a sentiment or emotion of a user. For example, if something bad happens, the statement ‘That's just what I need, great! Terrific!’ is expressed in a sarcastic tone which could be misread as a positive sign by any text-based analyzer. In this paper, we are presenting a unique real time ‘word with its tone’ spotting technique which would provide the sentiment analysis for a tone or pitch of a voice in combination with the words being expressed. This hybrid approach increases the probability for identification of special sentiment like sarcasm much closer to the real world than by mining text or speech individually. The system uses a tone analyzer such as YIN-FFT which extracts pitch segment-wise that would be used in parallel with a speech recognition system. The clustered data is classified for sentiments and sarcasm score for each of it determined. Our Simulations demonstrates the improvement in f-measure of around 12% compared to existing detection techniques with increased precision and recall.

Keywords: sarcasm recognition, tone-word spotting, natural language processing, pitch analyzer

Procedia PDF Downloads 272
7010 Using Optical Character Recognition to Manage the Unstructured Disaster Data into Smart Disaster Management System

Authors: Dong Seop Lee, Byung Sik Kim

Abstract:

In the 4th Industrial Revolution, various intelligent technologies have been developed in many fields. These artificial intelligence technologies are applied in various services, including disaster management. Disaster information management does not just support disaster work, but it is also the foundation of smart disaster management. Furthermore, it gets historical disaster information using artificial intelligence technology. Disaster information is one of important elements of entire disaster cycle. Disaster information management refers to the act of managing and processing electronic data about disaster cycle from its’ occurrence to progress, response, and plan. However, information about status control, response, recovery from natural and social disaster events, etc. is mainly managed in the structured and unstructured form of reports. Those exist as handouts or hard-copies of reports. Such unstructured form of data is often lost or destroyed due to inefficient management. It is necessary to manage unstructured data for disaster information. In this paper, the Optical Character Recognition approach is used to convert handout, hard-copies, images or reports, which is printed or generated by scanners, etc. into electronic documents. Following that, the converted disaster data is organized into the disaster code system as disaster information. Those data are stored in the disaster database system. Gathering and creating disaster information based on Optical Character Recognition for unstructured data is important element as realm of the smart disaster management. In this paper, Korean characters were improved to over 90% character recognition rate by using upgraded OCR. In the case of character recognition, the recognition rate depends on the fonts, size, and special symbols of character. We improved it through the machine learning algorithm. These converted structured data is managed in a standardized disaster information form connected with the disaster code system. The disaster code system is covered that the structured information is stored and retrieve on entire disaster cycle such as historical disaster progress, damages, response, and recovery. The expected effect of this research will be able to apply it to smart disaster management and decision making by combining artificial intelligence technologies and historical big data.

Keywords: disaster information management, unstructured data, optical character recognition, machine learning

Procedia PDF Downloads 104
7009 To Study the New Invocation of Biometric Authentication Technique

Authors: Aparna Gulhane

Abstract:

Biometrics is the science and technology of measuring and analyzing biological data form the basis of research in biological measuring techniques for the purpose of people identification and recognition. In information technology, biometrics refers to technologies that measure and analyze human body characteristics, such as DNA, fingerprints, eye retinas and irises, voice patterns, facial patterns and hand measurements. Biometric systems are used to authenticate the person's identity. The idea is to use the special characteristics of a person to identify him. These papers present a biometric authentication techniques and actual deployment of potential by overall invocation of biometrics recognition, with an independent testing of various biometric authentication products and technology.

Keywords: types of biometrics, importance of biometric, review for biometrics and getting a new implementation, biometric authentication technique

Procedia PDF Downloads 294
7008 DBN-Based Face Recognition System Using Light Field

Authors: Bing Gu

Abstract:

Abstract—Most of Conventional facial recognition systems are based on image features, such as LBP, SIFT. Recently some DBN-based 2D facial recognition systems have been proposed. However, we find there are few DBN-based 3D facial recognition system and relative researches. 3D facial images include all the individual biometric information. We can use these information to build more accurate features, So we present our DBN-based face recognition system using Light Field. We can see Light Field as another presentation of 3D image, and Light Field Camera show us a way to receive a Light Field. We use the commercially available Light Field Camera to act as the collector of our face recognition system, and the system receive a state-of-art performance as convenient as conventional 2D face recognition system.

Keywords: DBN, face recognition, light field, Lytro

Procedia PDF Downloads 435
7007 UKIYO-E: User Knowledge Improvement Based on Youth Oriented Entertainment, Art Appreciation Support by Interacting with Picture

Authors: Haruya Tamaki, Tsugunosuke Sakai, Ryuichi Yoshida, Ryohei Egusa, Shigenori Inagaki, Etsuji Yamaguchi, Fusako Kusunoki, Miki Namatame, Masanori Sugimoto, Hiroshi Mizoguchi

Abstract:

Art appreciation is important as part of children education. Art appreciation can enrich sensibility and creativity. To enrich sensibility and creativity, the children have to learning knowledge of picture such as social and historical backgrounds and author intention. High learning effect can acquire by actively learning. In short, it is important that encourage learning of the knowledge about pictures actively. It is necessary that children feel like interest to encourage learning of the knowledge about pictures actively. In a general art museum, comments on pictures are done through writing. Thus, we expect that this method cannot arouse the interest of the children in pictures, because children feel like boring. In brief, learning about the picture information is difficult. Therefore, we are developing an art-appreciation support system that will encourage learning of the knowledge about pictures actively by children feel like interest. This system uses that Interacting with Pictures to learning of the knowledge about pictures. To Interacting with Pictures, children have to utterance by themselves. We expect that will encourage learning of the knowledge about pictures actively by Interacting with Pictures. To more actively learning, children can choose who talking with by information that location and movement of the children. This system must be able to acquire real-time knowledge of the location, movement, and voice of the children. We utilize the Microsoft’s Kinect v2 sensor and its library, namely, Kinect for Windows SDK and Speech Platform SDK v11 for this purpose. By using these sensor and library, we can determine the location, movement, and voice of the children. As the first step of this system, we developed ukiyo-e game that use ukiyo-e to appreciation object. Ukiyo-e is a traditional Japanese graphic art that has influenced the western society. Therefore, we believe that the ukiyo-e game will be appreciated. In this study, we applied talking to pictures to learn information about the pictures because we believe that learning information about the pictures by talking to the pictures is more interesting than commenting on the pictures using only texts. However, we cannot confirm if talking to the pictures is more interesting than commenting using texts only. Thus, we evaluated through EDA measurement whether the user develops an interest in the pictures while talking to them using voice recognition or by commenting on the pictures using texts only. Hence, we evaluated that children have interest to picture while talking to them using voice recognition through EDA measurement. In addition, we quantitatively evaluate that enjoyed this game or not and learning information about the pictures for primary schoolchildren. In this paper, we summarize these two evaluation results.

Keywords: actively learning, art appreciation, EDA, Kinect V2

Procedia PDF Downloads 267
7006 Analysis of Vocal Fold Vibrations from High-Speed Digital Images Based on Dynamic Time Warping

Authors: A. I. A. Rahman, Sh-Hussain Salleh, K. Ahmad, K. Anuar

Abstract:

Analysis of vocal fold vibration is essential for understanding the mechanism of voice production and for improving clinical assessment of voice disorders. This paper presents a Dynamic Time Warping (DTW) based approach to analyze and objectively classify vocal fold vibration patterns. The proposed technique was designed and implemented on a Glottal Area Waveform (GAW) extracted from high-speed laryngeal images by delineating the glottal edges for each image frame. Feature extraction from the GAW was performed using Linear Predictive Coding (LPC). Several types of voice reference templates from simulations of clear, breathy, fry, pressed and hyperfunctional voice productions were used. The patterns of the reference templates were first verified using the analytical signal generated through Hilbert transformation of the GAW. Samples from normal speakers’ voice recordings were then used to evaluate and test the effectiveness of this approach. The classification of the voice patterns using the technique of LPC and DTW gave the accuracy of 81%.

Keywords: dynamic time warping, glottal area waveform, linear predictive coding, high-speed laryngeal images, Hilbert transform

Procedia PDF Downloads 220
7005 Lip Localization Technique for Myanmar Consonants Recognition Based on Lip Movements

Authors: Thein Thein, Kalyar Myo San

Abstract:

Lip reading system is one of the different supportive technologies for hearing impaired, or elderly people or non-native speakers. For normal hearing persons in noisy environments or in conditions where the audio signal is not available, lip reading techniques can be used to increase their understanding of spoken language. Hearing impaired persons have used lip reading techniques as important tools to find out what was said by other people without hearing voice. Thus, visual speech information is important and become active research area. Using visual information from lip movements can improve the accuracy and robustness of a speech recognition system and the need for lip reading system is ever increasing for every language. However, the recognition of lip movement is a difficult task because of the region of interest (ROI) is nonlinear and noisy. Therefore, this paper proposes method to detect the accurate lips shape and to localize lip movement towards automatic lip tracking by using the combination of Otsu global thresholding technique and Moore Neighborhood Tracing Algorithm. Proposed method shows how accurate lip localization and tracking which is useful for speech recognition. In this work of study and experiments will be carried out the automatic lip localizing the lip shape for Myanmar consonants using the only visual information from lip movements which is useful for visual speech of Myanmar languages.

Keywords: lip reading, lip localization, lip tracking, Moore neighborhood tracing algorithm

Procedia PDF Downloads 332
7004 Face Tracking and Recognition Using Deep Learning Approach

Authors: Degale Desta, Cheng Jian

Abstract:

The most important factor in identifying a person is their face. Even identical twins have their own distinct faces. As a result, identification and face recognition are needed to tell one person from another. A face recognition system is a verification tool used to establish a person's identity using biometrics. Nowadays, face recognition is a common technique used in a variety of applications, including home security systems, criminal identification, and phone unlock systems. This system is more secure because it only requires a facial image instead of other dependencies like a key or card. Face detection and face identification are the two phases that typically make up a human recognition system.The idea behind designing and creating a face recognition system using deep learning with Azure ML Python's OpenCV is explained in this paper. Face recognition is a task that can be accomplished using deep learning, and given the accuracy of this method, it appears to be a suitable approach. To show how accurate the suggested face recognition system is, experimental results are given in 98.46% accuracy using Fast-RCNN Performance of algorithms under different training conditions.

Keywords: deep learning, face recognition, identification, fast-RCNN

Procedia PDF Downloads 110
7003 English Learning Speech Assistant Speak Application in Artificial Intelligence

Authors: Albatool Al Abdulwahid, Bayan Shakally, Mariam Mohamed, Wed Almokri

Abstract:

Artificial intelligence has infiltrated every part of our life and every field we can think of. With technical developments, artificial intelligence applications are becoming more prevalent. We chose ELSA speak because it is a magnificent example of Artificial intelligent applications, ELSA speak is a smartphone application that is free to download on both IOS and Android smartphones. ELSA speak utilizes artificial intelligence to help non-native English speakers pronounce words and phrases similar to a native speaker, as well as enhance their English skills. It employs speech-recognition technology that aids the application to excel the pronunciation of its users. This remarkable feature distinguishes ELSA from other voice recognition algorithms and increase the efficiency of the application. This study focused on evaluating ELSA speak application, by testing the degree of effectiveness based on survey questions. The results of the questionnaire were variable. The generality of the participants strongly agreed that ELSA has helped them enhance their pronunciation skills. However, a few participants were unconfident about the application’s ability to assist them in their learning journey.

Keywords: ELSA speak application, artificial intelligence, speech-recognition technology, language learning, english pronunciation

Procedia PDF Downloads 81
7002 Work with Children's Music Group: Important Aspects of Didactic and Artistic Performance

Authors: Eudjen Cinc

Abstract:

Work with a human voice, especially with a child s voice and cultivating the sound of the choir, presents an area of crucial importance for a conductor. We use the term conductor because it needs to be understood that regardless of whether we have in front of us an amateur or a professional choir, whether they are singers with a wealth of experience or children who are still developing and educating their inner ear so that in the future they could contribute to the development of choir music, the person who stands in front of the group and works with them, needs to have the characteristics of a conductor. Voice formation is a long-term process, without which there is no success in both solo and collective music performance.

Keywords: music group, conductor, collective, performance

Procedia PDF Downloads 202
7001 A Review on Predictive Sound Recognition System

Authors: Ajay Kadam, Ramesh Kagalkar

Abstract:

The proposed research objective is to add to a framework for programmed recognition of sound. In this framework the real errand is to distinguish any information sound stream investigate it & anticipate the likelihood of diverse sounds show up in it. To create and industrially conveyed an adaptable sound web crawler a flexible sound search engine. The calculation is clamor and contortion safe, computationally productive, and hugely adaptable, equipped for rapidly recognizing a short portion of sound stream caught through a phone microphone in the presence of frontal area voices and other predominant commotion, and through voice codec pressure, out of a database of over accessible tracks. The algorithm utilizes a combinatorial hashed time-recurrence group of stars examination of the sound, yielding ordinary properties, for example, transparency, in which numerous tracks combined may each be distinguished.

Keywords: fingerprinting, pure tone, white noise, hash function

Procedia PDF Downloads 303
7000 Lovely, Lyrical, Lilting: Kubrick’s Translation of Lolita’s Voice

Authors: Taylor La Carriere

Abstract:

“What I had madly possessed was not she, but my own creation, another, fanciful Lolita perhaps, more real than Lolita; overlapping, encasing he and having no will, no consciousness indeed, no life of her own,” Vladimir Nabokov writes in his seminal work, Lolita. Throughout Nabokov’s novel, the eponymous character is rendered nonexistent through unreliable narrator Humbert Humbert’s impenetrable narrative, infused with lyrical rationalization. Instead, Lolita is “safely solipsised,” as Humbert muses, solidifying the potential for the erasure of Lolita’s agency and identity. In this literary work, Lolita’s voice is reduced to a nearly invisible presence, only seen through the eyes of her captor. However, in Stanley Kubrick’s film adaptation of Lolita (1962), the “nymphet,” as Nabokov coins, reemerges with a voice of her own, fueled by a lyric impulse, that displaces Humbert’s first-person narration. The lyric, as defined by Catherine Ing, is the voice of the invisible; it is also characterized by performance, the concentrated utterance of individual emotion, and the appearance of spontaneity. The novel’s lyricism is largely in the service of Humbert’s “seductive” voice, while the film reorients it more to Lolita’s subjectivity. Through a close analysis of Kubrick’s cinematic techniques, this paper examines the emergence and translation of Lolita’s voice in contrast with Humbert’s attempts to silence her in Nabokov’s Lolita, hypothesizing that Kubrick translates Lolita’s presence into a visual and aural voice with lyrical attributes, exemplified through the establishment of an altered power dynamic, Sue Lyon’s transformative performance as the titular character, Nelson Riddle and Bob Harris’ musical score, and the omission of Humbert’s first-person point-of-view. In doing so, the film reclaims Lolita’s agency by taking instances of Lolita’s voice in the novel as depicted in the last half of the work and expanding upon them in a way only cinematic depictions could allow. The results of this study suggest that Lolita’s voice in Kubrick’s adaptation functions without disrupting the lyricism present in Nabokov’s source text, materializing through the actions, expressions, and performance of Sue Lyon in the film. This voice, fueled by a lyric impulse of its own, refutes the silence bestowed upon the titular character and enables its ultimate reclamation upon the silver screen.

Keywords: cinema, adaptation, Lolita, lyric voice

Procedia PDF Downloads 171