Search results for: automated facial recognition
2197 Power Quality Modeling Using Recognition Learning Methods for Waveform Disturbances
Authors: Sang-Keun Moon, Hong-Rok Lim, Jin-O Kim
Abstract:
This paper presents a Power Quality (PQ) modeling and filtering processes for the distribution system disturbances using recognition learning methods. Typical PQ waveforms with mathematical applications and gathered field data are applied to the proposed models. The objective of this paper is analyzing PQ data with respect to monitoring, discriminating, and evaluating the waveform of power disturbances to ensure the system preventative system failure protections and complex system problem estimations. Examined signal filtering techniques are used for the field waveform noises and feature extractions. Using extraction and learning classification techniques, the efficiency was verified for the recognition of the PQ disturbances with focusing on interactive modeling methods in this paper. The waveform of selected 8 disturbances is modeled with randomized parameters of IEEE 1159 PQ ranges. The range, parameters, and weights are updated regarding field waveform obtained. Along with voltages, currents have same process to obtain the waveform features as the voltage apart from some of ratings and filters. Changing loads are causing the distortion in the voltage waveform due to the drawing of the different patterns of current variation. In the conclusion, PQ disturbances in the voltage and current waveforms indicate different types of patterns of variations and disturbance, and a modified technique based on the symmetrical components in time domain was proposed in this paper for the PQ disturbances detection and then classification. Our method is based on the fact that obtained waveforms from suggested trigger conditions contain potential information for abnormality detections. The extracted features are sequentially applied to estimation and recognition learning modules for further studies.Keywords: power quality recognition, PQ modeling, waveform feature extraction, disturbance trigger condition, PQ signal filtering
Procedia PDF Downloads 1862196 Comparison of Nucleic Acid Extraction Platforms On Tissue Samples
Authors: Siti Rafeah Md Rafei, Karen Wang Yanping, Park Mi Kyoung
Abstract:
Tissue samples are precious supply for molecular studies or disease identification diagnosed using molecular assays, namely real-time PCR (qPCR). It is critical to establish the most favorable nucleic acid extraction that gives the PCR-amplifiable genomic DNA. Furthermore, automated nucleic acid extraction is an appealing alternative to labor-intensive manual methods. Operational complexity, defined as the number of steps required to obtain an extracted sample, is one of the criteria in the comparison. Here we are comparing the One BioMed’s automated X8 platform with the commercially available manual-operated kits from QIAGEN Mini Kit and Roche. We extracted DNA from rat fresh-frozen tissue (from different type of organs) in the matrices. After tissue pre-treatment, it is added to the One BioMed’s X8 pre-filled cartridge, and the QIAGEN QIAmp column respectively. We found that the results after subjecting the eluates to the Real Time PCR using BIORAD CFX are comparable.Keywords: DNA extraction, frozen tissue, PCR, qPCR, rat
Procedia PDF Downloads 1612195 Combined Automatic Speech Recognition and Machine Translation in Business Correspondence Domain for English-Croatian
Authors: Sanja Seljan, Ivan Dunđer
Abstract:
The paper presents combined automatic speech recognition (ASR) for English and machine translation (MT) for English and Croatian in the domain of business correspondence. The first part presents results of training the ASR commercial system on two English data sets, enriched by error analysis. The second part presents results of machine translation performed by online tool Google Translate for English and Croatian and Croatian-English language pairs. Human evaluation in terms of usability is conducted and internal consistency calculated by Cronbach's alpha coefficient, enriched by error analysis. Automatic evaluation is performed by WER (Word Error Rate) and PER (Position-independent word Error Rate) metrics, followed by investigation of Pearson’s correlation with human evaluation.Keywords: automatic machine translation, integrated language technologies, quality evaluation, speech recognition
Procedia PDF Downloads 4842194 Determination of the Thermally Comfortable Air Temperature with Consideration of Individual Clothing and Activity as Preparation for a New Smart Home Heating System
Authors: Alexander Peikos, Carole Binsfeld
Abstract:
The aim of this paper is to determine a thermally comfortable air temperature in an automated living room. This calculated temperature should serve as input for a user-specific and dynamic heating control in such a living space. In addition to the usual physical factors (air temperature, humidity, air velocity, and radiation temperature), individual clothing and activity should be taken into account. The calculation of such a temperature is based on different methods and indices which are usually used for the evaluation of the thermal comfort. The thermal insulation of the worn clothing is determined with a Radio Frequency Identification system. The activity performed is only taken into account indirectly through the generated heart rate. All these methods are ultimately very well suited for use in temperature regulation in an automated home, but still require further research and extensive evaluation.Keywords: smart home, thermal comfort, predicted mean vote, radio frequency identification
Procedia PDF Downloads 1592193 Multimodal Deep Learning for Human Activity Recognition
Authors: Ons Slimene, Aroua Taamallah, Maha Khemaja
Abstract:
In recent years, human activity recognition (HAR) has been a key area of research due to its diverse applications. It has garnered increasing attention in the field of computer vision. HAR plays an important role in people’s daily lives as it has the ability to learn advanced knowledge about human activities from data. In HAR, activities are usually represented by exploiting different types of sensors, such as embedded sensors or visual sensors. However, these sensors have limitations, such as local obstacles, image-related obstacles, sensor unreliability, and consumer concerns. Recently, several deep learning-based approaches have been proposed for HAR and these approaches are classified into two categories based on the type of data used: vision-based approaches and sensor-based approaches. This research paper highlights the importance of multimodal data fusion from skeleton data obtained from videos and data generated by embedded sensors using deep neural networks for achieving HAR. We propose a deep multimodal fusion network based on a twostream architecture. These two streams use the Convolutional Neural Network combined with the Bidirectional LSTM (CNN BILSTM) to process skeleton data and data generated by embedded sensors and the fusion at the feature level is considered. The proposed model was evaluated on a public OPPORTUNITY++ dataset and produced a accuracy of 96.77%.Keywords: human activity recognition, action recognition, sensors, vision, human-centric sensing, deep learning, context-awareness
Procedia PDF Downloads 1012192 The Integration of Digital Humanities into the Sociology of Knowledge Approach to Discourse Analysis
Authors: Gertraud Koch, Teresa Stumpf, Alejandra Tijerina García
Abstract:
Discourse analysis research approaches belong to the central research strategies applied throughout the humanities; they focus on the countless forms and ways digital texts and images shape present-day notions of the world. Despite the constantly growing number of relevant digital, multimodal discourse resources, digital humanities (DH) methods are thus far not systematically developed and accessible for discourse analysis approaches. Specifically, the significance of multimodality and meaning plurality modelling are yet to be sufficiently addressed. In order to address this research gap, the D-WISE project aims to develop a prototypical working environment as digital support for the sociology of knowledge approach to discourse analysis and new IT-analysis approaches for the use of context-oriented embedding representations. Playing an essential role throughout our research endeavor is the constant optimization of hermeneutical methodology in the use of (semi)automated processes and their corresponding epistemological reflection. Among the discourse analyses, the sociology of knowledge approach to discourse analysis is characterised by the reconstructive and accompanying research into the formation of knowledge systems in social negotiation processes. The approach analyses how dominant understandings of a phenomenon develop, i.e., the way they are expressed and consolidated by various actors in specific arenas of discourse until a specific understanding of the phenomenon and its socially accepted structure are established. This article presents insights and initial findings from D-WISE, a joint research project running since 2021 between the Institute of Anthropological Studies in Culture and History and the Language Technology Group of the Department of Informatics at the University of Hamburg. As an interdisciplinary team, we develop central innovations with regard to the availability of relevant DH applications by building up a uniform working environment, which supports the procedure of the sociology of knowledge approach to discourse analysis within open corpora and heterogeneous, multimodal data sources for researchers in the humanities. We are hereby expanding the existing range of DH methods by developing contextualized embeddings for improved modelling of the plurality of meaning and the integrated processing of multimodal data. The alignment of this methodological and technical innovation is based on the epistemological working methods according to grounded theory as a hermeneutic methodology. In order to systematically relate, compare, and reflect the approaches of structural-IT and hermeneutic-interpretative analysis, the discourse analysis is carried out both manually and digitally. Using the example of current discourses on digitization in the healthcare sector and the associated issues regarding data protection, we have manually built an initial data corpus of which the relevant actors and discourse positions are analysed in conventional qualitative discourse analysis. At the same time, we are building an extensive digital corpus on the same topic based on the use and further development of entity-centered research tools such as topic crawlers and automated newsreaders. In addition to the text material, this consists of multimodal sources such as images, video sequences, and apps. In a blended reading process, the data material is filtered, annotated, and finally coded with the help of NLP tools such as dependency parsing, named entity recognition, co-reference resolution, entity linking, sentiment analysis, and other project-specific tools that are being adapted and developed. The coding process is carried out (semi-)automated by programs that propose coding paradigms based on the calculated entities and their relationships. Simultaneously, these can be specifically trained by manual coding in a closed reading process and specified according to the content issues. Overall, this approach enables purely qualitative, fully automated, and semi-automated analyses to be compared and reflected upon.Keywords: entanglement of structural IT and hermeneutic-interpretative analysis, multimodality, plurality of meaning, sociology of knowledge approach to discourse analysis
Procedia PDF Downloads 2262191 Multimodal Sentiment Analysis With Web Based Application
Authors: Shreyansh Singh, Afroz Ahmed
Abstract:
Sentiment Analysis intends to naturally reveal the hidden mentality that we hold towards an entity. The total of this assumption over a populace addresses sentiment surveying and has various applications. Current text-based sentiment analysis depends on the development of word embeddings and Machine Learning models that take in conclusion from enormous text corpora. Sentiment Analysis from text is presently generally utilized for consumer loyalty appraisal and brand insight investigation. With the expansion of online media, multimodal assessment investigation is set to carry new freedoms with the appearance of integral information streams for improving and going past text-based feeling examination using the new transforms methods. Since supposition can be distinguished through compelling follows it leaves, like facial and vocal presentations, multimodal opinion investigation offers good roads for examining facial and vocal articulations notwithstanding the record or printed content. These methodologies use the Recurrent Neural Networks (RNNs) with the LSTM modes to increase their performance. In this study, we characterize feeling and the issue of multimodal assessment investigation and audit ongoing advancements in multimodal notion examination in various spaces, including spoken surveys, pictures, video websites, human-machine, and human-human connections. Difficulties and chances of this arising field are additionally examined, promoting our theory that multimodal feeling investigation holds critical undiscovered potential.Keywords: sentiment analysis, RNN, LSTM, word embeddings
Procedia PDF Downloads 1192190 Behavioral and EEG Reactions in Children during Recognition of Emotionally Colored Sentences That Describe the Choice Situation
Authors: Tuiana A. Aiusheeva, Sergey S. Tamozhnikov, Alexander E. Saprygin, Arina A. Antonenko, Valentina V. Stepanova, Natalia N. Tolstykh, Alexander N. Savostyanov
Abstract:
Situation of choice is an important condition for the formation of essential character qualities of a child, such as being initiative, responsible, hard-working. We have studied the behavioral and EEG reactions in Russian schoolchildren during recognition of syntactic errors in emotionally colored sentences that describe the choice situation. Twenty healthy children (mean age 9,0±0,3 years, 12 boys, 8 girls) were examined. Forty sentences were selected for the experiment; the half of them contained a syntactic error. The experiment additionally had the hidden condition: 50% of the sentences described the children's own choice and were emotionally colored (positive or negative). The other 50% of the sentences described the forced-choice situation, also with positive or negative coloring. EEG were recorded during execution of error-recognition task. Reaction time and quality of syntactic error detection were chosen as behavioral measures. Event-related spectral perturbation (ERSP) was applied to characterize the oscillatory brain activity of children. There were two time-frequency intervals in EEG reactions: (1) 500-800 ms in the 3-7 Hz frequency range (theta synchronization) and (2) 500-1000 ms in the 8-12 Hz range (alpha desynchronization). We found out that behavioral and brain reactions in child brain during recognition of positive and negative sentences describing forced-choice situation did not have significant differences. Theta synchronization and alpha desynchronization were stronger during recognition of sentences with children's own choice, especially with negative coloring. Also, the quality and execution time of the task were higher for this types of sentences. The results of our study will be useful for improvement of teaching methods and diagnostics of children affective disorders.Keywords: choice situation, electroencephalogram (EEG), emotionally colored sentences, schoolchildren
Procedia PDF Downloads 2692189 Optimizing the Capacity of a Convolutional Neural Network for Image Segmentation and Pattern Recognition
Authors: Yalong Jiang, Zheru Chi
Abstract:
In this paper, we study the factors which determine the capacity of a Convolutional Neural Network (CNN) model and propose the ways to evaluate and adjust the capacity of a CNN model for best matching to a specific pattern recognition task. Firstly, a scheme is proposed to adjust the number of independent functional units within a CNN model to make it be better fitted to a task. Secondly, the number of independent functional units in the capsule network is adjusted to fit it to the training dataset. Thirdly, a method based on Bayesian GAN is proposed to enrich the variances in the current dataset to increase its complexity. Experimental results on the PASCAL VOC 2010 Person Part dataset and the MNIST dataset show that, in both conventional CNN models and capsule networks, the number of independent functional units is an important factor that determines the capacity of a network model. By adjusting the number of functional units, the capacity of a model can better match the complexity of a dataset.Keywords: CNN, convolutional neural network, capsule network, capacity optimization, character recognition, data augmentation, semantic segmentation
Procedia PDF Downloads 1532188 Antioxidant Face Mask from Purple Sweet Potato (Ipomea Batatas) with Oleum Cytrus
Authors: Lilis Kistriyani, Dine Olisvia, Lutfa Rahmawati
Abstract:
Facial mask is an important part of every beauty treatment because it will give a smooth and gentle effect on the face. This research is done to make edible film that will be applied for face mask. The main ingredient in making this edible film is purple sweet potato powder with the addition of glycerol as plasticizer. One of the ingredients in purple sweet potato is a flavonoid compound. The purpose of this study was to determine the effect of increasing the amount of glycerol to flavonoids release and the effect on the physical properties and biological properties of edible film produced. The stages of this research are the making of edible film, then perform some analysis, among others, spectrophotometer UV-vis analysis to find out how many flavonoids can be released into facial skin, tensile strength and elongation of break analysis, biodegradability analysis, and microbiological analysis. The variation of edible film is the volume of glycerol that is 1 ml, 2 ml, 3 ml. The results of spectrophotometer UV-vis analysis showed that the most flavonoid release concentration is 20.33 ppm in the 2 ml glycerol variation. The best tensile strength value is 8,502 N, and the greatest elongation of break value is 14% in 1 ml glycerol variation. In the biodegradability test, the more volume of glycerol added the faster the edible film is degraded. The results of microbiological analysis showed that purple sweet potato extract has the ability to inhibit the growth of Propionibacterium acnes seen in the presence of inhibiting zone which is 18.9 mm.Keywords: face mask, edible film, plasticizer, flavonoid
Procedia PDF Downloads 1762187 Keypoint Detection Method Based on Multi-Scale Feature Fusion of Attention Mechanism
Authors: Xiaoxiao Li, Shuangcheng Jia, Qian Li
Abstract:
Keypoint detection has always been a challenge in the field of image recognition. This paper proposes a novelty keypoint detection method which is called Multi-Scale Feature Fusion Convolutional Network with Attention (MFFCNA). We verified that the multi-scale features with the attention mechanism module have better feature expression capability. The feature fusion between different scales makes the information that the network model can express more abundant, and the network is easier to converge. On our self-made street sign corner dataset, we validate the MFFCNA model with an accuracy of 97.8% and a recall of 81%, which are 5 and 8 percentage points higher than the HRNet network, respectively. On the COCO dataset, the AP is 71.9%, and the AR is 75.3%, which are 3 points and 2 points higher than HRNet, respectively. Extensive experiments show that our method has a remarkable improvement in the keypoint recognition tasks, and the recognition effect is better than the existing methods. Moreover, our method can be applied not only to keypoint detection but also to image classification and semantic segmentation with good generality.Keywords: keypoint detection, feature fusion, attention, semantic segmentation
Procedia PDF Downloads 1192186 Optimized Brain Computer Interface System for Unspoken Speech Recognition: Role of Wernicke Area
Authors: Nassib Abdallah, Pierre Chauvet, Abd El Salam Hajjar, Bassam Daya
Abstract:
In this paper, we propose an optimized brain computer interface (BCI) system for unspoken speech recognition, based on the fact that the constructions of unspoken words rely strongly on the Wernicke area, situated in the temporal lobe. Our BCI system has four modules: (i) the EEG Acquisition module based on a non-invasive headset with 14 electrodes; (ii) the Preprocessing module to remove noise and artifacts, using the Common Average Reference method; (iii) the Features Extraction module, using Wavelet Packet Transform (WPT); (iv) the Classification module based on a one-hidden layer artificial neural network. The present study consists of comparing the recognition accuracy of 5 Arabic words, when using all the headset electrodes or only the 4 electrodes situated near the Wernicke area, as well as the selection effect of the subbands produced by the WPT module. After applying the articial neural network on the produced database, we obtain, on the test dataset, an accuracy of 83.4% with all the electrodes and all the subbands of 8 levels of the WPT decomposition. However, by using only the 4 electrodes near Wernicke Area and the 6 middle subbands of the WPT, we obtain a high reduction of the dataset size, equal to approximately 19% of the total dataset, with 67.5% of accuracy rate. This reduction appears particularly important to improve the design of a low cost and simple to use BCI, trained for several words.Keywords: brain-computer interface, speech recognition, artificial neural network, electroencephalography, EEG, wernicke area
Procedia PDF Downloads 2722185 The Wear Recognition on Guide Surface Based on the Feature of Radar Graph
Authors: Youhang Zhou, Weimin Zeng, Qi Xie
Abstract:
Abstract: In order to solve the wear recognition problem of the machine tool guide surface, a new machine tool guide surface recognition method based on the radar-graph barycentre feature is presented in this paper. Firstly, the gray mean value, skewness, projection variance, flat degrees and kurtosis features of the guide surface image data are defined as primary characteristics. Secondly, data Visualization technology based on radar graph is used. The visual barycentre graphical feature is demonstrated based on the radar plot of multi-dimensional data. Thirdly, a classifier based on the support vector machine technology is used, the radar-graph barycentre feature and wear original feature are put into the classifier separately for classification and comparative analysis of classification and experiment results. The calculation and experimental results show that the method based on the radar-graph barycentre feature can detect the guide surface effectively.Keywords: guide surface, wear defects, feature extraction, data visualization
Procedia PDF Downloads 5192184 Host-Assisted Delivery of a Model Drug to Genomic DNA: Key Information From Ultrafast Spectroscopy and in Silico Study
Authors: Ria Ghosh, Soumendra Singh, Dipanjan Mukherjee, Susmita Mondal, Monojit Das, Uttam Pal, Aniruddha Adhikari, Aman Bhushan, Surajit Bose, Siddharth Sankar Bhattacharyya, Debasish Pal, Tanusri Saha-Dasgupta, Maitree Bhattacharyya, Debasis Bhattacharyya, Asim Kumar Mallick, Ranjan Das, Samir Kumar Pal
Abstract:
Drug delivery to a target without adverse effects is one of the major criteria for clinical use. Herein, we have made an attempt to explore the delivery efficacy of SDS surfactant in a monomer and micellar stage during the delivery of the model drug, Toluidine Blue (TB) from the micellar cavity to DNA. Molecular recognition of pre-micellar SDS encapsulated TB with DNA occurs at a rate constant of k1 ~652 s 1. However, no significant release of encapsulated TB at micellar concentration was observed within the experimental time frame. This originated from the higher binding affinity of TB towards the nano-cavity of SDS at micellar concentration which does not allow the delivery of TB from the nano-cavity of SDS micelles to DNA. Thus, molecular recognition controls the extent of DNA recognition by TB which in turn modulates the rate of delivery of TB from SDS in a concentration-dependent manner.Keywords: DNA, drug delivery, micelle, pre-micelle, SDS, toluidine blue
Procedia PDF Downloads 1132183 A Comparative Study of k-NN and MLP-NN Classifiers Using GA-kNN Based Feature Selection Method for Wood Recognition System
Authors: Uswah Khairuddin, Rubiyah Yusof, Nenny Ruthfalydia Rosli
Abstract:
This paper presents a comparative study between k-Nearest Neighbour (k-NN) and Multi-Layer Perceptron Neural Network (MLP-NN) classifier using Genetic Algorithm (GA) as feature selector for wood recognition system. The features have been extracted from the images using Grey Level Co-Occurrence Matrix (GLCM). The use of GA based feature selection is mainly to ensure that the database used for training the features for the wood species pattern classifier consists of only optimized features. The feature selection process is aimed at selecting only the most discriminating features of the wood species to reduce the confusion for the pattern classifier. This feature selection approach maintains the ‘good’ features that minimizes the inter-class distance and maximizes the intra-class distance. Wrapper GA is used with k-NN classifier as fitness evaluator (GA-kNN). The results shows that k-NN is the best choice of classifier because it uses a very simple distance calculation algorithm and classification tasks can be done in a short time with good classification accuracy.Keywords: feature selection, genetic algorithm, optimization, wood recognition system
Procedia PDF Downloads 5452182 A Supervised Learning Data Mining Approach for Object Recognition and Classification in High Resolution Satellite Data
Authors: Mais Nijim, Rama Devi Chennuboyina, Waseem Al Aqqad
Abstract:
Advances in spatial and spectral resolution of satellite images have led to tremendous growth in large image databases. The data we acquire through satellites, radars and sensors consists of important geographical information that can be used for remote sensing applications such as region planning, disaster management. Spatial data classification and object recognition are important tasks for many applications. However, classifying objects and identifying them manually from images is a difficult task. Object recognition is often considered as a classification problem, this task can be performed using machine-learning techniques. Despite of many machine-learning algorithms, the classification is done using supervised classifiers such as Support Vector Machines (SVM) as the area of interest is known. We proposed a classification method, which considers neighboring pixels in a region for feature extraction and it evaluates classifications precisely according to neighboring classes for semantic interpretation of region of interest (ROI). A dataset has been created for training and testing purpose; we generated the attributes by considering pixel intensity values and mean values of reflectance. We demonstrated the benefits of using knowledge discovery and data-mining techniques, which can be on image data for accurate information extraction and classification from high spatial resolution remote sensing imagery.Keywords: remote sensing, object recognition, classification, data mining, waterbody identification, feature extraction
Procedia PDF Downloads 3402181 Effect of Lullabies on Babies Stress and Relaxation Symptoms in the Neonatal Intensive Care Units
Authors: Meltem Kürtüncü, Işın Alkan
Abstract:
Objective: This study was carried out with an experimental design in order to determine whether the lullaby, which was listened from mother’s voice and a stranger’s voice to the babies born at term and hospitalized in neonatal intensive care unit, had an effect on stress and relaxation symptoms of the infants. Method: Data from the study were obtained from 90 newborn babies who were hospitalized in Neonatal Intensive Care Unit of Zonguldak Maternity And Children Hospital between September 2015-January 2016 and who met the eligibility criteria. Lullaby concert was performed by choosing one of the suitable care hours. Stress and relaxation symptoms were recorded by the researcher on “Newborn response follow-up form” at pre-care and post-care. Results: After lullaby concert when stress symptoms compared to infants in the experimental and control groups before the care was not detected statistically significant difference between crying, contraction, facial grimacing, flushing, cyanosis and the rates of increase in temperature. After care, crying, contractions, facial grimacing, flushing, and restlessness revealed a statistically significant difference between the groups, but as the cyanosis and temperature increased stress responses did not result in a significant difference between the groups. In the control group babies the crying, contraction, facial grimacing, flushing, and restlessness behaviors rates were found to be significantly higher than experimental group babies. After lullaby concert when relaxation symptoms compared to infants in the experimental and control groups before the care, eye contact rates who listen to lullaby from mother’s voice was found to be significantly higher than infants who listen to lullaby from stranger’s voice and infants in the control group. After care as eye contact, smiling, sucking/searching, yawning, non-crying and sleep behaviors relaxation symptoms revealed statistically significant results. In the control group, these behaviors were found statistically lower degree than the experimental groups. Conclusion: Lullaby concerts as masking the ambient noise, reducing the stress symptoms and increasing the relaxation symptoms, and also for soothing and stimulant affects, due to ease the transition to the sleep state should be preferred in the neonatal intensive care units.Keywords: lullaby, mother voice, relaxation, stress
Procedia PDF Downloads 2312180 Positive Bias and Length Bias in Deep Neural Networks for Premises Selection
Authors: Jiaqi Huang, Yuheng Wang
Abstract:
Premises selection, the task of selecting a set of axioms for proving a given conjecture, is a major bottleneck in automated theorem proving. An array of deep-learning-based methods has been established for premises selection, but a perfect performance remains challenging. Our study examines the inaccuracy of deep neural networks in premises selection. Through training network models using encoded conjecture and axiom pairs from the Mizar Mathematical Library, two potential biases are found: the network models classify more premises as necessary than unnecessary, referred to as the ‘positive bias’, and the network models perform better in proving conjectures that paired with more axioms, referred to as ‘length bias’. The ‘positive bias’ and ‘length bias’ discovered could inform the limitation of existing deep neural networks.Keywords: automated theorem proving, premises selection, deep learning, interpreting deep learning
Procedia PDF Downloads 1832179 Distributed Processing for Content Based Lecture Video Retrieval on Hadoop Framework
Authors: U. S. N. Raju, Kothuri Sai Kiran, Meena G. Kamal, Vinay Nikhil Pabba, Suresh Kanaparthi
Abstract:
There is huge amount of lecture video data available for public use, and many more lecture videos are being created and uploaded every day. Searching for videos on required topics from this huge database is a challenging task. Therefore, an efficient method for video retrieval is needed. An approach for automated video indexing and video search in large lecture video archives is presented. As the amount of video lecture data is huge, it is very inefficient to do the processing in a centralized computation framework. Hence, Hadoop Framework for distributed computing for Big Video Data is used. First, step in the process is automatic video segmentation and key-frame detection to offer a visual guideline for the video content navigation. In the next step, we extract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames. The OCR and detected slide text line types are adopted for keyword extraction, by which both video- and segment-level keywords are extracted for content-based video browsing and search. The performance of the indexing process can be improved for a large database by using distributed computing on Hadoop framework.Keywords: video lectures, big video data, video retrieval, hadoop
Procedia PDF Downloads 5342178 "Revolutionizing Geographic Data: CADmapper's Automated Precision in CAD Drawing Transformation"
Authors: Toleen Alaqqad, Kadi Alshabramiy, Suad Zaafarany, Basma Musallam
Abstract:
CADmapper is a significant tool of software for transforming geographic data into realistic CAD drawings. It speeds up and simplifies the conversion process by automating it. This allows architects, urban planners, engineers, and geographic information system (GIS) experts to solely concentrate on the imaginative and scientific parts of their projects. While the future incorporation of AI has the potential for further improvements, CADmapper's current capabilities make it an indispensable asset in the business. It covers a combination of 2D and 3D city and urban area models. The user can select a specific square section of the map to view, and the fee is based on the dimensions of the area being viewed. The procedure is straightforward: you choose the area you want, then pick whether or not to include topography. 3D architectural data (if available), followed by selecting whatever design program or CAD style you want to publish the document which contains more than 200 free broad town plans in DXF format. If you desire to specify a bespoke area, it's free up to 1 km2.Keywords: cadmaper, gdata, 2d and 3d data conversion, automated cad drawing, urban planning software
Procedia PDF Downloads 682177 Little Retrieval Augmented Generation for Named Entity Recognition: Toward Lightweight, Generative, Named Entity Recognition Through Prompt Engineering, and Multi-Level Retrieval Augmented Generation
Authors: Sean W. T. Bayly, Daniel Glover, Don Horrell, Simon Horrocks, Barnes Callum, Stuart Gibson, Mac Misuira
Abstract:
We assess suitability of recent, ∼7B parameter, instruction-tuned Language Models Mistral-v0.3, Llama-3, and Phi-3, for Generative Named Entity Recognition (GNER). Our proposed Multi-Level Information Retrieval method achieves notable improvements over finetuned entity-level and sentence-level methods. We consider recent developments at the cross roads of prompt engineering and Retrieval Augmented Generation (RAG), such as EmotionPrompt. We conclude that language models directed toward this task are highly capable when distinguishing between positive classes (precision). However, smaller models seem to struggle to find all entities (recall). Poorly defined classes such as ”Miscellaneous” exhibit substantial declines in performance, likely due to the ambiguity it introduces to the prompt. This is partially resolved through a self verification method using engineered prompts containing knowledge of the stricter class definitions, particularly in areas where their boundaries are in danger of overlapping, such as the conflation between the location ”Britain” and the nationality ”British”. Finally, we explore correlations between model performance on the GNER task with performance on relevant academic benchmarks.Keywords: generative named entity recognition, information retrieval, lightweight artificial intelligence, prompt engineering, personal information identification, retrieval augmented generation, self verification
Procedia PDF Downloads 462176 Fight the Burnout: Phase Two of a NICU Nurse Wellness Bundle
Authors: Megan Weisbart
Abstract:
Background/Significance: The Intensive Care Unit (ICU) environment contributes to nurse burnout. Burnout costs include decreased employee compassion, missed workdays, worse patient outcomes, diminished job performance, high turnover, and higher organizational cost. Meaningful recognition, nurturing of interpersonal connections, and mindfulness-based interventions are associated with decreased burnout. The purpose of this quality improvement project was to decrease Neonatal ICU (NICU) nurse burnout using a Wellness Bundle that fosters meaningful recognition, interpersonal connections and includes mindfulness-based interventions. Methods: The Professional Quality of Life Scale Version 5 (ProQOL5) was used to measure burnout before Wellness Bundle implementation, after six months, and will be given yearly for three years. Meaningful recognition bundle items include Online submission and posting of staff shoutouts, recognition events, Nurses Week and Unit Practice Council member gifts, and an employee recognition program. Fostering of interpersonal connections bundle items include: Monthly staff games with prizes, social events, raffle fundraisers, unit blog, unit wellness basket, and a wellness resource sheet. Quick coherence techniques were implemented at staff meetings and huddles as a mindfulness-based intervention. Findings: The mean baseline burnout score of 14 NICU nurses was 20.71 (low burnout). The baseline range was 13-28, with 11 nurses experiencing low burnout, three nurses experiencing moderate burnout, and zero nurses experiencing high burnout. After six months of the Wellness Bundle Implementation, the mean burnout score of 39 NICU nurses was 22.28 (low burnout). The range was 14-31, with 22 nurses experiencing low burnout, 17 nurses experiencing moderate burnout, and zero nurses experiencing high burnout. Conclusion: A NICU Wellness Bundle that incorporated meaningful recognition, fostering of interpersonal connections, and mindfulness-based activities was implemented to improve work environments and decrease nurse burnout. Participation bias and low baseline response rate may have affected the reliability of the data and necessitate another comparative measure of burnout in one year.Keywords: burnout, NICU, nurse, wellness
Procedia PDF Downloads 872175 Effect of Monotonically Decreasing Parameters on Margin Softmax for Deep Face Recognition
Authors: Umair Rashid
Abstract:
Normally softmax loss is used as the supervision signal in face recognition (FR) system, and it boosts the separability of features. In the last two years, a number of techniques have been proposed by reformulating the original softmax loss to enhance the discriminating power of Deep Convolutional Neural Networks (DCNNs) for FR system. To learn angularly discriminative features Cosine-Margin based softmax has been adjusted as monotonically decreasing angular function, that is the main challenge for angular based softmax. On that issue, we propose monotonically decreasing element for Cosine-Margin based softmax and also, we discussed the effect of different monotonically decreasing parameters on angular Margin softmax for FR system. We train the model on publicly available dataset CASIA- WebFace via our proposed monotonically decreasing parameters for cosine function and the tests on YouTube Faces (YTF, Labeled Face in the Wild (LFW), VGGFace1 and VGGFace2 attain the state-of-the-art performance.Keywords: deep convolutional neural networks, cosine margin face recognition, softmax loss, monotonically decreasing parameter
Procedia PDF Downloads 1012174 Image Processing of Scanning Electron Microscope Micrograph of Ferrite and Pearlite Steel for Recognition of Micro-Constituents
Authors: Subir Gupta, Subhas Ganguly
Abstract:
In this paper, we demonstrate the new area of application of image processing in metallurgical images to develop the more opportunity for structure-property correlation based approaches of alloy design. The present exercise focuses on the development of image processing tools suitable for phrase segmentation, grain boundary detection and recognition of micro-constituents in SEM micrographs of ferrite and pearlite steels. A comprehensive data of micrographs have been experimentally developed encompassing the variation of ferrite and pearlite volume fractions and taking images at different magnification (500X, 1000X, 15000X, 2000X, 3000X and 5000X) under scanning electron microscope. The variation in the volume fraction has been achieved using four different plain carbon steel containing 0.1, 0.22, 0.35 and 0.48 wt% C heat treated under annealing and normalizing treatments. The obtained data pool of micrographs arbitrarily divided into two parts to developing training and testing sets of micrographs. The statistical recognition features for ferrite and pearlite constituents have been developed by learning from training set of micrographs. The obtained features for microstructure pattern recognition are applied to test set of micrographs. The analysis of the result shows that the developed strategy can successfully detect the micro constitutes across the wide range of magnification and variation of volume fractions of the constituents in the structure with an accuracy of about +/- 5%.Keywords: SEM micrograph, metallurgical image processing, ferrite pearlite steel, microstructure
Procedia PDF Downloads 1992173 A Stepwise Approach to Automate the Search for Optimal Parameters in Seasonal ARIMA Models
Authors: Manisha Mukherjee, Diptarka Saha
Abstract:
Reliable forecasts of univariate time series data are often necessary for several contexts. ARIMA models are quite popular among practitioners in this regard. Hence, choosing correct parameter values for ARIMA is a challenging yet imperative task. Thus, a stepwise algorithm is introduced to provide automatic and robust estimates for parameters (p; d; q)(P; D; Q) used in seasonal ARIMA models. This process is focused on improvising the overall quality of the estimates, and it alleviates the problems induced due to the unidimensional nature of the methods that are currently used such as auto.arima. The fast and automated search of parameter space also ensures reliable estimates of the parameters that possess several desirable qualities, consequently, resulting in higher test accuracy especially in the cases of noisy data. After vigorous testing on real as well as simulated data, the algorithm doesn’t only perform better than current state-of-the-art methods, it also completely obviates the need for human intervention due to its automated nature.Keywords: time series, ARIMA, auto.arima, ARIMA parameters, forecast, R function
Procedia PDF Downloads 1652172 Altasreef: Automated System of Quran Verbs for Urdu Language
Authors: Haq Nawaz, Muhammad Amjad Iqbal, Kamran Malik
Abstract:
"Altasreef" is an automated system available for Web and Android users which provide facility to the users to learn the Quran verbs. It provides the facility to the users to practice the learned material and also provide facility of exams of Arabic verbs variation focusing on Quran text. Arabic is a highly inflectional language. Almost all of its words connect to roots of three, four or five letters which approach the meaning of all their inflectional forms. In Arabic, a verb is formed by inserting the consonants into one of a set of verb patterns. Suffixes and prefixes are then added to generate the meaning of number, person, and gender. The active/passive voice and perfective aspect and other patterns are than generated. This application is designed for learners of Quranic Arabic who already have learn basics of Arabic conjugation. Application also provides the facility of translation of generated patterns. These translations are generated with the help of rule-based approach to give 100% results to the learners.Keywords: NLP, Quran, Computational Linguistics, E Learning
Procedia PDF Downloads 1672171 Automated Prepaid Billing Subscription System
Authors: Adekunle K. O, Adeniyi A. E, Kolawole E
Abstract:
One of the most dramatic trends in the communications market in recent years has been the growth of prepaid services. Today, prepaid no longer constitutes the low-revenue, basic-service segment. It is driven by a high margin, value-add service customers who view it as a convenient way of retaining control over their usage and communication spending while expecting high service levels. To service providers, prepaid services offer the advantage of reducing bad accounts while allowing them to predict usage and plan network resources. Yet, the real-time demands of prepaid services require a scalable, real-time platform to manage customers through their entire life cycle. It delivers integrated real-time rating, voucher management, recharge management, customer care and service provisioning for the generation of new prepaid services. It carries high scalability that can handle millions of prepaid customers in real-time through their entire life cycle.Keywords: prepaid billing, voucher management, customers, automated, security
Procedia PDF Downloads 1152170 Using Speech Emotion Recognition as a Longitudinal Biomarker for Alzheimer’s Diseases
Authors: Yishu Gong, Liangliang Yang, Jianyu Zhang, Zhengyu Chen, Sihong He, Xusheng Zhang, Wei Zhang
Abstract:
Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that affects millions of people worldwide and is characterized by cognitive decline and behavioral changes. People living with Alzheimer’s disease often find it hard to complete routine tasks. However, there are limited objective assessments that aim to quantify the difficulty of certain tasks for AD patients compared to non-AD people. In this study, we propose to use speech emotion recognition (SER), especially the frustration level, as a potential biomarker for quantifying the difficulty patients experience when describing a picture. We build an SER model using data from the IEMOCAP dataset and apply the model to the DementiaBank data to detect the AD/non-AD group difference and perform longitudinal analysis to track the AD disease progression. Our results show that the frustration level detected from the SER model can possibly be used as a cost-effective tool for objective tracking of AD progression in addition to the Mini-Mental State Examination (MMSE) score.Keywords: Alzheimer’s disease, speech emotion recognition, longitudinal biomarker, machine learning
Procedia PDF Downloads 1132169 Automated Parking System
Authors: N. Arunraj, C. P. V. Paul, D. M. D. Jayawardena, W. N. D. Fernando
Abstract:
Traffic congestion with increased numbers of vehicles is already a serious issue for many countries. The absence of sufficient parking spaces adds to the issue. Motorists are forced to wait in long queues to park their vehicles. This adds to the inconvenience faced by a motorist, kept waiting for a slot allocation, manually done along with the parking payment calculation. In Sri Lanka, nowadays, parking systems use barcode technology to identify the vehicles at both the entrance and the exit points. Customer management is handled by the use of man power. A parking space is, generally permanently sub divided according to the vehicle type. Here, again, is an issue. Parking spaces are not utilized to the maximum. The current arrangement leaves room for unutilized parking spaces. Accordingly, there is a need to manage the parking space dynamically. As a vehicle enters the parking area, available space has to be assigned for the vehicle according to the vehicle type. The system, Automated Parking System (APS), provides an automated solution using RFID Technology to identify the vehicles. Simultaneously, an algorithm manages the space allocation dynamically. With this system, there is no permanent parking slot allocation for a vehicle type. A desktop application manages the customer. A Web application is used to manage the external users with their reservations. The system also has an android application to view the nearest parking area from the current location. APS is built using java and php. It uses LED panels to guide the user inside the parking area to find the allocated parking slot accurately. The system ensures efficient performance, saving precious time for a customer. Compared with the current parking systems, APS interacts with users and increases customer satisfaction as well.Keywords: RFID, android, web based system, barcode, algorithm, LED panels
Procedia PDF Downloads 5992168 Non-Invasive Characterization of the Mechanical Properties of Arterial Walls
Authors: Bruno RamaëL, GwenaëL Page, Catherine Knopf-Lenoir, Olivier Baledent, Anne-Virginie Salsac
Abstract:
No routine technique currently exists for clinicians to measure the mechanical properties of vascular walls non-invasively. Most of the data available in the literature come from traction or dilatation tests conducted ex vivo on native blood vessels. The objective of the study is to develop a non-invasive characterization technique based on Magnetic Resonance Imaging (MRI) measurements of the deformation of vascular walls under pulsating blood flow conditions. The goal is to determine the mechanical properties of the vessels by inverse analysis, coupling imaging measurements and numerical simulations of the fluid-structure interactions. The hyperelastic properties are identified using Solidworks and Ansys workbench (ANSYS Inc.) solving an optimization technique. The vessel of interest targeted in the study is the common carotid artery. In vivo MRI measurements of the vessel anatomy and inlet velocity profiles was acquired along the facial vascular network on a cohort of 30 healthy volunteers: - The time-evolution of the blood vessel contours and, thus, of the cross-section surface area was measured by 3D imaging angiography sequences of phase-contrast MRI. - The blood flow velocity was measured using a 2D CINE MRI phase contrast (PC-MRI) method. Reference arterial pressure waveforms were simultaneously measured in the brachial artery using a sphygmomanometer. The three-dimensional (3D) geometry of the arterial network was reconstructed by first creating an STL file from the raw MRI data using the open source imaging software ITK-SNAP. The resulting geometry was then transformed with Solidworks into volumes that are compatible with Ansys softwares. Tetrahedral meshes of the wall and fluid domains were built using the ANSYS Meshing software, with a near-wall mesh refinement method in the case of the fluid domain to improve the accuracy of the fluid flow calculations. Ansys Structural was used for the numerical simulation of the vessel deformation and Ansys CFX for the simulation of the blood flow. The fluid structure interaction simulations showed that the systolic and diastolic blood pressures of the common carotid artery could be taken as reference pressures to identify the mechanical properties of the different arteries of the network. The coefficients of the hyperelastic law were identified using Ansys Design model for the common carotid. Under large deformations, a stiffness of 800 kPa is measured, which is of the same order of magnitude as the Young modulus of collagen fibers. Areas of maximum deformations were highlighted near bifurcations. This study is a first step towards patient-specific characterization of the mechanical properties of the facial vessels. The method is currently applied on patients suffering from facial vascular malformations and on patients scheduled for facial reconstruction. Information on the blood flow velocity as well as on the vessel anatomy and deformability will be key to improve surgical planning in the case of such vascular pathologies.Keywords: identification, mechanical properties, arterial walls, MRI measurements, numerical simulations
Procedia PDF Downloads 319