Search results for: taste machine
1074 Effects of Abiotic Stress on the Phytochemical Content and Bioactivity of Pistacia lentiscus L.
Authors: S. Mamoucha, N. Tsafantakis, Α. Ioannidis, S. Chatzipanagiotou, C. Nikolaou, L. Skaltsounis, N. Fokialakis, N. Christodoulakis
Abstract:
Introduction: Plant secondary metabolites (SM) can be grouped into three chemically distinct groups: terpenes, phenolics, and nitrogen-containing compounds. For many years the adaptive significance of SM was unknown. They were thought to be functionless end-products. Currently it is accepted that many secondary metabolites (also known as natural products) have important ecological roles in plants. For instance, they serve as attractants (odor, color, taste) for pollinators and seed-dispersing animals. Moreover, they protect plants from herbivores, microbial pathogens and from environmental stress (high and low temperatures, drought, alkalinity, salinity, radiation etc). It is well known that both biotic and abiotic stress often increase the accumulation of SM. The local climatic conditions, seasonal changes, external factors such as light, temperature, humidity affect the biosynthesis and composition of secondary metabolites. A well known dioecious evergreen plant, Pistacia lentiscus L. (mastic tree), was selected in order to study the metabolic variations occur in response to the different climate conditions, due to the seasonal variation and its effect on the biosynthesis of bioactive compounds. Materials-methods: Young and mature leaves were collected in January and July 2014, dried and extracted by accelerated solvent extraction (Dionex ASE™ 350) using solvents of increased polarity (DCM, MeOH, and H2O). GC-MS and UHPLC-HRMS analysis were carried out in order to define the nature and the relative abundance of SM. The antibacterial activity was evaluated by using the Agar Disc Diffusion Assay against ATCC and clinical isolates strains: Escherichia coli, Staphylococcus aureus, Pseudomonas aeruginosa, Candida albicans, Streptococcus mutans and Klebsiella pneumoniae. All tests were carried out in duplicate and the average radii of the inhibition zones were calculated for each extract. Results: According to the phytochemical profile obtained from each extract, the biosynthesis of SM varied both qualitatively and quantitatively under the two different types of seasonal stress. With exception of the biologically inactive nonpolar DCM extract of July, all extracts inhibited the growth of most of the investigated microorganisms. A clear positive correlation has been observed between the relative abundance of SM and the bioactivity of the DCM extracts of January and July. Observed changes during phytochemical analysis were mainly focused on the triterpenoid content. On the other hand, the bioactivity of the polar extracts (MeOH and H2O) of January and July resulted practically invariable against most of the microorganisms, besides the significant variation of the SM content due to the seasonal variation. Conclusion: Our results clearly confirmed the hypothesis of abiotic stress as an important regulating factor that significantly affects the biosynthesis of secondary metabolites and thus the presence of bioactive compounds. Acknowledgment: This work was supported by IKY - State Scholarship Foundation, Athens, Greece.Keywords: antibacterial screening, phytochemical profile, Pistacia lentiscus, abiotic stress
Procedia PDF Downloads 2561073 Efficient Motion Estimation by Fast Three Step Search Algorithm
Authors: S. M. Kulkarni, D. S. Bormane, S. L. Nalbalwar
Abstract:
The rapid development in the technology have dramatic impact on the medical health care field. Medical data base obtained with latest machines like CT Machine, MRI scanner requires large amount of memory storage and also it requires large bandwidth for transmission of data in telemedicine applications. Thus, there is need for video compression. As the database of medical images contain number of frames (slices), hence while coding of these images there is need of motion estimation. Motion estimation finds out movement of objects in an image sequence and gets motion vectors which represents estimated motion of object in the frame. In order to reduce temporal redundancy between successive frames of video sequence, motion compensation is preformed. In this paper three step search (TSS) block matching algorithm is implemented on different types of video sequences. It is shown that three step search algorithm produces better quality performance and less computational time compared with exhaustive full search algorithm.Keywords: block matching, exhaustive search motion estimation, three step search, video compression
Procedia PDF Downloads 4911072 Investigation of the Multiaxial Pedicle Screw Tulip Design Using Finite Element Analysis
Authors: S. Daqiqeh Rezaei, S. Mohajerzadeh, M. R. Sharifi
Abstract:
Pedicle screws are used to stabilize vertebrae and treat several types of spinal diseases and injuries. Multiaxial pedicle screws are a type of pedicle screw that increase surgical versatility, but they also increase design complexity. Failure of multiaxial pedicle screws caused by static loading, dynamic loading and fatigue can lead to irreparable damage to the patient. Inappropriate deformation of the multiaxial pedicle screw tulip can cause system failure. Investigation of deformation and stress in these tulips can be employed to optimize multiaxial pedicle screw design. The sensitivity of this matter necessitates precise analyzing and modeling of pedicle screws. In this work, three commercial multiaxial pedicle screw tulips and a newly designed tulip are investigated using finite element analysis. Employing video measuring machine (VMM), tulips are modeled. Afterwards, utilizing ANSYS, static analysis is performed on these models. In the end, stresses and displacements of the models are compared.Keywords: pedicle screw, multiaxial pedicle screw, finite element analysis, static analysis
Procedia PDF Downloads 3681071 Automated Adaptions of Semantic User- and Service Profile Representations by Learning the User Context
Authors: Nicole Merkle, Stefan Zander
Abstract:
Ambient Assisted Living (AAL) describes a technological and methodological stack of (e.g. formal model-theoretic semantics, rule-based reasoning and machine learning), different aspects regarding the behavior, activities and characteristics of humans. Hence, a semantic representation of the user environment and its relevant elements are required in order to allow assistive agents to recognize situations and deduce appropriate actions. Furthermore, the user and his/her characteristics (e.g. physical, cognitive, preferences) need to be represented with a high degree of expressiveness in order to allow software agents a precise evaluation of the users’ context models. The correct interpretation of these context models highly depends on temporal, spatial circumstances as well as individual user preferences. In most AAL approaches, model representations of real world situations represent the current state of a universe of discourse at a given point in time by neglecting transitions between a set of states. However, the AAL domain currently lacks sufficient approaches that contemplate on the dynamic adaptions of context-related representations. Semantic representations of relevant real-world excerpts (e.g. user activities) help cognitive, rule-based agents to reason and make decisions in order to help users in appropriate tasks and situations. Furthermore, rules and reasoning on semantic models are not sufficient for handling uncertainty and fuzzy situations. A certain situation can require different (re-)actions in order to achieve the best results with respect to the user and his/her needs. But what is the best result? To answer this question, we need to consider that every smart agent requires to achieve an objective, but this objective is mostly defined by domain experts who can also fail in their estimation of what is desired by the user and what not. Hence, a smart agent has to be able to learn from context history data and estimate or predict what is most likely in certain contexts. Furthermore, different agents with contrary objectives can cause collisions as their actions influence the user’s context and constituting conditions in unintended or uncontrolled ways. We present an approach for dynamically updating a semantic model with respect to the current user context that allows flexibility of the software agents and enhances their conformance in order to improve the user experience. The presented approach adapts rules by learning sensor evidence and user actions using probabilistic reasoning approaches, based on given expert knowledge. The semantic domain model consists basically of device-, service- and user profile representations. In this paper, we present how this semantic domain model can be used in order to compute the probability of matching rules and actions. We apply this probability estimation to compare the current domain model representation with the computed one in order to adapt the formal semantic representation. Our approach aims at minimizing the likelihood of unintended interferences in order to eliminate conflicts and unpredictable side-effects by updating pre-defined expert knowledge according to the most probable context representation. This enables agents to adapt to dynamic changes in the environment which enhances the provision of adequate assistance and affects positively the user satisfaction.Keywords: ambient intelligence, machine learning, semantic web, software agents
Procedia PDF Downloads 2811070 Feature Extraction Based on Contourlet Transform and Log Gabor Filter for Detection of Ulcers in Wireless Capsule Endoscopy
Authors: Nimisha Elsa Koshy, Varun P. Gopi, V. I. Thajudin Ahamed
Abstract:
The entire visualization of GastroIntestinal (GI) tract is not possible with conventional endoscopic exams. Wireless Capsule Endoscopy (WCE) is a low risk, painless, noninvasive procedure for diagnosing diseases such as bleeding, polyps, ulcers, and Crohns disease within the human digestive tract, especially the small intestine that was unreachable using the traditional endoscopic methods. However, analysis of massive images of WCE detection is tedious and time consuming to physicians. Hence, researchers have developed software methods to detect these diseases automatically. Thus, the effectiveness of WCE can be improved. In this paper, a novel textural feature extraction method is proposed based on Contourlet transform and Log Gabor filter to distinguish ulcer regions from normal regions. The results show that the proposed method performs well with a high accuracy rate of 94.16% using Support Vector Machine (SVM) classifier in HSV colour space.Keywords: contourlet transform, log gabor filter, ulcer, wireless capsule endoscopy
Procedia PDF Downloads 5401069 Recognition of Voice Commands of Mentor Robot in Noisy Environment Using Hidden Markov Model
Authors: Khenfer Koummich Fatma, Hendel Fatiha, Mesbahi Larbi
Abstract:
This paper presents an approach based on Hidden Markov Models (HMM: Hidden Markov Model) using HTK tools. The goal is to create a human-machine interface with a voice recognition system that allows the operator to teleoperate a mentor robot to execute specific tasks as rotate, raise, close, etc. This system should take into account different levels of environmental noise. This approach has been applied to isolated words representing the robot commands pronounced in two languages: French and Arabic. The obtained recognition rate is the same in both speeches, Arabic and French in the neutral words. However, there is a slight difference in favor of the Arabic speech when Gaussian white noise is added with a Signal to Noise Ratio (SNR) equals 30 dB, in this case; the Arabic speech recognition rate is 69%, and the French speech recognition rate is 80%. This can be explained by the ability of phonetic context of each speech when the noise is added.Keywords: Arabic speech recognition, Hidden Markov Model (HMM), HTK, noise, TIMIT, voice command
Procedia PDF Downloads 3861068 'Low Electronic Noise' Detector Technology in Computed Tomography
Authors: A. Ikhlef
Abstract:
Image noise in computed tomography, is mainly caused by the statistical noise, system noise reconstruction algorithm filters. Since last few years, low dose x-ray imaging became more and more desired and looked as a technical differentiating technology among CT manufacturers. In order to achieve this goal, several technologies and techniques are being investigated, including both hardware (integrated electronics and photon counting) and software (artificial intelligence and machine learning) based solutions. From a hardware point of view, electronic noise could indeed be a potential driver for low and ultra-low dose imaging. We demonstrated that the reduction or elimination of this term could lead to a reduction of dose without affecting image quality. Also, in this study, we will show that we can achieve this goal using conventional electronics (low cost and affordable technology), designed carefully and optimized for maximum detective quantum efficiency. We have conducted the tests using large imaging objects such as 30 cm water and 43 cm polyethylene phantoms. We compared the image quality with conventional imaging protocols with radiation as low as 10 mAs (<< 1 mGy). Clinical validation of such results has been performed as well.Keywords: computed tomography, electronic noise, scintillation detector, x-ray detector
Procedia PDF Downloads 1261067 Unearthing Air Traffic Control Officers Decision Instructional Patterns From Simulator Data for Application in Human Machine Teams
Authors: Zainuddin Zakaria, Sun Woh Lye
Abstract:
Despite the continuous advancements in automated conflict resolution tools, there is still a low rate of adoption of automation from Air Traffic Control Officers (ATCOs). Trust or acceptance in these tools and conformance to the individual ATCO preferences in strategy execution for conflict resolution are two key factors that impact their use. This paper proposes a methodology to unearth and classify ATCO conflict resolution strategies from simulator data of trained and qualified ATCOs. The methodology involves the extraction of ATCO executive control actions and the establishment of a system of strategy resolution classification based on ATCO radar commands and prevailing flight parameters in deconflicting a pair of aircraft. Six main strategies used to handle various categories of conflict were identified and discussed. It was found that ATCOs were about twice more likely to choose only vertical maneuvers in conflict resolution compared to horizontal maneuvers or a combination of both vertical and horizontal maneuvers.Keywords: air traffic control strategies, conflict resolution, simulator data, strategy classification system
Procedia PDF Downloads 1481066 Perspectives of Computational Modeling in Sanskrit Lexicons
Authors: Baldev Ram Khandoliyan, Ram Kishor
Abstract:
India has a classical tradition of Sanskrit Lexicons. Research work has been done on the study of Indian lexicography. India has seen amazing strides in Information and Communication Technology (ICT) applications for Indian languages in general and for Sanskrit in particular. Since Machine Translation from Sanskrit to other Indian languages is often the desired goal, traditional Sanskrit lexicography has attracted a lot of attention from the ICT and Computational Linguistics community. From Nighaŋţu and Nirukta to Amarakośa and Medinīkośa, Sanskrit owns a rich history of lexicography. As these kośas do not follow the same typology or standard in the selection and arrangement of the words and the information related to them, several types of Kośa-styles have emerged in this tradition. The model of a grammar given by Aṣṭādhyāyī is well appreciated by Indian and western linguists and grammarians. But the different models provided by lexicographic tradition also have importance. The general usefulness of Sanskrit traditional Kośas is well discussed by some scholars. That is most of the matter made available in the text. Some also have discussed the good arrangement of lexica. This paper aims to discuss some more use of the different models of Sanskrit lexicography especially focusing on its computational modeling and its use in different computational operations.Keywords: computational lexicography, Sanskrit Lexicons, nighanṭu, kośa, Amarkosa
Procedia PDF Downloads 1651065 Primary and Secondary Big Bangs Theory of Creation of Universe
Authors: Shyam Sunder Gupta
Abstract:
The current theory for the creation of the universe, the Big Bang theory, is widely accepted but leaves some unanswered questions. It does not explain the origin of the singularity or what causes the Big Bang. The theory of the Big Bang also does not explain why there is such a huge amount of dark energy and dark matter in our universe. Also, there is a question related to one universe or multiple universes which needs to be answered. This research addresses these questions using the Bhagvat Puran and other Vedic scriptures as the basis. There is a Unique Pure Energy Field that is eternal, infinite, and finest of all and never transforms when in its original form. The Carrier Particles of Unique Pure Energy are Param-anus- Fundamental Energy Particles. Param-anus and a combination of these particles create bigger particles from which the Universe gets created. For creation to initiate, Unique Pure Energy is represented in three phases: positive phase energy, neutral phase eternal time energy and negative phase energy. Positive phase energy further expands in three forms of creative energies (CE1, CE2andCE3). From CE1 energy, three energy modes, mode of activation, mode of action, and mode of darkness, were created. From these three modes, 16 Principles, subtlest forms of energies, namely Pradhan, Mahat-tattva, Time, Ego, Intellect, Mind, Sound, Space, Touch, Air, Form, Fire, Taste, Water, Smell, and Earth, get created. In the Mahat-tattva, dominant in the Mode of Darkness, CE1 energy creates innumerable primary singularities from seven principles: Pradhan, Mahat-tattva, Ego, Sky, Air, Fire, and Water. CE1 energy gets divided as CE2 and enters, along with three modes and time, in each singularity, and primary Big Bang takes place, and innumerable Invisible Universes get created. Each Universe has seven coverings of 7 principles, and each layer is 10 times thicker than the previous layer. By energy CE2, space in Invisible Universe under the coverings is divided into two halves. In the lower half, the process of evolution gets initiated, and seeds of 24 elements get created, out of which 5 fundamental elements, building blocks of matter, Sky, Air, Fire, Water and Earth, create seeds of stars, planets, galaxies and all other matter. Since 5 fundamental elements get created out of the mode of darkness, it explains why there is so much dark energy and dark matter in our Universe. This process of creation, in the lower half of Invisible universe continues for 2.16 billion years. Further, in the lower part of the energy field, exactly at the Centre of Invisible Universe, Secondary Singularity is created, through which, by force of Mode of Action, Secondary Big Bang takes place and Visible Universe gets created in the shape of Lotus Flower, expanding into upper part. Visible matter starts appearing after a gap of 360,000 years. Within the Visible Universe, a small part gets created known as the Phenomenal Material World, which is our Solar System, the sun being in the Centre. Diameter of Solar planetary system is 6.4 billion km.Keywords: invisible universe, phenomenal material world, primary Big Bang, secondary Big Bang, singularities, visible universe
Procedia PDF Downloads 891064 Wear Performance of Stellite 21 Cladded Overlay on Aisi 304L
Authors: Sandeep Singh Sandhua, Karanvir Singh Ghuman, Arun Kumar
Abstract:
Stellite 21 is cobalt based super alloy used in improving the wear performance of stainless steel engineering components subjected to harsh environmental conditions. This piece of research focuses on the wear analysis of satellite 21 cladded on AISI 304 L substrate using SMAW process. Bead on plate experiments were carried out by varying current and electrode manipulation techniques to optimize the dilution and microhardness. 80 Amp current and weaving technique was found to be optimum set of parameters for overlaying which were further used for multipass multilayer cladding of AISI 304 L substrate. The wear performance was examined on pin on dics wear testing machine under room temperature conditions. The results from this study show that Stellite 21 overlays show a significant improvement in the frictional wear resistance after TIG remelting. It is also established that low dilution procedures are important in controlling the metallurgical composition of these overlays which has a consequent effect in enhancing hardness and wear resistance of these overlays.Keywords: surfacing, stellite 21, dilution, SMAW, frictional wear, micro-hardness
Procedia PDF Downloads 2501063 A Resource Optimization Strategy for CPU (Central Processing Unit) Intensive Applications
Authors: Junjie Peng, Jinbao Chen, Shuai Kong, Danxu Liu
Abstract:
On the basis of traditional resource allocation strategies, the usage of resources on physical servers in cloud data center is great uncertain. It will cause waste of resources if the assignment of tasks is not enough. On the contrary, it will cause overload if the assignment of tasks is too much. This is especially obvious when the applications are the same type because of its resource preferences. Considering CPU intensive application is one of the most common types of application in the cloud, we studied the optimization strategy for CPU intensive applications on the same server. We used resource preferences to analyze the case that multiple CPU intensive applications run simultaneously, and put forward a model which can predict the execution time for CPU intensive applications which run simultaneously. Based on the prediction model, we proposed the method to select the appropriate number of applications for a machine. Experiments show that the model can predict the execution time accurately for CPU intensive applications. To improve the execution efficiency of applications, we propose a scheduling model based on priority for CPU intensive applications. Extensive experiments verify the validity of the scheduling model.Keywords: cloud computing, CPU intensive applications, resource optimization, strategy
Procedia PDF Downloads 2781062 User-Based Cannibalization Mitigation in an Online Marketplace
Authors: Vivian Guo, Yan Qu
Abstract:
Online marketplaces are not only digital places where consumers buy and sell merchandise, and they are also destinations for brands to connect with real consumers at the moment when customers are in the shopping mindset. For many marketplaces, brands have been important partners through advertising. There can be, however, a risk of advertising impacting a consumer’s shopping journey if it hurts the use experience or takes the user away from the site. Both could lead to the loss of transaction revenue for the marketplace. In this paper, we present user-based methods for cannibalization control by selectively turning off ads to users who are likely to be cannibalized by ads subject to business objectives. We present ways of measuring cannibalization of advertising in the context of an online marketplace and propose novel ways of measuring cannibalization through purchase propensity and uplift modeling. A/B testing has shown that our methods can significantly improve user purchase and engagement metrics while operating within business objectives. To our knowledge, this is the first paper that addresses cannibalization mitigation at the user-level in the context of advertising.Keywords: cannibalization, machine learning, online marketplace, revenue optimization, yield optimization
Procedia PDF Downloads 1601061 The Impact of Recurring Events in Fake News Detection
Authors: Ali Raza, Shafiq Ur Rehman Khan, Raja Sher Afgun Usmani, Asif Raza, Basit Umair
Abstract:
Detection of Fake news and missing information is gaining popularity, especially after the advancement in social media and online news platforms. Social media platforms are the main and speediest source of fake news propagation, whereas online news websites contribute to fake news dissipation. In this study, we propose a framework to detect fake news using the temporal features of text and consider user feedback to identify whether the news is fake or not. In recent studies, the temporal features in text documents gain valuable consideration from Natural Language Processing and user feedback and only try to classify the textual data as fake or true. This research article indicates the impact of recurring and non-recurring events on fake and true news. We use two models BERT and Bi-LSTM to investigate, and it is concluded from BERT we get better results and 70% of true news are recurring and rest of 30% are non-recurring.Keywords: natural language processing, fake news detection, machine learning, Bi-LSTM
Procedia PDF Downloads 221060 Corrosion Fatigue of Al-Mg Alloy 5052 in Sodium Chloride Solution Contains Some Inhibitors
Authors: Khalid Ahmed Eldwaib
Abstract:
In this study, Al-Mg alloy 5052 was used as the testing material. Corrosion fatigue life was studied for the alloy in 3.5% NaCl (pH=1, 3, 5, 7, 9, and 11), and 3.5% NaCl (pH=1) with inhibitors. The compound inhibitors were composed mainly of phosphate (PO4³-), adding a certain proportion of other nontoxic inhibitors so as to select alternatives to environmentally hazardous chromate (Cr2O7²-). The inhibitors were sodium dichromate Na2Cr2O7, sodium phosphate Na3PO4, sodium molybdate Na2MoO4, and sodium citrate Na3C6H5O7. The total amount of inhibiting pigments was at different concentrations (250,500,750, and 1000 ppm) in the solutions. Corrosion fatigue behavior was studied by using plane-bending corrosion fatigue machine with stress ratio R=0.5 and under the constant frequency of 13.3 Hz. Results show that in 3.5% NaCl the highest fatigue life (number of cycles to failure Nf) is obtained at pH=5 where the oxide film on aluminum has very low solubility, and the lowest number of cycles is obtained at pH=1, where the media is too aggressive (extremely acidic). When the concentration of inhibitor increases the cycles to failure increase. The surface morphology and fracture section of the specimens had been characterized through scanning electron microscope (SEM).Keywords: Al-Mg alloy 5052, corrosion, fatigue, inhibitors
Procedia PDF Downloads 4601059 Cross Coupling Sliding Mode Synchronization Control of Dual-Driving Feed System
Authors: Hong Lu, Wei Fan, Yongquan Zhang, Junbo Zhang
Abstract:
A cross coupling sliding synchronization control strategy is proposed for the dual-driving feed system. This technology will minimize the position error oscillation and achieve the precise synchronization performance in the high speed and high precision drive system, especially some high speed and high precision machine. Moreover, a cross coupling compensation matrix is provided to offset the mismatched disturbance and the disturbance observer is established to eliminate the chattering phenomenon. Performance comparisons of proposed dual-driving cross coupling sliding mode control (CCSMC), normal cross coupling control (CCC) strategy with PID control, and electronic virtual main shaft control (EVMSC) strategy with SMC control are investigated by simulation and a dual-driving control system; the results show the effectiveness of the proposed control scheme.Keywords: cross coupling matrix, dual motors, synchronization control, sliding mode control
Procedia PDF Downloads 3651058 On the Bias and Predictability of Asylum Cases
Authors: Panagiota Katsikouli, William Hamilton Byrne, Thomas Gammeltoft-Hansen, Tijs Slaats
Abstract:
An individual who demonstrates a well-founded fear of persecution or faces real risk of being subjected to torture is eligible for asylum. In Danish law, the exact legal thresholds reflect those established by international conventions, notably the 1951 Refugee Convention and the 1950 European Convention for Human Rights. These international treaties, however, remain largely silent when it comes to how states should assess asylum claims. As a result, national authorities are typically left to determine an individual’s legal eligibility on a narrow basis consisting of an oral testimony, which may itself be hampered by several factors, including imprecise language interpretation, insecurity or lacking trust towards the authorities among applicants. The leaky ground, on which authorities must assess their subjective perceptions of asylum applicants' credibility, questions whether, in all cases, adjudicators make the correct decision. Moreover, the subjective element in these assessments raises questions on whether individual asylum cases could be afflicted by implicit biases or stereotyping amongst adjudicators. In fact, recent studies have uncovered significant correlations between decision outcomes and the experience and gender of the assigned judge, as well as correlations between asylum outcomes and entirely external events such as weather and political elections. In this study, we analyze a publicly available dataset containing approximately 8,000 summaries of asylum cases, initially rejected, and re-tried by the Refugee Appeals Board (RAB) in Denmark. First, we look for variations in the recognition rates, with regards to a number of applicants’ features: their country of origin/nationality, their identified gender, their identified religion, their ethnicity, whether torture was mentioned in their case and if so, whether it was supported or not, and the year the applicant entered Denmark. In order to extract those features from the text summaries, as well as the final decision of the RAB, we applied natural language processing and regular expressions, adjusting for the Danish language. We observed interesting variations in recognition rates related to the applicants’ country of origin, ethnicity, year of entry and the support or not of torture claims, whenever those were made in the case. The appearance (or not) of significant variations in the recognition rates, does not necessarily imply (or not) bias in the decision-making progress. None of the considered features, with the exception maybe of the torture claims, should be decisive factors for an asylum seeker’s fate. We therefore investigate whether the decision can be predicted on the basis of these features, and consequently, whether biases are likely to exist in the decisionmaking progress. We employed a number of machine learning classifiers, and found that when using the applicant’s country of origin, religion, ethnicity and year of entry with a random forest classifier, or a decision tree, the prediction accuracy is as high as 82% and 85% respectively. tentially predictive properties with regards to the outcome of an asylum case. Our analysis and findings call for further investigation on the predictability of the outcome, on a larger dataset of 17,000 cases, which is undergoing.Keywords: asylum adjudications, automated decision-making, machine learning, text mining
Procedia PDF Downloads 951057 Clarifier Dialogue Interface to resolve linguistic ambiguities in E-Learning Environment
Authors: Dalila Souilem, Salma Boumiza, Abdelkarim Abdelkader
Abstract:
The Clarifier Dialogue Interface (CDI) is a part of an online teaching system based on human-machine communication in learning situation. This interface used in the system during the learning action specifically in the evaluation step, to clarify ambiguities in the learner's response. The CDI can generate patterns allowing access to an information system, using the selectors associated with lexical units. To instantiate these patterns, the user request (especially learner’s response), must be analyzed and interpreted to deduce the canonical form, the semantic form and the subject of the sentence. For the efficiency of this interface at the interpretation level, a set of substitution operators is carried out in order to extend the possibilities of manipulation with a natural language. A second approach that will be presented in this paper focuses on the object languages with new prospects such as combination of natural language with techniques of handling information system in the area of online education. So all operators, the CDI and other interfaces associated to the domain expertise and teaching strategies will be unified using FRAME representation form.Keywords: dialogue, e-learning, FRAME, information system, natural language
Procedia PDF Downloads 3771056 Automatic Classification of Periodic Heart Sounds Using Convolutional Neural Network
Authors: Jia Xin Low, Keng Wah Choo
Abstract:
This paper presents an automatic normal and abnormal heart sound classification model developed based on deep learning algorithm. MITHSDB heart sounds datasets obtained from the 2016 PhysioNet/Computing in Cardiology Challenge database were used in this research with the assumption that the electrocardiograms (ECG) were recorded simultaneously with the heart sounds (phonocardiogram, PCG). The PCG time series are segmented per heart beat, and each sub-segment is converted to form a square intensity matrix, and classified using convolutional neural network (CNN) models. This approach removes the need to provide classification features for the supervised machine learning algorithm. Instead, the features are determined automatically through training, from the time series provided. The result proves that the prediction model is able to provide reasonable and comparable classification accuracy despite simple implementation. This approach can be used for real-time classification of heart sounds in Internet of Medical Things (IoMT), e.g. remote monitoring applications of PCG signal.Keywords: convolutional neural network, discrete wavelet transform, deep learning, heart sound classification
Procedia PDF Downloads 3481055 Detecting Paraphrases in Arabic Text
Authors: Amal Alshahrani, Allan Ramsay
Abstract:
Paraphrasing is one of the important tasks in natural language processing; i.e. alternative ways to express the same concept by using different words or phrases. Paraphrases can be used in many natural language applications, such as Information Retrieval, Machine Translation, Question Answering, Text Summarization, or Information Extraction. To obtain pairs of sentences that are paraphrases we create a system that automatically extracts paraphrases from a corpus, which is built from different sources of news article since these are likely to contain paraphrases when they report the same event on the same day. There are existing simple standard approaches (e.g. TF-IDF vector space, cosine similarity) and alignment technique (e.g. Dynamic Time Warping (DTW)) for extracting paraphrase which have been applied to the English. However, the performance of these approaches could be affected when they are applied to another language, for instance Arabic language, due to the presence of phenomena which are not present in English, such as Free Word Order, Zero copula, and Pro-dropping. These phenomena will affect the performance of these algorithms. Thus, if we can analysis how the existing algorithms for English fail for Arabic then we can find a solution for Arabic. The results are promising.Keywords: natural language processing, TF-IDF, cosine similarity, dynamic time warping (DTW)
Procedia PDF Downloads 3861054 Study of Atmospheric Cascades Generated by Primary Comic Rays, from Simulations in Corsika for the City of Tunja in Colombia
Authors: Tathiana Yesenia Coy Mondragón, Jossitt William Vargas Cruz, Cristian Leonardo Gutiérrez Gómez
Abstract:
The study of cosmic rays is based on two fundamental pillars: the detection of secondary cosmic rays on the Earth's surface and the detection of the source and origin of the cascade. In addition, the constant flow of RC generates a lot of interest for study due to the incidence of various natural phenomena, which makes it relevant to characterize their incidence parameters to determine their effect not only at subsoil or terrestrial surface levels but also throughout the atmosphere. To determine the physical parameters of the primary cosmic ray, the implementation of robust algorithms capable of reconstructing the cascade from the measured values is required, with a high level of reliability. Therefore, it is proposed to build a machine learning system that will be fed from the cosmic ray simulations in CORSIKA at different energies that lie in a range [10⁹-10¹²] eV. in order to generate a trained particle and pattern recognition system to obtain greater efficiency when inferring the nature of the origin of the cascade for EAS in the atmosphere considering atmospheric models.Keywords: CORSIKA, cosmic rays, eas, Colombia
Procedia PDF Downloads 811053 Surface Roughness Modeling in Dry Face Milling of Annealed and Hardened AISI 52100 Steel
Authors: Mohieddine Benghersallah, Mohamed Zakaria Zahaf, Ali Medjber, Idriss Tibakh
Abstract:
The objective of this study is to analyse the effects of cutting parameters on surface roughness in dry face milling using statistical techniques. We studied the effect of the microstructure of AISI 52100 steel on machinability before and after hardening. The machining tests were carried out on a high rigidity vertical milling machine with a 25 mm diameter face milling cutter equipped with micro-grain bicarbide inserts with PVD (Ti, AlN) coating in GC1030 grade. A Taguchi L9 experiment plan is adopted. Analysis of variance (ANOVA) was used to determine the effects of cutting parameters (Vc, fz, ap) on the roughness (Ra) of the machined surface. Regression analysis to assess the machinability of steel presented mathematical models of roughness and the combination of parameters to minimize it. The recorded results show that feed per tooth has the most significant effect on the surface condition for both steel treatment conditions. The best roughnesses were obtained for the hardened AISI 52100 steel.Keywords: machinability, heat treatment, microstructure, surface roughness, Taguchi method
Procedia PDF Downloads 1471052 COVID-19 Analysis with Deep Learning Model Using Chest X-Rays Images
Authors: Uma Maheshwari V., Rajanikanth Aluvalu, Kumar Gautam
Abstract:
The COVID-19 disease is a highly contagious viral infection with major worldwide health implications. The global economy suffers as a result of COVID. The spread of this pandemic disease can be slowed if positive patients are found early. COVID-19 disease prediction is beneficial for identifying patients' health problems that are at risk for COVID. Deep learning and machine learning algorithms for COVID prediction using X-rays have the potential to be extremely useful in solving the scarcity of doctors and clinicians in remote places. In this paper, a convolutional neural network (CNN) with deep layers is presented for recognizing COVID-19 patients using real-world datasets. We gathered around 6000 X-ray scan images from various sources and split them into two categories: normal and COVID-impacted. Our model examines chest X-ray images to recognize such patients. Because X-rays are commonly available and affordable, our findings show that X-ray analysis is effective in COVID diagnosis. The predictions performed well, with an average accuracy of 99% on training photographs and 88% on X-ray test images.Keywords: deep CNN, COVID–19 analysis, feature extraction, feature map, accuracy
Procedia PDF Downloads 791051 Classification of Political Affiliations by Reduced Number of Features
Authors: Vesile Evrim, Aliyu Awwal
Abstract:
By the evolvement in technology, the way of expressing opinions switched the direction to the digital world. The domain of politics as one of the hottest topics of opinion mining research merged together with the behavior analysis for affiliation determination in text which constitutes the subject of this paper. This study aims to classify the text in news/blogs either as Republican or Democrat with the minimum number of features. As an initial set, 68 features which 64 are constituted by Linguistic Inquiry and Word Count (LIWC) features are tested against 14 benchmark classification algorithms. In the later experiments, the dimensions of the feature vector reduced based on the 7 feature selection algorithms. The results show that Decision Tree, Rule Induction and M5 Rule classifiers when used with SVM and IGR feature selection algorithms performed the best up to 82.5% accuracy on a given dataset. Further tests on a single feature and the linguistic based feature sets showed the similar results. The feature “function” as an aggregate feature of the linguistic category, is obtained as the most differentiating feature among the 68 features with 81% accuracy by itself in classifying articles either as Republican or Democrat.Keywords: feature selection, LIWC, machine learning, politics
Procedia PDF Downloads 3821050 An Approach to Make an Adaptive Immunoassay to Detect an Unknown Disease
Authors: Josselyn Mata Calidonio, Arianna I. Maddox, Kimberly Hamad-Schifferli
Abstract:
Rapid diagnostics are critical infectious disease tools that are designed to detect a known biomarker using antibodies specific to that biomarker. However, a way to detect unknown viruses has not yet been achieved in a paper test format. We describe here a route to make an adaptable paper immunoassay that can detect an unknown biomarker, demonstrating it on SARS-CoV-2 variants. The immunoassay repurposes cross-reactive antibodies raised against the alpha variant. Gold nanoparticles of two different colors conjugated to two different antibodies create a colorimetric signal, and machine learning of the resulting colorimetric pattern is used to train the assay to discriminate between variants of alpha and Omicron BA.5. By using principal component analysis, the colorimetric test patterns can pick up and discriminate an unknown that it has not encountered before, Omicron BA.1. The test has an accuracy of 100% and a potential calculated discriminatory power of 900. We show that it can be used adaptively and that it can be used to pick up emerging variants without the need to raise new antibodies.Keywords: adaptive immunoassay, detecting unknown viruses, gold nanoparticles, paper immunoassay, repurposing antibodies
Procedia PDF Downloads 1141049 Semi-Supervised Learning for Spanish Speech Recognition Using Deep Neural Networks
Authors: B. R. Campomanes-Alvarez, P. Quiros, B. Fernandez
Abstract:
Automatic Speech Recognition (ASR) is a machine-based process of decoding and transcribing oral speech. A typical ASR system receives acoustic input from a speaker or an audio file, analyzes it using algorithms, and produces an output in the form of a text. Some speech recognition systems use Hidden Markov Models (HMMs) to deal with the temporal variability of speech and Gaussian Mixture Models (GMMs) to determine how well each state of each HMM fits a short window of frames of coefficients that represents the acoustic input. Another way to evaluate the fit is to use a feed-forward neural network that takes several frames of coefficients as input and produces posterior probabilities over HMM states as output. Deep neural networks (DNNs) that have many hidden layers and are trained using new methods have been shown to outperform GMMs on a variety of speech recognition systems. Acoustic models for state-of-the-art ASR systems are usually training on massive amounts of data. However, audio files with their corresponding transcriptions can be difficult to obtain, especially in the Spanish language. Hence, in the case of these low-resource scenarios, building an ASR model is considered as a complex task due to the lack of labeled data, resulting in an under-trained system. Semi-supervised learning approaches arise as necessary tasks given the high cost of transcribing audio data. The main goal of this proposal is to develop a procedure based on acoustic semi-supervised learning for Spanish ASR systems by using DNNs. This semi-supervised learning approach consists of: (a) Training a seed ASR model with a DNN using a set of audios and their respective transcriptions. A DNN with a one-hidden-layer network was initialized; increasing the number of hidden layers in training, to a five. A refinement, which consisted of the weight matrix plus bias term and a Stochastic Gradient Descent (SGD) training were also performed. The objective function was the cross-entropy criterion. (b) Decoding/testing a set of unlabeled data with the obtained seed model. (c) Selecting a suitable subset of the validated data to retrain the seed model, thereby improving its performance on the target test set. To choose the most precise transcriptions, three confidence scores or metrics, regarding the lattice concept (based on the graph cost, the acoustic cost and a combination of both), was performed as selection technique. The performance of the ASR system will be calculated by means of the Word Error Rate (WER). The test dataset was renewed in order to extract the new transcriptions added to the training dataset. Some experiments were carried out in order to select the best ASR results. A comparison between a GMM-based model without retraining and the DNN proposed system was also made under the same conditions. Results showed that the semi-supervised ASR-model based on DNNs outperformed the GMM-model, in terms of WER, in all tested cases. The best result obtained an improvement of 6% relative WER. Hence, these promising results suggest that the proposed technique could be suitable for building ASR models in low-resource environments.Keywords: automatic speech recognition, deep neural networks, machine learning, semi-supervised learning
Procedia PDF Downloads 3391048 Distributed Manufacturing (DM)- Smart Units and Collaborative Processes
Authors: Hermann Kuehnle
Abstract:
Developments in ICT totally reshape manufacturing as machines, objects and equipment on the shop floors will be smart and online. Interactions with virtualizations and models of a manufacturing unit will appear exactly as interactions with the unit itself. These virtualizations may be driven by providers with novel ICT services on demand that might jeopardize even well established business models. Context aware equipment, autonomous orders, scalable machine capacity or networkable manufacturing unit will be the terminology to get familiar with in manufacturing and manufacturing management. Such newly appearing smart abilities with impact on network behavior, collaboration procedures and human resource development will make distributed manufacturing a preferred model to produce. Computing miniaturization and smart devices revolutionize manufacturing set ups, as virtualizations and atomization of resources unwrap novel manufacturing principles. Processes and resources obey novel specific laws and have strategic impact on manufacturing and major operational implications. Mechanisms from distributed manufacturing engaging interacting smart manufacturing units and decentralized planning and decision procedures already demonstrate important effects from this shift of focus towards collaboration and interoperability.Keywords: autonomous unit, networkability, smart manufacturing unit, virtualization
Procedia PDF Downloads 5261047 O.MG- It’s a Cyber-Enabled Fraud
Authors: Damola O. Lawal, David W. Gresty, Diane E. Gan, Louise Hewitt
Abstract:
This paper investigates the feasibility of using a programmable USB such as the O.MG Cable to perform a file tampering attack. Here, the O.MG Cable, an apparently harmless mobile device charger, is used in an unauthorized way to alter the content of a file (accounts record-January_Contributions.xlsx). The aim is to determine if a forensics analyst can reliably determine who has altered the target file; the O.MG Cable or the user of the machine. This work highlights some of the traces of the O.MG Cable left behind on the target computer itself, such as the Product ID (PID) and Vendor ID (ID). Also discussed is the O.MG Cable’s behavior during the experiments. We determine if a forensics analyst could identify if any evidence has been left behind by the programmable device on the target file once it has been removed from the computer to establish if the analyst would be able to link the traces left by the O.MG Cable to the file tampering. It was discovered that the forensic analyst might mistake the actions of the O.MG Cable for the computer users. Experiments carried out in this work could further the discussion as to whether an innocent user could be punished for the unauthorized changes made by a programmable device.Keywords: O.MG cable, programmable USB, file tampering attack, digital evidence credibility, miscarriage of justice, cyber fraud
Procedia PDF Downloads 1591046 Productivity Improvement of Faffa Food Share Company Using a Computerized Maintenance Management System
Authors: Gadisa Alemayehu, Muralidhar Avvari, Atkilt Mulu G.
Abstract:
Since 1962 EC, the Faffa Food Share Company has been producing and supplying flour (famix) and value-added flour (baby food) in Ethiopia. It meets nearly all of the country's total flour demand, both for relief and commercial markets. However, it is incompetent in the international market due to a poor maintenance management system. The results of recorded documents and stopwatches revealed that frequent failure machines, as well as a poor maintenance management system, cause increased production downtimes, resulting in a 29.19 percent decrease in production from the planned production. As a result, the current study's goal is to recommend newly developed software for use in and as a Computerized Maintenance Management System (CMMS). As a result, the system increases machine reliability and decreases the frequency of equipment failure, reducing breakdown time and maintenance costs. The company's overall manufacturing performance improved by 4.45 percent, particularly after the implementation of the CMMS.Keywords: CMMS, manufacturing performance, delivery, availability, flexibility, Faffa Food Share Company
Procedia PDF Downloads 1371045 Red Dawn in the Desert: A World-Systems Analysis of the Maritime Silk Road Initiative
Authors: Toufic Sarieddine
Abstract:
The current debate on the hegemonic impact of China’s Belt and Road Initiative (BRI) is of two opposing strands: Resilient and absolute US hegemony on the one hand and various models of multipolar hegemony such as bifurcation on the other. Bifurcation theories illustrate an unprecedented division of hegemonic functions between China and the US, whereby Beijing becomes the world’s economic hegemon, leaving Washington the world’s military hegemon and security guarantor. While consensus points to China being the main driver of unipolarity’s rupturing, the debate among bifurcationists is on the location of the first rupture. In this regard, the Middle East and North Africa (MENA) region has seen increasing Chinese foreign direct investment in recent years while that to other regions has declined, ranking it second in 2018 as part of the financing for the Maritime Silk Road Initiative (MSRI). China has also become the top trade partner of 11 states in the MENA region, as well as its top source of machine imports, surpassing the US and achieving an overall trade surplus almost double that of Washington’s. These are among other features outlined in world-systems analysis (WSA) literature which correspond with the emergence of a new hegemon. WSA is further utilized to gauge other facets of China’s increasing involvement in MENA and assess whether bifurcation is unfolding therein. These features of hegemony include the adoption of China’s modi operandi, economic dominance in production, trade, and finance, military capacity, cultural hegemony in ideology, education, and language, and the promotion of a general interest around which to rally potential peripheries (MENA states in this case). China’s modi operandi has seen some adoption with regards to support against the United Nations Convention on the Law of the Sea, oil bonds denominated in the yuan, and financial institutions such as the Shanghai Gold Exchange enjoying increasing Arab patronage. However, recent elections in Qatar, as well as liberal reforms in Saudi Arabia, demonstrate Washington’s stronger normative influence. Meanwhile, Washington’s economic dominance is challenged by China’s sizable machine exports, increasing overall imports, and widening trade surplus, but retains some clout via dominant arms and transport exports, as well as free-trade deals across the region. Militarily, Washington bests Beijing’s arms exports, has a dominant and well-established presence in the region, and successfully blocked Beijing’s attempt to penetrate through the UAE. Culturally, Beijing enjoys higher favorability in Arab public opinion, and its broadcast networks have found some resonance with Arab audiences. In education, the West remains MENA students’ preferred destination. Further, while Mandarin has become increasingly available in schools across MENA, its usage and availability still lag far behind English. Finally, Beijing’s general interest in infrastructure provision and prioritizing economic development over social justice and democracy provides an avenue for increased incorporation between Beijing and the MENA region. The overall analysis shows solid progress towards bifurcation in MENA.Keywords: belt and road initiative, hegemony, Middle East and North Africa, world-systems analysis
Procedia PDF Downloads 108