Search results for: processing based on signal identification
31670 A Cost Effective Approach to Develop Mid-Size Enterprise Software Adopted the Waterfall Model
Authors: Mohammad Nehal Hasnine, Md Kamrul Hasan Chayon, Md Mobasswer Rahman
Abstract:
Organizational tendencies towards computer-based information processing have been observed noticeably in the third-world countries. Many enterprises are taking major initiatives towards computerized working environment because of massive benefits of computer-based information processing. However, designing and developing information resource management software for small and mid-size enterprises under budget costs and strict deadline is always challenging for software engineers. Therefore, we introduced an approach to design mid-size enterprise software by using the Waterfall model, which is one of the SDLC (Software Development Life Cycles), in a cost effective way. To fulfill research objectives, in this study, we developed mid-sized enterprise software named “BSK Management System” that assists enterprise software clients with information resource management and perform complex organizational tasks. Waterfall model phases have been applied to ensure that all functions, user requirements, strategic goals, and objectives are met. In addition, Rich Picture, Structured English, and Data Dictionary have been implemented and investigated properly in engineering manner. Furthermore, an assessment survey with 20 participants has been conducted to investigate the usability and performance of the proposed software. The survey results indicated that our system featured simple interfaces, easy operation and maintenance, quick processing, and reliable and accurate transactions.Keywords: end-user application development, enterprise software design, information resource management, usability
Procedia PDF Downloads 43831669 Identification Strategies for Unknown Victims from Mass Disasters and Unknown Perpetrators from Violent Crime or Terrorist Attacks
Authors: Michael Josef Schwerer
Abstract:
Background: The identification of unknown victims from mass disasters, violent crimes, or terrorist attacks is frequently facilitated through information from missing persons lists, portrait photos, old or recent pictures showing unique characteristics of a person such as scars or tattoos, or simply reference samples from blood relatives for DNA analysis. In contrast, the identification or at least the characterization of an unknown perpetrator from criminal or terrorist actions remains challenging, particularly in the absence of material or data for comparison, such as fingerprints, which had been previously stored in criminal records. In scenarios that result in high levels of destruction of the perpetrator’s corpse, for instance, blast or fire events, the chance for a positive identification using standard techniques is further impaired. Objectives: This study shows the forensic genetic procedures in the Legal Medicine Service of the German Air Force for the identification of unknown individuals, including such cases in which reference samples are not available. Scenarios requiring such efforts predominantly involve aircraft crash investigations, which are routinely carried out by the German Air Force Centre of Aerospace Medicine as one of the Institution’s essential missions. Further, casework by military police or military intelligence is supported based on administrative cooperation. In the talk, data from study projects, as well as examples from real casework, will be demonstrated and discussed with the audience. Methods: Forensic genetic identification in our laboratories involves the analysis of Short Tandem Repeats and Single Nucleotide Polymorphisms in nuclear DNA along with mitochondrial DNA haplotyping. Extended DNA analysis involves phenotypic markers for skin, hair, and eye color together with the investigation of a person’s biogeographic ancestry. Assessment of the biological age of an individual employs CpG-island methylation analysis using bisulfite-converted DNA. Forensic Investigative Genealogy assessment allows the detection of an unknown person’s blood relatives in reference databases. Technically, end-point-PCR, real-time PCR, capillary electrophoresis, pyrosequencing as well as next generation sequencing using flow-cell-based and chip-based systems are used. Results and Discussion: Optimization of DNA extraction from various sources, including difficult matrixes like formalin-fixed, paraffin-embedded tissues, degraded specimens from decomposed bodies or from decedents exposed to blast or fire events, provides soil for successful PCR amplification and subsequent genetic profiling. For cases with extremely low yields of extracted DNA, whole genome preamplification protocols are successfully used, particularly regarding genetic phenotyping. Improved primer design for CpG-methylation analysis, together with validated sampling strategies for the analyzed substrates from, e.g., lymphocyte-rich organs, allows successful biological age estimation even in bodies with highly degraded tissue material. Conclusions: Successful identification of unknown individuals or at least their phenotypic characterization using pigmentation markers together with age-informative methylation profiles, possibly supplemented by family tree search employing Forensic Investigative Genealogy, can be provided in specialized laboratories. However, standard laboratory procedures must be adapted to work with difficult and highly degraded sample materials.Keywords: identification, forensic genetics, phenotypic markers, CPG methylation, biological age estimation, forensic investigative genealogy
Procedia PDF Downloads 5131668 Video Heart Rate Measurement for the Detection of Trauma-Related Stress States
Authors: Jarek Krajewski, David Daxberger, Luzi Beyer
Abstract:
Finding objective and non-intrusive measurements of emotional and psychopathological states (e.g., post-traumatic stress disorder, PTSD) is an important challenge. Thus, the proposed approach here uses Photoplethysmographic imaging (PPGI) applying facial RGB Cam videos to estimate heart rate levels. A pipeline for the signal processing of the raw image has been proposed containing different preprocessing approaches, e.g., Independent Component Analysis, Non-negative Matrix factorization, and various other artefact correction approaches. Under resting and constant light conditions, we reached a sensitivity of 84% for pulse peak detection. The results indicate that PPGI can be a suitable solution for providing heart rate data derived from these indirectly post-traumatic stress states.Keywords: heart rate, PTSD, PPGI, stress, preprocessing
Procedia PDF Downloads 12431667 Prediction of Product Size Distribution of a Vertical Stirred Mill Based on Breakage Kinetics
Authors: C. R. Danielle, S. Erik, T. Patrick, M. Hugh
Abstract:
In the last decade there has been an increase in demand for fine grinding due to the depletion of coarse-grained orebodies and an increase of processing fine disseminated minerals and complex orebodies. These ores have provided new challenges in concentrator design because fine and ultra-fine grinding is required to achieve acceptable recovery rates. Therefore, the correct design of a grinding circuit is important for minimizing unit costs and increasing product quality. The use of ball mills for grinding in fine size ranges is inefficient and, therefore, vertical stirred grinding mills are becoming increasingly popular in the mineral processing industry due to its already known high energy efficiency. This work presents a hypothesis of a methodology to predict the product size distribution of a vertical stirred mill using a Bond ball mill. The Population Balance Model (PBM) was used to empirically analyze the performance of a vertical mill and a Bond ball mill. The breakage parameters obtained for both grinding mills are compared to determine the possibility of predicting the product size distribution of a vertical mill based on the results obtained from the Bond ball mill. The biggest advantage of this methodology is that most of the minerals processing laboratories already have a Bond ball mill to perform the tests suggested in this study. Preliminary results show the possibility of predicting the performance of a laboratory vertical stirred mill using a Bond ball mill.Keywords: bond ball mill, population balance model, product size distribution, vertical stirred mill
Procedia PDF Downloads 29231666 Interaction between Cognitive Control and Language Processing in Non-Fluent Aphasia
Authors: Izabella Szollosi, Klara Marton
Abstract:
Aphasia can be defined as a weakness in accessing linguistic information. Accessing linguistic information is strongly related to information processing, which in turn is associated with the cognitive control system. According to the literature, a deficit in the cognitive control system interferes with language processing and contributes to non-fluent speech performance. The aim of our study was to explore this hypothesis by investigating how cognitive control interacts with language performance in participants with non-fluent aphasia. Cognitive control is a complex construct that includes working memory (WM) and the ability to resist proactive interference (PI). Based on previous research, we hypothesized that impairments in domain-general (DG) cognitive control abilities have negative effects on language processing. In contrast, better DG cognitive control functioning supports goal-directed behavior in language-related processes as well. Since stroke itself might slow down information processing, it is important to examine its negative effects on both cognitive control and language processing. Participants (N=52) in our study were individuals with non-fluent Broca’s aphasia (N = 13), with transcortical motor aphasia (N=13), individuals with stroke damage without aphasia (N=13), and unimpaired speakers (N = 13). All participants performed various computer-based tasks targeting cognitive control functions such as WM and resistance to PI in both linguistic and non-linguistic domains. Non-linguistic tasks targeted primarily DG functions, while linguistic tasks targeted more domain specific (DS) processes. The results showed that participants with Broca’s aphasia differed from the other three groups in the non-linguistic tasks. They performed significantly worse even in the baseline conditions. In contrast, we found a different performance profile in the linguistic domain, where the control group differed from all three stroke-related groups. The three groups with impairment performed more poorly than the controls but similar to each other in the verbal baseline condition. In the more complex verbal PI condition, however, participants with Broca’s aphasia performed significantly worse than all the other groups. Participants with Broca’s aphasia demonstrated the most severe language impairment and the highest vulnerability in tasks measuring DG cognitive control functions. Results support the notion that the more severe the cognitive control impairment, the more severe the aphasia. Thus, our findings suggest a strong interaction between cognitive control and language. Individuals with the most severe and most general cognitive control deficit - participants with Broca’s aphasia - showed the most severe language impairment. Individuals with better DG cognitive control functions demonstrated better language performance. While all participants with stroke damage showed impaired cognitive control functions in the linguistic domain, participants with better language skills performed also better in tasks that measured non-linguistic cognitive control functions. The overall results indicate that the level of cognitive control deficit interacts with the language functions in individuals along with the language spectrum (from severe to no impairment). However, future research is needed to determine any directionality.Keywords: cognitive control, information processing, language performance, non-fluent aphasia
Procedia PDF Downloads 12331665 EGFR Signal Induced-Nuclear Translocation of Beta-catenin and PKM2 Promotes HCC Malignancy and Indicates Early Recurrence After Curative Resection
Authors: Fangtian Fan, Zhaoguo Liu, Yin Lu
Abstract:
Early recurrence (ER) (< 1 year) after liver resection is one of the most important factors that impacts the prognosis of patients with hepatocellular carcinoma (HCC). However, the molecular mechanisms and predictive indexes of ER after curative resection remain largely unknown. The present study aimed to exploit the role of EGFR signaling in EMT and early recurrence of HCC after curative resection and elucidate the molecular mechanisms. Our results showed that nuclear beta-catenin / PKM2 was a independent predictor of early recurrence after curative resection in EGFR-overexpressed HCC. Mechanistic investigation indicated that nuclear accumulation of beta-catenin and PKM2 induced by EGFR signal promoted HCC cell invasion and proliferation, which were required for early recurrence of HCC. These effects were mediated by PI3K/AKT and ERK pathways rather than the canonical Wnt signaling. In conclusions, EGFR signal induced-nuclear translocation of beta-catenin and PKM2 promotes HCC malignancy and indicates early recurrence after curative resection.Keywords: beta-catenin, early recurrence, hepatocellular carcinoma, malignancy, PKM2
Procedia PDF Downloads 35731664 Image Denoising Using Spatial Adaptive Mask Filter for Medical Images
Authors: R. Sumalatha, M. V. Subramanyam
Abstract:
In medical image processing the quality of the image is degraded in the presence of noise. Especially in ultra sound imaging and Magnetic resonance imaging the data was corrupted by signal dependent noise known as salt and pepper noise. Removal of noise from the medical images is a critical issue for researchers. In this paper, a new type of technique Adaptive Spatial Mask Filter (ASMF) has been proposed. The proposed filter is used to increase the quality of MRI and ultra sound images. Experimental results show that the proposed filter outperforms the implementation of mean, median, adaptive median filters in terms of MSE and PSNR.Keywords: salt and pepper noise, ASMF, PSNR, MSE
Procedia PDF Downloads 43631663 Thermomechanical Processing of a CuZnAl Shape-Memory Alloy
Authors: Pedro Henrique Alves Martins, Paulo Guilherme Ferreira De Siqueira, Franco De Castro Bubani, Maria Teresa Paulino Aguilar, Paulo Roberto Cetlin
Abstract:
Cu-base shape-memory alloys (CuZnAl, CuAlNi, CuAlBe, etc.) are promising engineering materials for several unconventional devices, such as sensors, actuators, and mechanical vibration dampers. Brittleness is one of the factors that limit the commercial use of these alloys, as it makes thermomechanical processing difficult. In this work, a method for the hot extrusion of a 75.50% Cu, 16,74% Zn, 7,76% Al (weight %) alloy is presented. The effects of the thermomechanical processing in the microstructure and the pseudoelastic behavior of the alloy are assessed by optical metallography, compression and hardness tests. Results show that hot extrusion is a suitable method to obtain severe cross-section reductions in the CuZnAl shape-memory alloy studied. The alloy maintained its pseudoelastic effect after the extrusion and the modifications in the mechanical behavior caused by precipitation during hot extrusion can be minimized by a suitable precipitate dissolution heat treatment.Keywords: hot extrusion, pseudoelastic, shape-memory alloy, thermomechanical processing
Procedia PDF Downloads 37431662 Application of a Synthetic DNA Reference Material for Optimisation of DNA Extraction and Purification for Molecular Identification of Medicinal Plants
Authors: Mina Kalantarzadeh, Claire Lockie-Williams, Caroline Howard
Abstract:
DNA barcoding is increasingly used for identification of medicinal plants worldwide. In the last decade, a large number of DNA barcodes have been generated, and their application in species identification explored. The success of DNA barcoding process relies on the accuracy of the results from polymerase chain reaction (PCR) amplification step which could be negatively affected due to a presence of inhibitors or degraded DNA in herbal samples. An established DNA reference material can be used to support molecular characterisation protocols and prove system suitability, for fast and accurate identification of plant species. The present study describes the use of a novel reference material, the trnH-psbA British Pharmacopoeia Nucleic Acid Reference Material (trnH-psbA BPNARM), which was produced to aid in the identification of Ocimum tenuiflorum L., a widely used herb. During DNA barcoding of O. tenuiflorum, PCR amplifications of isolated DNA produced inconsistent results, suggesting an issue with either the method or DNA quality of the tested samples. The trnH-psbA BPNARM was produced and tested to check for the issues caused during PCR amplification. It was added to the plant material as control DNA before extraction and was co-extracted and amplified by PCR. PCR analyses revealed that the amplification was not as successful as expected which suggested that the amplification is affected by presence of inhibitors co-extracted from plant materials. Various potential issues were assessed during DNA extraction and optimisations were made accordingly. A DNA barcoding protocol for O. tenuiflorum was published in the British Pharmacopoeia 2016, which included the reference sequence. The trnH-psbA BPNARM accelerated degradation test which investigates the stability of the reference material over time demonstrated that it has been stable when stored at 56 °C for a year. Using this protocol and trnH-psbA reference material provides a fast and accurate method for identification of O. tenuiflorum. The optimisations of the DNA extraction using the trnH-psbA BPNARM provided a signposting method which can assist in overcoming common problems encountered when using molecular methods with medicinal plants.Keywords: degradation, DNA extraction, nucleic acid reference material, trnH-psbA
Procedia PDF Downloads 19931661 RBF Neural Network Based Adaptive Robust Control for Bounded Position/Force Control of Bilateral Teleoperation Arms
Authors: Henni Mansour Abdelwaheb
Abstract:
This study discusses the design of a bounded position/force feedback controller developed to ensure position and force tracking for bilateral teleoperation arms operating with variable delay, and actuator saturation. Also, an adaptive robust Radial Basis Function (RBF) neural network is used to estimate the environment torque. The parameters of the environment torque are then sent from the slave site to the master site as a non-power signal to avoid passivity problems. Moreover, a nonlinear function is applied to each controller term as a smooth saturation function, providing a bounded control signal and preserving the system’s actuators. Lastly, the Lyapunov approach demonstrates the global stability of the controlled system, and numerical experiment results further confirm the validity of the presented strategy.Keywords: teleoperation manipulators system, time-varying delay, actuator saturation, adaptive robust rbf neural network approximation, uncertainties
Procedia PDF Downloads 7631660 Assessment of Barriers to the Clinical Adoption of Cell-Based Therapeutics
Authors: David Pettitt, Benjamin Davies, Georg Holländer, David Brindley
Abstract:
Cellular based therapies, whose origins can be traced from the intertwined concepts of tissue engineering and regenerative medicine, have the potential to transform the current medical landscape and offer an approach to managing what were once considered untreatable diseases. However, despite a large increase in basic science activity in the cell therapy arena alongside a growing portfolio of cell therapy trials, the number of industry products available for widespread clinical use correlates poorly with such a magnitude of activity, with the number of cell-based therapeutics in mainstream use remaining comparatively low. This research serves to quantitatively assess the barriers to the clinical adoption of cell-based therapeutics through identification of unique barriers, specific challenges and opportunities facing the development and adoption of such therapies.Keywords: cell therapy, clinical adoption, commercialization, translation
Procedia PDF Downloads 40031659 A Controlled Natural Language Assisted Approach for the Design and Automated Processing of Service Level Agreements
Authors: Christopher Schwarz, Katrin Riegler, Erwin Zinser
Abstract:
The management of outsourcing relationships between IT service providers and their customers proofs to be a critical issue that has to be stipulated by means of Service Level Agreements (SLAs). Since service requirements differ from customer to customer, SLA content and language structures vary largely, standardized SLA templates may not be used and an automated processing of SLA content is not possible. Hence, SLA management is usually a time-consuming and inefficient manual process. For overcoming these challenges, this paper presents an innovative and ITIL V3-conform approach for automated SLA design and management using controlled natural language in enterprise collaboration portals. The proposed novel concept is based on a self-developed controlled natural language that follows a subject-predicate-object approach to specify well-defined SLA content structures that act as templates for customized contracts and support automated SLA processing. The derived results eventually enable IT service providers to automate several SLA request, approval and negotiation processes by means of workflows and business rules within an enterprise collaboration portal. The illustrated prototypical realization gives evidence of the practical relevance in service-oriented scenarios as well as the high flexibility and adaptability of the presented model. Thus, the prototype enables the automated creation of well defined, customized SLA documents, providing a knowledge representation that is both human understandable and machine processable.Keywords: automated processing, controlled natural language, knowledge representation, information technology outsourcing, service level management
Procedia PDF Downloads 43231658 Big Data Analysis with RHadoop
Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim
Abstract:
It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop
Procedia PDF Downloads 43731657 High Speed Image Rotation Algorithm
Authors: Hee-Choul Kwon, Hyungjin Cho, Heeyong Kwon
Abstract:
Image rotation is one of main pre-processing step in image processing or image pattern recognition. It is implemented with rotation matrix multiplication. However it requires lots of floating point arithmetic operations and trigonometric function calculations, so it takes long execution time. We propose a new high speed image rotation algorithm without two major time-consuming operations. We compare the proposed algorithm with the conventional rotation one with various size images. Experimental results show that the proposed algorithm is superior to the conventional rotation ones.Keywords: high speed rotation operation, image processing, image rotation, pattern recognition, transformation matrix
Procedia PDF Downloads 50631656 Pre-Processing of Ultrasonography Image Quality Improvement in Cases of Cervical Cancer Using Image Enhancement
Authors: Retno Supriyanti, Teguh Budiono, Yogi Ramadhani, Haris B. Widodo, Arwita Mulyawati
Abstract:
Cervical cancer is the leading cause of mortality in cancer-related diseases. In this diagnosis doctors usually perform several tests to determine the presence of cervical cancer in a patient. However, these checks require support equipment to get the results in more detail. One is by using ultrasonography. However, for the developing countries most of the existing ultrasonography has a low resolution. The goal of this research is to obtain abnormalities on low-resolution ultrasound images especially for cervical cancer case. In this paper, we emphasize our work to use Image Enhancement for pre-processing image quality improvement. The result shows that pre-processing stage is promising to support further analysis.Keywords: cervical cancer, mortality, low-resolution, image enhancement.
Procedia PDF Downloads 63631655 A Comparative Analysis on QRS Peak Detection Using BIOPAC and MATLAB Software
Authors: Chandra Mukherjee
Abstract:
The present paper is a representation of the work done in the field of ECG signal analysis using MATLAB 7.1 Platform. An accurate and simple ECG feature extraction algorithm is presented in this paper and developed algorithm is validated using BIOPAC software. To detect the QRS peak, ECG signal is processed by following mentioned stages- First Derivative, Second Derivative and then squaring of that second derivative. Efficiency of developed algorithm is tested on ECG samples from different database and real time ECG signals acquired using BIOPAC system. Firstly we have lead wise specified threshold value the samples above that value is marked and in the original signal, where these marked samples face change of slope are spotted as R-peak. On the left and right side of the R-peak, faces change of slope identified as Q and S peak, respectively. Now the inbuilt Detection algorithm of BIOPAC software is performed on same output sample and both outputs are compared. ECG baseline modulation correction is done after detecting characteristics points. The efficiency of the algorithm is tested using some validation parameters like Sensitivity, Positive Predictivity and we got satisfied value of these parameters.Keywords: first derivative, variable threshold, slope reversal, baseline modulation correction
Procedia PDF Downloads 41131654 Speech Motor Processing and Animal Sound Communication
Authors: Ana Cleide Vieira Gomes Guimbal de Aquino
Abstract:
Sound communication is present in most vertebrates, from fish, mainly in species that live in murky waters, to some species of reptiles, anuran amphibians, birds, and mammals, including primates. There are, in fact, relevant similarities between human language and animal sound communication, and among these similarities are the vocalizations called calls. The first specific call in human babies is crying, which has a characteristic prosodic contour and is motivated most of the time by the need for food and by affecting the puppy-caregiver interaction, with a view to communicating the necessities and food requests and guaranteeing the survival of the species. The present work aims to articulate speech processing in the motor context with aspects of the project entitled emotional states and vocalization: a comparative study of the prosodic contours of crying in human and non-human animals. First, concepts of speech motor processing and general aspects of speech evolution will be presented to relate these two approaches to animal sound communication.Keywords: speech motor processing, animal communication, animal behaviour, language acquisition
Procedia PDF Downloads 8931653 Cardiokey: A Binary and Multi-Class Machine Learning Approach to Identify Individuals Using Electrocardiographic Signals on Wearable Devices
Authors: S. Chami, J. Chauvin, T. Demarest, Stan Ng, M. Straus, W. Jahner
Abstract:
Biometrics tools such as fingerprint and iris are widely used in industry to protect critical assets. However, their vulnerability and lack of robustness raise several worries about the protection of highly critical assets. Biometrics based on Electrocardiographic (ECG) signals is a robust identification tool. However, most of the state-of-the-art techniques have worked on clinical signals, which are of high quality and less noisy, extracted from wearable devices like a smartwatch. In this paper, we are presenting a complete machine learning pipeline that identifies people using ECG extracted from an off-person device. An off-person device is a wearable device that is not used in a medical context such as a smartwatch. In addition, one of the main challenges of ECG biometrics is the variability of the ECG of different persons and different situations. To solve this issue, we proposed two different approaches: per person classifier, and one-for-all classifier. The first approach suggests making binary classifier to distinguish one person from others. The second approach suggests a multi-classifier that distinguishes the selected set of individuals from non-selected individuals (others). The preliminary results, the binary classifier obtained a performance 90% in terms of accuracy within a balanced data. The second approach has reported a log loss of 0.05 as a multi-class score.Keywords: biometrics, electrocardiographic, machine learning, signals processing
Procedia PDF Downloads 14231652 A Comparative Study on Automatic Feature Classification Methods of Remote Sensing Images
Authors: Lee Jeong Min, Lee Mi Hee, Eo Yang Dam
Abstract:
Geospatial feature extraction is a very important issue in the remote sensing research. In the meantime, the image classification based on statistical techniques, but, in recent years, data mining and machine learning techniques for automated image processing technology is being applied to remote sensing it has focused on improved results generated possibility. In this study, artificial neural network and decision tree technique is applied to classify the high-resolution satellite images, as compared to the MLC processing result is a statistical technique and an analysis of the pros and cons between each of the techniques.Keywords: remote sensing, artificial neural network, decision tree, maximum likelihood classification
Procedia PDF Downloads 34731651 A Comparison of Inverse Simulation-Based Fault Detection in a Simple Robotic Rover with a Traditional Model-Based Method
Authors: Murray L. Ireland, Kevin J. Worrall, Rebecca Mackenzie, Thaleia Flessa, Euan McGookin, Douglas Thomson
Abstract:
Robotic rovers which are designed to work in extra-terrestrial environments present a unique challenge in terms of the reliability and availability of systems throughout the mission. Should some fault occur, with the nearest human potentially millions of kilometres away, detection and identification of the fault must be performed solely by the robot and its subsystems. Faults in the system sensors are relatively straightforward to detect, through the residuals produced by comparison of the system output with that of a simple model. However, faults in the input, that is, the actuators of the system, are harder to detect. A step change in the input signal, caused potentially by the loss of an actuator, can propagate through the system, resulting in complex residuals in multiple outputs. These residuals can be difficult to isolate or distinguish from residuals caused by environmental disturbances. While a more complex fault detection method or additional sensors could be used to solve these issues, an alternative is presented here. Using inverse simulation (InvSim), the inputs and outputs of the mathematical model of the rover system are reversed. Thus, for a desired trajectory, the corresponding actuator inputs are obtained. A step fault near the input then manifests itself as a step change in the residual between the system inputs and the input trajectory obtained through inverse simulation. This approach avoids the need for additional hardware on a mass- and power-critical system such as the rover. The InvSim fault detection method is applied to a simple four-wheeled rover in simulation. Additive system faults and an external disturbance force and are applied to the vehicle in turn, such that the dynamic response and sensor output of the rover are impacted. Basic model-based fault detection is then employed to provide output residuals which may be analysed to provide information on the fault/disturbance. InvSim-based fault detection is then employed, similarly providing input residuals which provide further information on the fault/disturbance. The input residuals are shown to provide clearer information on the location and magnitude of an input fault than the output residuals. Additionally, they can allow faults to be more clearly discriminated from environmental disturbances.Keywords: fault detection, ground robot, inverse simulation, rover
Procedia PDF Downloads 30831650 A New Distributed Computing Environment Based On Mobile Agents for Massively Parallel Applications
Authors: Fatéma Zahra Benchara, Mohamed Youssfi, Omar Bouattane, Hassan Ouajji, Mohamed Ouadi Bensalah
Abstract:
In this paper, we propose a new distributed environment for High Performance Computing (HPC) based on mobile agents. It allows us to perform parallel programs execution as distributed one over a flexible grid constituted by a cooperative mobile agent team works. The distributed program to be performed is encapsulated on team leader agent which deploys its team workers as Agent Virtual Processing Unit (AVPU). Each AVPU is asked to perform its assigned tasks and provides the computational results which make the data and team works tasks management difficult for the team leader agent and that influence the performance computing. In this work we focused on the implementation of the Mobile Provider Agent (MPA) in order to manage the distribution of data and instructions and to ensure a load balancing model. It grants also some interesting mechanisms to manage the others computing challenges thanks to the mobile agents several skills.Keywords: image processing, distributed environment, mobile agents, parallel and distributed computing
Procedia PDF Downloads 41031649 All-Optical Function Based on Self-Similar Spectral Broadening for 2R Regeneration in High-Bit-Rate Optical Transmission Systems
Authors: Leila Graini
Abstract:
In this paper, we demonstrate basic all-optical functions for 2R regeneration (Re-amplification and Re-shaping) based on self-similar spectral broadening in low normal dispersion and highly nonlinear fiber (ND-HNLF) to regenerate the signal through optical filtering including the transfer function characteristics, and output extinction ratio. Our approach of all-optical 2R regeneration is based on those of Mamyshev. The numerical study reveals the self-similar spectral broadening very effective for 2R all-optical regeneration; the proposed design presents high stability compared to a conventional regenerator using SPM broadening with reduction of the intensity fluctuations and improvement of the extinction ratio.Keywords: all-optical function, 2R optical regeneration, self-similar broadening, Mamyshev regenerator
Procedia PDF Downloads 18631648 A Review on Aluminium Metal Matric Composites
Authors: V. Singh, S. Singh, S. S. Garewal
Abstract:
Metal matrix composites with aluminum as the matrix material have been heralded as the next great development in advanced engineering materials. Aluminum metal matrix composites (AMMC) refer to the class of light weight high performance material systems. Properties of AMMCs can be tailored to the demands of different industrial applications by suitable combinations of matrix, reinforcement and processing route. AMMC finds its application in automotive, aerospace, defense, sports and structural areas. This paper presents an overview of AMMC material systems on aspects relating to processing, types and applications with case studies.Keywords: aluminum metal matrix composites, applications of aluminum metal matrix composites, lighting material processing of aluminum metal matrix composites
Procedia PDF Downloads 46531647 Competition between Verb-Based Implicit Causality and Theme Structure's Influence on Anaphora Bias in Mandarin Chinese Sentences: Evidence from Corpus
Authors: Linnan Zhang
Abstract:
Linguists, as well as psychologists, have shown great interests in implicit causality in reference processing. However, most frequently-used approaches to this issue are psychological experiments (such as eye tracking or self-paced reading, etc.). This research is a corpus-based one and is assisted with statistical tool – software R. The main focus of the present study is about the competition between verb-based implicit causality and theme structure’s influence on anaphora bias in Mandarin Chinese sentences. In Accessibility Theory, it is believed that salience, which is also known as accessibility, and relevance are two important factors in reference processing. Theme structure, which is a special syntactic structure in Chinese, determines the salience of an antecedent on the syntactic level while verb-based implicit causality is a key factor to the relevance between antecedent and anaphora. Therefore, it is a study about anaphora, combining psychology with linguistics. With analysis of the sentences from corpus as well as the statistical analysis of Multinomial Logistic Regression, major findings of the present study are as follows: 1. When the sentence is stated in a ‘cause-effect’ structure, the theme structure will always be the antecedent no matter forward biased verbs or backward biased verbs co-occur; in non-theme structure, the anaphora bias will tend to be the opposite of the verb bias; 2. When the sentence is stated in a ‘effect-cause’ structure, theme structure will not always be the antecedent and the influence of verb-based implicit causality will outweigh that of theme structure; moreover, the anaphora bias will be the same with the bias of verbs. All the results indicate that implicit causality functions conditionally and the noun in theme structure will not be the high-salience antecedent under any circumstances.Keywords: accessibility theory, anaphora, theme strcture, verb-based implicit causality
Procedia PDF Downloads 19831646 Graph-Oriented Summary for Optimized Resource Description Framework Graphs Streams Processing
Authors: Amadou Fall Dia, Maurras Ulbricht Togbe, Aliou Boly, Zakia Kazi Aoul, Elisabeth Metais
Abstract:
Existing RDF (Resource Description Framework) Stream Processing (RSP) systems allow continuous processing of RDF data issued from different application domains such as weather station measuring phenomena, geolocation, IoT applications, drinking water distribution management, and so on. However, processing window phase often expires before finishing the entire session and RSP systems immediately delete data streams after each processed window. Such mechanism does not allow optimized exploitation of the RDF data streams as the most relevant and pertinent information of the data is often not used in a due time and almost impossible to be exploited for further analyzes. It should be better to keep the most informative part of data within streams while minimizing the memory storage space. In this work, we propose an RDF graph summarization system based on an explicit and implicit expressed needs through three main approaches: (1) an approach for user queries (SPARQL) in order to extract their needs and group them into a more global query, (2) an extension of the closeness centrality measure issued from Social Network Analysis (SNA) to determine the most informative parts of the graph and (3) an RDF graph summarization technique combining extracted user query needs and the extended centrality measure. Experiments and evaluations show efficient results in terms of memory space storage and the most expected approximate query results on summarized graphs compared to the source ones.Keywords: centrality measures, RDF graphs summary, RDF graphs stream, SPARQL query
Procedia PDF Downloads 20331645 Quantification of E-Waste: A Case Study in Federal University of Espírito Santo, Brazil
Authors: Andressa S. T. Gomes, Luiza A. Souza, Luciana H. Yamane, Renato R. Siman
Abstract:
The segregation of waste of electrical and electronic equipment (WEEE) in the generating source, its characterization (quali-quantitative) and identification of origin, besides being integral parts of classification reports, are crucial steps to the success of its integrated management. The aim of this paper was to count WEEE generation at the Federal University of Espírito Santo (UFES), Brazil, as well as to define sources, temporary storage sites, main transportations routes and destinations, the most generated WEEE and its recycling potential. Quantification of WEEE generated at the University in the years between 2010 and 2015 was performed using data analysis provided by UFES’s sector of assets management. EEE and WEEE flow in the campuses information were obtained through questionnaires applied to the University workers. It was recorded 6028 WEEEs units of data processing equipment disposed by the university between 2010 and 2015. Among these waste, the most generated were CRT screens, desktops, keyboards and printers. Furthermore, it was observed that these WEEEs are temporarily stored in inappropriate places at the University campuses. In general, these WEEE units are donated to NGOs of the city, or sold through auctions (2010 and 2013). As for recycling potential, from the primary processing and further sale of printed circuit boards (PCB) from the computers, the amount collected could reach U$ 27,839.23. The results highlight the importance of a WEEE management policy at the University.Keywords: solid waste, waste of electrical and electronic equipment, waste management, institutional solid waste generation
Procedia PDF Downloads 26031644 Unified Public Transportation System for Mumbai Using Radio Frequency Identification
Authors: Saurabh Parkhedkar, Rajanikant Tenguria
Abstract:
The paper proposes revamping the public transportation system in Mumbai with the use of Radio Frequency Identification (RFID) technology in order to provide better integration and compatibility across various modes of transport. In Mumbai, mass transport system suffers from poor inter-compatible ticketing system, subpar money collection techniques, and lack of planning for optimum utilization of resources. Development of suburbs and growth in population will result in growing demand for mass transportation networks. Hence, the growing demand for the already overburdened public transportation system is only going to worsen the scenario. Thus, a superior system is essential in order to regulate, manage and supervise future transportation needs. The proposed RFID based system integrates Mumbai Suburban Railway, BEST (Brihanmumbai Electric Supply and Transport Undertaking transport wing) Bus, Mumbai Monorail and Mumbai Metro systems into a Unified Public Transportation System (UPTS). The UTPS takes into account various drawbacks of the present day system and offers solution, suitable for the modern age Mumbai.Keywords: urbanization, transportation, RFID, Mumbai, public transportation, smart city.
Procedia PDF Downloads 41231643 A Paper Based Sensor for Mercury Ion Detection
Authors: Emine G. Cansu Ergun
Abstract:
Conjugated system based sensors for selective detection of metal ions have been taking attention during last two decades. Fluorescent sensors are the promising candidates for ion detection due to their high selectivity towards metal ions, and rapid response times. Detection of mercury in an environmenet is important since mercury is a toxic element for human. Beyond the maximum allowable limit, mercury may cause serious problems in human health by spreading into the atmosphere, water and the food chain. In this study, a quinoxaline and 3,4-ethylenedioxy thiophene based donor-acceptor-donor type conjugated molecule used as a fluorescent sensor for detecting the mercury ion in aqueous medium. Among other various cations, existence of mercury resulted in a full quenching of the fluorescence signal. Then, a paper based sensor is constructed and used for mercury detection. As a result it is concluded that the offering sensor is a good candidate for selective mercury detection in aqueous media both in solution and paper based forms.Keywords: Conjugated molecules , fluorescence quenching, metal ion detection , sensors
Procedia PDF Downloads 15931642 Influence of Radio Frequency Identification Technology at Cost of Supply Chain as a Driver for the Generation of Competitive Advantage
Authors: Mona Baniahmadi, Saied Haghanifar
Abstract:
Radio Frequency Identification (RFID) is regarded as a promising technology for the optimization of supply chain processes since it improves manufacturing and retail operations from forecasting demand for planning, managing inventory, and distribution. This study precisely aims at learning to know the RFID technology and at explaining how it can concretely be used for supply chain management and how it can help improving it in the case of Hejrat Company which is located in Iran and works on the distribution of medical drugs and cosmetics. This study uses some statistical analysis to calculate the expected benefits of an integrated RFID system on supply chain obtained through competitive advantages increases with decreasing cost factor. The study investigates how the cost of storage process, labor cost, the cost of missing goods, inventory management optimization, on-time delivery, order cost, lost sales and supply process optimization affect the performance of the integrated RFID supply chain regarding cost factors and provides a competitive advantage.Keywords: cost, competitive advantage, radio frequency identification, supply chain
Procedia PDF Downloads 27631641 Effect of Traffic Composition on Delay and Saturation Flow at Signal Controlled Intersections
Authors: Arpita Saha, Apoorv Jain, Satish Chandra, Indrajit Ghosh
Abstract:
Level of service at a signal controlled intersection is directly measured from the delay. Similarly, saturation flow rate is a fundamental parameter to measure the intersection capacity. The present study calculates vehicle arrival rate, departure rate, and queue length for every five seconds interval in each cycle. Based on the queue lengths, the total delay of the cycle has been calculated using Simpson’s 1/3rd rule. Saturation flow has been estimated in terms of veh/hr of green/lane for every five seconds interval of the green period until at least three vehicles are left to cross the stop line. Vehicle composition shows an immense effect on total delay and saturation flow rate. The increase in two-wheeler proportion increases the saturation flow rate and reduces the total delay per vehicle significantly. Additionally, an increase in the heavy vehicle proportion reduces the saturation flow rate and increases the total delay for each vehicle.Keywords: delay, saturation flow, signalised intersection, vehicle composition
Procedia PDF Downloads 464