Search results for: noun based filtering
28151 Designing a Patient Monitoring System Using Cloud and Semantic Web Technologies
Authors: Chryssa Thermolia, Ekaterini S. Bei, Stelios Sotiriadis, Kostas Stravoskoufos, Euripides G. M. Petrakis
Abstract:
Moving into a new era of healthcare, new tools and devices are developed to extend and improve health services, such as remote patient monitoring and risk prevention. In this concept, Internet of Things (IoT) and Cloud Computing present great advantages by providing remote and efficient services, as well as cooperation between patients, clinicians, researchers and other health professionals. This paper focuses on patients suffering from bipolar disorder, a brain disorder that belongs to a group of conditions called effective disorders, which is characterized by great mood swings.We exploit the advantages of Semantic Web and Cloud Technologies to develop a patient monitoring system to support clinicians. Based on intelligently filtering of evidence-knowledge and individual-specific information we aim to provide treatment notifications and recommended function tests at appropriate times or concluding into alerts for serious mood changes and patient’s non-response to treatment. We propose an architecture, as the back-end part of a cloud platform for IoT, intertwining intelligence devices with patients’ daily routine and clinicians’ support.Keywords: bipolar disorder, intelligent systems patient monitoring, semantic web technologies, healthcare
Procedia PDF Downloads 50828150 Raman Scattering Broadband Spectrum Generation in Compact Yb-Doped Fiber Laser
Authors: Yanrong Song, Zikai Dong, Runqin Xu, Jinrong Tian, Kexuan Li
Abstract:
Nonlinear polarization rotation (NPR) technique has become one of the main techniques to achieve mode-locked fiber lasers for its compactness, implementation, and low cost. In this paper, we demonstrate a compact mode-locked Yb-doped fiber laser based on NPR technique in the all normal dispersion (ANDi) regime. In the laser cavity, there are no physical filter and polarization controller in laser cavity. Mode-locked pulse train is achieved in ANDi regime based on NPR technique. The fiber birefringence induced filtering effect is the mainly reason for mode-locking. After that, an extra 20 m long single-mode fiber is inserted in two different positions, dissipative soliton operation and noise like pulse operations are achieved correspondingly. The nonlinear effect is obviously enhanced in the noise like pulse regime and broadband spectrum generated owing to enhanced stimulated Raman scattering effect. When the pump power is 210 mW, the central wavelength is 1030 nm, and the corresponding 1st order Raman scattering stokes wave generates and locates at 1075 nm. When the pump power is 370 mW, the 1st and 2nd order Raman scattering stokes wave generate and locate at 1080 nm, 1126 nm respectively. When the pump power is 600 mW, the Raman continuum is generated with cascaded multi-order stokes waves, and the spectrum extends to 1188 nm. The total flat spectrum is from 1000nm to 1200nm. The maximum output average power and pulse energy are 18.0W and 14.75nJ, respectively.Keywords: fiber laser, mode-locking, nonlinear polarization rotation, Raman scattering
Procedia PDF Downloads 22128149 Design and Implementation of a Software Platform Based on Artificial Intelligence for Product Recommendation
Authors: Giuseppina Settanni, Antonio Panarese, Raffaele Vaira, Maurizio Galiano
Abstract:
Nowdays, artificial intelligence is used successfully in academia and industry for its ability to learn from a large amount of data. In particular, in recent years the use of machine learning algorithms in the field of e-commerce has spread worldwide. In this research study, a prototype software platform was designed and implemented in order to suggest to users the most suitable products for their needs. The platform includes a chatbot and a recommender system based on artificial intelligence algorithms that provide suggestions and decision support to the customer. The recommendation systems perform the important function of automatically filtering and personalizing information, thus allowing to manage with the IT overload to which the user is exposed on a daily basis. Recently, international research has experimented with the use of machine learning technologies with the aim to increase the potential of traditional recommendation systems. Specifically, support vector machine algorithms have been implemented combined with natural language processing techniques that allow the user to interact with the system, express their requests and receive suggestions. The interested user can access the web platform on the internet using a computer, tablet or mobile phone, register, provide the necessary information and view the products that the system deems them most appropriate. The platform also integrates a dashboard that allows the use of the various functions, which the platform is equipped with, in an intuitive and simple way. Artificial intelligence algorithms have been implemented and trained on historical data collected from user browsing. Finally, the testing phase allowed to validate the implemented model, which will be further tested by letting customers use it.Keywords: machine learning, recommender system, software platform, support vector machine
Procedia PDF Downloads 13428148 Correction of Frequent English Writing Errors by Using Coded Indirect Corrective Feedback and Error Treatment: The Case of Reading and Writing English for Academic Purposes II
Authors: Chaiwat Tantarangsee
Abstract:
The purposes of this study are 1) to study the frequent English writing errors of students registering the course: Reading and Writing English for Academic Purposes II, and 2) to find out the results of writing error correction by using coded indirect corrective feedback and writing error treatments. Samples include 28 2nd year English Major students, Faculty of Education, Suan Sunandha Rajabhat University. Tool for experimental study includes the lesson plan of the course; Reading and Writing English for Academic Purposes II, and tool for data collection includes 4 writing tests of short texts. The research findings disclose that frequent English writing errors found in this course comprise 7 types of grammatical errors, namely Fragment sentence, Subject-verb agreement, Wrong form of verb tense, Singular or plural noun endings, Run-ons sentence, Wrong form of verb pattern and Lack of parallel structure. Moreover, it is found that the results of writing error correction by using coded indirect corrective feedback and error treatment reveal the overall reduction of the frequent English writing errors and the increase of students’ achievement in the writing of short texts with the significance at .05.Keywords: coded indirect corrective feedback, error correction, error treatment, English writing
Procedia PDF Downloads 30628147 A Finite Memory Residual Generation Filter for Fault Detection
Authors: Pyung Soo Kim, Eung Hyuk Lee, Mun Suck Jang
Abstract:
In the current paper, a residual generation filter with finite memory structure is proposed for fault detection. The proposed finite memory residual generation filter provides the residual by real-time filtering of fault vector using only the most recent finite observations and inputs on the window. It is shown that the residual given by the proposed residual generation filter provides the exact fault for noise-free systems. Finally, to illustrate the capability of the proposed residual generation filter, numerical examples are performed for the discretized DC motor system having the multiple sensor faults.Keywords: residual generation filter, finite memory structure, kalman filter, fast detection
Procedia PDF Downloads 69828146 A Single Feature Probability-Object Based Image Analysis for Assessing Urban Landcover Change: A Case Study of Muscat Governorate in Oman
Authors: Salim H. Al Salmani, Kevin Tansey, Mohammed S. Ozigis
Abstract:
The study of the growth of built-up areas and settlement expansion is a major exercise that city managers seek to undertake to establish previous and current developmental trends. This is to ensure that there is an equal match of settlement expansion needs to the appropriate levels of services and infrastructure required. This research aims at demonstrating the potential of satellite image processing technique, harnessing the utility of single feature probability-object based image analysis technique in assessing the urban growth dynamics of the Muscat Governorate in Oman for the period 1990, 2002 and 2013. This need is fueled by the continuous expansion of the Muscat Governorate beyond predicted levels of infrastructural provision. Landsat Images of the years 1990, 2002 and 2013 were downloaded and preprocessed to forestall appropriate radiometric and geometric standards. A novel approach of probability filtering of the target feature segment was implemented to derive the spatial extent of the final Built-Up Area of the Muscat governorate for the three years period. This however proved to be a useful technique as high accuracy assessment results of 55%, 70%, and 71% were recorded for the Urban Landcover of 1990, 2002 and 2013 respectively. Furthermore, the Normalized Differential Built – Up Index for the various images were derived and used to consolidate the results of the SFP-OBIA through a linear regression model and visual comparison. The result obtained showed various hotspots where urbanization have sporadically taken place. Specifically, settlement in the districts (Wilayat) of AL-Amarat, Muscat, and Qurayyat experienced tremendous change between 1990 and 2002, while the districts (Wilayat) of AL-Seeb, Bawshar, and Muttrah experienced more sporadic changes between 2002 and 2013.Keywords: urban growth, single feature probability, object based image analysis, landcover change
Procedia PDF Downloads 27428145 Real-Time Radar Tracking Based on Nonlinear Kalman Filter
Authors: Milca F. Coelho, K. Bousson, Kawser Ahmed
Abstract:
To accurately track an aerospace vehicle in a time-critical situation and in a highly nonlinear environment, is one of the strongest interests within the aerospace community. The tracking is achieved by estimating accurately the state of a moving target, which is composed of a set of variables that can provide a complete status of the system at a given time. One of the main ingredients for a good estimation performance is the use of efficient estimation algorithms. A well-known framework is the Kalman filtering methods, designed for prediction and estimation problems. The success of the Kalman Filter (KF) in engineering applications is mostly due to the Extended Kalman Filter (EKF), which is based on local linearization. Besides its popularity, the EKF presents several limitations. To address these limitations and as a possible solution to tracking problems, this paper proposes the use of the Ensemble Kalman Filter (EnKF). Although the EnKF is being extensively used in the context of weather forecasting and it is being recognized for producing accurate and computationally effective estimation on systems with a very high dimension, it is almost unknown by the tracking community. The EnKF was initially proposed as an attempt to improve the error covariance calculation, which on the classic Kalman Filter is difficult to implement. Also, in the EnKF method the prediction and analysis error covariances have ensemble representations. These ensembles have sizes which limit the number of degrees of freedom, in a way that the filter error covariance calculations are a lot more practical for modest ensemble sizes. In this paper, a realistic simulation of a radar tracking was performed, where the EnKF was applied and compared with the Extended Kalman Filter. The results suggested that the EnKF is a promising tool for tracking applications, offering more advantages in terms of performance.Keywords: Kalman filter, nonlinear state estimation, optimal tracking, stochastic environment
Procedia PDF Downloads 14628144 The Trigger-DAQ System in the Mu2e Experiment
Authors: Antonio Gioiosa, Simone Doanti, Eric Flumerfelt, Luca Morescalchi, Elena Pedreschi, Gianantonio Pezzullo, Ryan A. Rivera, Franco Spinella
Abstract:
The Mu2e experiment at Fermilab aims to measure the charged-lepton flavour violating neutrino-less conversion of a negative muon into an electron in the field of an aluminum nucleus. With the expected experimental sensitivity, Mu2e will improve the previous limit of four orders of magnitude. The Mu2e data acquisition (DAQ) system provides hardware and software to collect digitized data from the tracker, calorimeter, cosmic ray veto, and beam monitoring systems. Mu2e’s trigger and data acquisition system (TDAQ) uses otsdaq as its solution. developed at Fermilab, otsdaq uses the artdaq DAQ framework and art analysis framework, under-the-hood, for event transfer, filtering, and processing. Otsdaq is an online DAQ software suite with a focus on flexibility and scalability while providing a multi-user, web-based interface accessible through the Chrome or Firefox web browser. The detector read out controller (ROC) from the tracker and calorimeter stream out zero-suppressed data continuously to the data transfer controller (DTC). Data is then read over the PCIe bus to a software filter algorithm that selects events which are finally combined with the data flux that comes from a cosmic ray veto system (CRV).Keywords: trigger, daq, mu2e, Fermilab
Procedia PDF Downloads 15528143 Lab Bench for Synthetic Aperture Radar Imaging System
Authors: Karthiyayini Nagarajan, P. V. Ramakrishna
Abstract:
Radar Imaging techniques provides extensive applications in the field of remote sensing, majorly Synthetic Aperture Radar (SAR) that provide high resolution target images. This paper work puts forward the effective and realizable signal generation and processing for SAR images. The major units in the system include camera, signal generation unit, signal processing unit and display screen. The real radio channel is replaced by its mathematical model based on optical image to calculate a reflected signal model in real time. Signal generation realizes the algorithm and forms the radar reflection model. Signal processing unit provides range and azimuth resolution through matched filtering and spectrum analysis procedure to form radar image on the display screen. The restored image has the same quality as that of the optical image. This SAR imaging system has been designed and implemented using MATLAB and Quartus II tools on Stratix III device as a System (Lab Bench) that works in real time to study/investigate on radar imaging rudiments and signal processing scheme for educational and research purposes.Keywords: synthetic aperture radar, radio reflection model, lab bench, imaging engineering
Procedia PDF Downloads 49728142 Design and Implementation of a Lab Bench for Synthetic Aperture Radar Imaging System
Authors: Karthiyayini Nagarajan, P. V. RamaKrishna
Abstract:
Radar Imaging techniques provides extensive applications in the field of remote sensing, majorly Synthetic Aperture Radar(SAR) that provide high resolution target images. This paper work puts forward the effective and realizable signal generation and processing for SAR images. The major units in the system include camera, signal generation unit, signal processing unit and display screen. The real radio channel is replaced by its mathematical model based on optical image to calculate a reflected signal model in real time. Signal generation realizes the algorithm and forms the radar reflection model. Signal processing unit provides range and azimuth resolution through matched filtering and spectrum analysis procedure to form radar image on the display screen. The restored image has the same quality as that of the optical image. This SAR imaging system has been designed and implemented using MATLAB and Quartus II tools on Stratix III device as a System(lab bench) that works in real time to study/investigate on radar imaging rudiments and signal processing scheme for educational and research purposes.Keywords: synthetic aperture radar, radio reflection model, lab bench
Procedia PDF Downloads 46828141 Measuring and Evaluating the Effectiveness of Mobile High Efficiency Particulate Air Filtering on Particulate Matter within the Road Traffic Network of a Sample of Non-Sparse and Sparse Urban Environments in the UK
Authors: Richard Maguire
Abstract:
This research evaluates the efficiency of using mobile HEPA filters to reduce localized Particulate Matter (PM), Total Volatile Organic Chemical (TVOC) and Formaldehyde (HCHO) Air Pollution. The research is being performed using a standard HEPA filter that is tube fitted and attached to a motor vehicle. The velocity of the vehicle is used to generate the pressure difference that allows the filter to remove PM, VOC and HCOC pollution from the localized atmosphere of a road transport traffic route. The testing has been performed on a sample of traffic routes in Non-Sparse and Sparse urban environments within the UK. Pre and Post filter measuring of the PM2.5 Air Quality has been carried out along with demographics of the climate environment, including live filming of the traffic conditions. This provides a base line for future national and international research. The effectiveness measurement is generated through evaluating the difference in PM2.5 Air Quality measured pre- and post- the mobile filter test equipment. A series of further research opportunities and future exploitation options are made based on the results of the research.Keywords: high efficiency particulate air, HEPA filter, particulate matter, traffic pollution
Procedia PDF Downloads 12328140 Modeling of a UAV Longitudinal Dynamics through System Identification Technique
Authors: Asadullah I. Qazi, Mansoor Ahsan, Zahir Ashraf, Uzair Ahmad
Abstract:
System identification of an Unmanned Aerial Vehicle (UAV), to acquire its mathematical model, is a significant step in the process of aircraft flight automation. The need for reliable mathematical model is an established requirement for autopilot design, flight simulator development, aircraft performance appraisal, analysis of aircraft modifications, preflight testing of prototype aircraft and investigation of fatigue life and stress distribution etc. This research is aimed at system identification of a fixed wing UAV by means of specifically designed flight experiment. The purposely designed flight maneuvers were performed on the UAV and aircraft states were recorded during these flights. Acquired data were preprocessed for noise filtering and bias removal followed by parameter estimation of longitudinal dynamics transfer functions using MATLAB system identification toolbox. Black box identification based transfer function models, in response to elevator and throttle inputs, were estimated using least square error technique. The identification results show a high confidence level and goodness of fit between the estimated model and actual aircraft response.Keywords: fixed wing UAV, system identification, black box modeling, longitudinal dynamics, least square error
Procedia PDF Downloads 32528139 Video Text Information Detection and Localization in Lecture Videos Using Moments
Authors: Belkacem Soundes, Guezouli Larbi
Abstract:
This paper presents a robust and accurate method for text detection and localization over lecture videos. Frame regions are classified into text or background based on visual feature analysis. However, lecture video shows significant degradation mainly related to acquisition conditions, camera motion and environmental changes resulting in low quality videos. Hence, affecting feature extraction and description efficiency. Moreover, traditional text detection methods cannot be directly applied to lecture videos. Therefore, robust feature extraction methods dedicated to this specific video genre are required for robust and accurate text detection and extraction. Method consists of a three-step process: Slide region detection and segmentation; Feature extraction and non-text filtering. For robust and effective features extraction moment functions are used. Two distinct types of moments are used: orthogonal and non-orthogonal. For orthogonal Zernike Moments, both Pseudo Zernike moments are used, whereas for non-orthogonal ones Hu moments are used. Expressivity and description efficiency are given and discussed. Proposed approach shows that in general, orthogonal moments show high accuracy in comparison to the non-orthogonal one. Pseudo Zernike moments are more effective than Zernike with better computation time.Keywords: text detection, text localization, lecture videos, pseudo zernike moments
Procedia PDF Downloads 15128138 A Microwave and Millimeter-Wave Transmit/Receive Switch Subsystem for Communication Systems
Authors: Donghyun Lee, Cam Nguyen
Abstract:
Multi-band systems offer a great deal of benefit in modern communication and radar systems. In particular, multi-band antenna-array radar systems with their extended frequency diversity provide numerous advantages in detection, identification, locating and tracking a wide range of targets, including enhanced detection coverage, accurate target location, reduced survey time and cost, increased resolution, improved reliability and target information. An accurate calibration is a critical issue in antenna array systems. The amplitude and phase errors in multi-band and multi-polarization antenna array transceivers result in inaccurate target detection, deteriorated resolution and reduced reliability. Furthermore, the digital beam former without the RF domain phase-shifting is less immune to unfiltered interference signals, which can lead to receiver saturation in array systems. Therefore, implementing integrated front-end architecture, which can support calibration function with low insertion and filtering function from the farthest end of an array transceiver is of great interest. We report a dual K/Ka-band T/R/Calibration switch module with quasi-elliptic dual-bandpass filtering function implementing a Q-enhanced metamaterial transmission line. A unique dual-band frequency response is incorporated in the reception and calibration path of the proposed switch module utilizing the composite right/left-handed meta material transmission line coupled with a Colpitts-style negative generation circuit. The fabricated fully integrated T/R/Calibration switch module in 0.18-μm BiCMOS technology exhibits insertion loss of 4.9-12.3 dB and isolation of more than 45 dB in the reception, transmission and calibration mode of operation. In the reception and calibration mode, the dual-band frequency response centered at 24.5 and 35 GHz exhibits out-of-band rejection of more than 30 dB compared to the pass bands below 10.5 GHz and above 59.5 GHz. The rejection between the pass bands reaches more than 50 dB. In all modes of operation, the IP1-dB is between 4 and 11 dBm. Acknowledgement: This paper was made possible by NPRP grant # 6-241-2-102 from the Qatar National Research Fund (a member of Qatar Foundation). The statements made herein are solely the responsibility of the authors.Keywords: microwaves, millimeter waves, T/R switch, wireless communications, wireless communications
Procedia PDF Downloads 16028137 Threshold (K, P) Quantum Distillation
Authors: Shashank Gupta, Carlos Cid, William John Munro
Abstract:
Quantum distillation is the task of concentrating quantum correlations present in N imperfect copies to M perfect copies (M < N) using free operations by involving all P the parties sharing the quantum correlation. We present a threshold quantum distillation task where the same objective is achieved but using lesser number of parties (K < P). In particular, we give an exact local filtering operations by the participating parties sharing high dimension multipartite entangled state to distill the perfect quantum correlation. Later, we bridge a connection between threshold quantum entanglement distillation and quantum steering distillation and show that threshold distillation might work in the scenario where general distillation protocol like DEJMPS does not work.Keywords: quantum networks, quantum distillation, quantum key distribution, entanglement distillation
Procedia PDF Downloads 4528136 Heliport Remote Safeguard System Based on Real-Time Stereovision 3D Reconstruction Algorithm
Authors: Ł. Morawiński, C. Jasiński, M. Jurkiewicz, S. Bou Habib, M. Bondyra
Abstract:
With the development of optics, electronics, and computers, vision systems are increasingly used in various areas of life, science, and industry. Vision systems have a huge number of applications. They can be used in quality control, object detection, data reading, e.g., QR-code, etc. A large part of them is used for measurement purposes. Some of them make it possible to obtain a 3D reconstruction of the tested objects or measurement areas. 3D reconstruction algorithms are mostly based on creating depth maps from data that can be acquired from active or passive methods. Due to the specific appliance in airfield technology, only passive methods are applicable because of other existing systems working on the site, which can be blinded on most spectral levels. Furthermore, reconstruction is required to work long distances ranging from hundreds of meters to tens of kilometers with low loss of accuracy even with harsh conditions such as fog, rain, or snow. In response to those requirements, HRESS (Heliport REmote Safeguard System) was developed; which main part is a rotational head with a two-camera stereovision rig gathering images around the head in 360 degrees along with stereovision 3D reconstruction and point cloud combination. The sub-pixel analysis introduced in the HRESS system makes it possible to obtain an increased distance measurement resolution and accuracy of about 3% for distances over one kilometer. Ultimately, this leads to more accurate and reliable measurement data in the form of a point cloud. Moreover, the program algorithm introduces operations enabling the filtering of erroneously collected data in the point cloud. All activities from the programming, mechanical and optical side are aimed at obtaining the most accurate 3D reconstruction of the environment in the measurement area.Keywords: airfield monitoring, artificial intelligence, stereovision, 3D reconstruction
Procedia PDF Downloads 12428135 REDUCER: An Architectural Design Pattern for Reducing Large and Noisy Data Sets
Authors: Apkar Salatian
Abstract:
To relieve the burden of reasoning on a point to point basis, in many domains there is a need to reduce large and noisy data sets into trends for qualitative reasoning. In this paper we propose and describe a new architectural design pattern called REDUCER for reducing large and noisy data sets that can be tailored for particular situations. REDUCER consists of 2 consecutive processes: Filter which takes the original data and removes outliers, inconsistencies or noise; and Compression which takes the filtered data and derives trends in the data. In this seminal article, we also show how REDUCER has successfully been applied to 3 different case studies.Keywords: design pattern, filtering, compression, architectural design
Procedia PDF Downloads 21228134 The Voiceless Dental- Alveolar Common Augment in Arabic and Other Semitic Languages, a Morphophonemic Comparison
Authors: Tarek Soliman Mostafa Soliman Al-Nana'i
Abstract:
There are non-steady voiced augments in the Semitic languages, and in the morphological and structural augmentation, two sounds were augments in all Semitic languages at the level of the spoken language and two letters at the level of the written language, which are the hamza and the ta’. This research studies only the second of them; Therefore, we defined it as “The Voiceless Dental- alveolar common augment” (VDACA) to distinguish it from the glottal sound “Hamza”, first, middle, or last, in a noun or in a verb, in Arabic and its equivalent in the Semitic languages. What is meant by “VDACA” is the ta’ that is in addition to the root of the word at the morphological level: the word “voiceless” takes out the voiced sounds that we studied before, and the “dental- alveolar common augment” takes out the laryngeal sound of them, which is the “Hamza”: and the word “common” brings out the uncommon voiceless sounds, which are sīn, shīn, and hā’. The study is limited to the ta' alone among the Arabic sounds, and this title faced a problem in identifying it with the ta'. Because the designation of the ta is not the same in most Semitic languages. Hebrew, for example, has “tav” and is pronounced with the voiced fa (v), which is not in Arabic. It is called different names in other Semitic languages, such as “taw” or “tAu” in old Syriac. And so on. This goes hand in hand with the insistence on distance from the written level and the reference to the phonetic aspect in this study that is closely and closely linked to the morphological level. Therefore, the study is “morphophonemic”. What is meant by Semitic languages in this study are the following: Akkadian, Ugaritic, Hebrew, Syriac, Mandaean, Ge'ez, and Amharic. The problem of the study is the agreement or difference between these languages in the position of that augment, first, middle, or last. And in determining the distinguishing characteristics of each language from the other. As for the study methodology, it is determined by the comparative approach in Semitic languages, which is based on the descriptive approach for each language. The study is divided into an introduction, four sections, and a conclusion: Introduction: It included the subject of the study, its importance, motives, problem, methodology, and division. The first section: VDACA as a non-common phoneme. The second: VDACA as a common phoneme. The third: VDACA as a functional morpheme. The fourth section: Commentary and conclusion with the most important results. The positions of VDACA in Arabic and other Semitic languages, and in nouns and verbs, were limited to first, middle, and last. The research identified the individual addition, which is common with other augments, and the research proved that this augmentation is constant in all Semitic languages, but there are characteristics that distinguish each language from the other.Keywords: voiceless -, dental- alveolar, augment, Arabic - semitic languages
Procedia PDF Downloads 7328133 The Research of the Game Interface Improvement Due to the Game Operation Dilemma of Player in the Side-Scrolling Shooting Game
Authors: Shih-Chieh Liao, Cheng-Yan Shuai
Abstract:
The feature of a side-scrolling shooting game is facing the surrounding enemy and barraging in entire screen. The player will be in trouble when they are trying to do complicated operations because of the physical and system limitations of the joystick in the games. This study designed the prototype of a new type of arcade stick by focus group and assessed by the expert. By filtering the most representative, and build up the control system for the arcade stick, and testing time and bullets consumed in two experiments, try to prove it works in the game. Finally, the prototype of L-1 solves the dilemma of scroll shooting games when the player uses the arcade stick and improves the function of the arcade stick.Keywords: arcade stick, joystick, user interface, 2D STG
Procedia PDF Downloads 8028132 The Role of Specificity in Mastering the English Article System
Authors: Sugene Kim
Abstract:
The English articles are taught as a binary system based on nominal countability and definiteness. Despite the detailed rules of prescriptive grammar, it has been consistently reported in the literature that their correct usage is extremely difficult to master even for advanced learners of English as a second language (ESL) or a foreign language (EFL). Given that an English sentence (except for an imperative) cannot be constructed without a noun, which is always paired with one of the indefinite, definite, and zero articles; it is essential to understand specifically what causes ESL/EFL learners to misuse them. To that end, this study examined EFL learners’ article use employing a one-group pre–post-test design. Forty-three Korean college students received instruction on correct English article usage for two 75-minute classes employing the binary schema set up for the study. They also practiced in class how to apply the rules as instructed. Then, the participants were assigned a forced-choice elicitation task, which was also used as a pre-test administered three months prior to the instruction. Unlike the pre-test on which they only chose the correct article for each of the 40 items, the post-instruction task additionally asked them to give written accounts of their decision-making procedure to choose the article as they did. The participants’ performance was scored manually by checking whether the answer given is correct or incorrect, and their written comments were first categorized using thematic analysis and then ranked by frequency. The analyses of the performance on the two tasks and the written think-aloud data suggested that EFL learners exhibit fluctuation between specificity and definiteness, overgeneralizing the use of the definite article for almost all cataphoric references. It was apparent that they have trouble distinguishing from the two concepts possibly because the former is almost never introduced in the grammar books or classes designed for ESL/EFL learners. Particularly, most participants were found to be ignorant of the possibility of using nouns as [+specific, –definite]. Not surprisingly, the correct answer rates for such nouns averaged out at 33% and 46% on the pre- and post-tests, respectively, which narrowly reach half the overall mean correct answer rates of 65% on the pre-test and 81% on the post-test. In addition, correct article use for specific indefinites was most impermeable to instruction when compared with nouns used as [–specific, –definite] or [± specific, +definite]. Such findings underline the necessity for expanding the binary schema to a ternary form that incorporates the specificity feature, albeit not morphologically marked in the English language.Keywords: countability, definiteness, English articles, specificity, ternary system
Procedia PDF Downloads 12428131 Improve Divers Tracking and Classification in Sonar Images Using Robust Diver Wake Detection Algorithm
Authors: Mohammad Tarek Al Muallim, Ozhan Duzenli, Ceyhun Ilguy
Abstract:
Harbor protection systems are so important. The need for automatic protection systems has increased over the last years. Diver detection active sonar has great significance. It used to detect underwater threats such as divers and autonomous underwater vehicle. To automatically detect such threats the sonar image is processed by algorithms. These algorithms used to detect, track and classify of underwater objects. In this work, divers tracking and classification algorithm is improved be proposing a robust wake detection method. To detect objects the sonar images is normalized then segmented based on fixed threshold. Next, the centroids of the segments are found and clustered based on distance metric. Then to track the objects linear Kalman filter is applied. To reduce effect of noise and creation of false tracks, the Kalman tracker is fine tuned. The tuning is done based on our active sonar specifications. After the tracks are initialed and updated they are subjected to a filtering stage to eliminate the noisy and unstable tracks. Also to eliminate object with a speed out of the diver speed range such as buoys and fast boats. Afterwards the result tracks are subjected to a classification stage to deiced the type of the object been tracked. Here the classification stage is to deice wither if the tracked object is an open circuit diver or a close circuit diver. At the classification stage, a small area around the object is extracted and a novel wake detection method is applied. The morphological features of the object with his wake is extracted. We used support vector machine to find the best classifier. The sonar training images and the test images are collected by ARMELSAN Defense Technologies Company using the portable diver detection sonar ARAS-2023. After applying the algorithm to the test sonar data, we get fine and stable tracks of the divers. The total classification accuracy achieved with the diver type is 97%.Keywords: harbor protection, diver detection, active sonar, wake detection, diver classification
Procedia PDF Downloads 23828130 Familial Exome Sequencing to Decipher the Complex Genetic Basis of Holoprosencephaly
Authors: Artem Kim, Clara Savary, Christele Dubourg, Wilfrid Carre, Houda Hamdi-Roze, Valerie Dupé, Sylvie Odent, Marie De Tayrac, Veronique David
Abstract:
Holoprosencephaly (HPE) is a rare congenital brain malformation resulting from the incomplete separation of the two cerebral hemispheres. It is characterized by a wide phenotypic spectrum and a high degree of locus heterogeneity. Genetic defects in 16 genes have already been implicated in HPE, but account for only 30% of cases, suggesting that a large part of genetic factors remains to be discovered. HPE has been recently redefined as a complex multigenic disorder, requiring the joint effect of multiple mutational events in genes belonging to one or several developmental pathways. The onset of HPE may result from accumulation of the effects of multiple rare variants in functionally-related genes, each conferring a moderate increase in the risk of HPE onset. In order to decipher the genetic basis of HPE, unconventional patterns of inheritance involving multiple genetic factors need to be considered. The primary objective of this study was to uncover possible disease causing combinations of multiple rare variants underlying HPE by performing trio-based Whole Exome Sequencing (WES) of familial cases where no molecular diagnosis could be established. 39 families were selected with no fully-penetrant causal mutation in known HPE gene, no chromosomic aberrations/copy number variants and without any implication of environmental factors. As the main challenge was to identify disease-related variants among a large number of nonpathogenic polymorphisms detected by WES classical scheme, a novel variant prioritization approach was established. It combined WES filtering with complementary gene-level approaches: transcriptome-driven (RNA-Seq data) and clinically-driven (public clinical data) strategies. Briefly, a filtering approach was performed to select variants compatible with disease segregation, population frequency and pathogenicity prediction to identify an exhaustive list of rare deleterious variants. The exome search space was then reduced by restricting the analysis to candidate genes identified by either transcriptome-driven strategy (genes sharing highly similar expression patterns with known HPE genes during cerebral development) or clinically-driven strategy (genes associated to phenotypes of interest overlapping with HPE). Deeper analyses of candidate variants were then performed on a family-by-family basis. These included the exploration of clinical information, expression studies, variant characteristics, recurrence of mutated genes and available biological knowledge. A novel bioinformatics pipeline was designed. Applied to the 39 families, this final integrated workflow identified an average of 11 candidate variants per family. Most of candidate variants were inherited from asymptomatic parents suggesting a multigenic inheritance pattern requiring the association of multiple mutational events. The manual analysis highlighted 5 new strong HPE candidate genes showing recurrences in distinct families. Functional validations of these genes are foreseen.Keywords: complex genetic disorder, holoprosencephaly, multiple rare variants, whole exome sequencing
Procedia PDF Downloads 20328129 Bayesian Inference for High Dimensional Dynamic Spatio-Temporal Models
Authors: Sofia M. Karadimitriou, Kostas Triantafyllopoulos, Timothy Heaton
Abstract:
Reduced dimension Dynamic Spatio-Temporal Models (DSTMs) jointly describe the spatial and temporal evolution of a function observed subject to noise. A basic state space model is adopted for the discrete temporal variation, while a continuous autoregressive structure describes the continuous spatial evolution. Application of such a DSTM relies upon the pre-selection of a suitable reduced set of basic functions and this can present a challenge in practice. In this talk, we propose an online estimation method for high dimensional spatio-temporal data based upon DSTM and we attempt to resolve this issue by allowing the basis to adapt to the observed data. Specifically, we present a wavelet decomposition in order to obtain a parsimonious approximation of the spatial continuous process. This parsimony can be achieved by placing a Laplace prior distribution on the wavelet coefficients. The aim of using the Laplace prior, is to filter wavelet coefficients with low contribution, and thus achieve the dimension reduction with significant computation savings. We then propose a Hierarchical Bayesian State Space model, for the estimation of which we offer an appropriate particle filter algorithm. The proposed methodology is illustrated using real environmental data.Keywords: multidimensional Laplace prior, particle filtering, spatio-temporal modelling, wavelets
Procedia PDF Downloads 42728128 Investigating Complement Clause Choice in Written Educated Nigerian English (ENE)
Authors: Juliet Udoudom
Abstract:
Inappropriate complement selection constitutes one of the major features of non-standard complementation in the Nigerian users of English output of sentence construction. This paper investigates complement clause choice in Written Educated Nigerian English (ENE) and offers some results. It aims at determining preferred and dispreferred patterns of complement clause selection in respect of verb heads in English by selected Nigerian users of English. The complementation data analyzed in this investigation were obtained from experimental tasks designed to elicit complement categories of Verb – Noun -, Adjective – and Prepositional – heads in English. Insights from the Government – Binding relations were employed in analyzing data, which comprised responses obtained from one hundred subjects to a picture elicitation exercise, a grammaticality judgement test, and a free composition task. The findings indicate a general tendency for clausal complements (CPs) introduced by the complementizer that to be preferred by the subjects studied. Of the 235 tokens of clausal complements which occurred in our corpus, 128 of them representing 54.46% were CPs headed by that, while whether – and if-clauses recorded 31.07% and 8.94%, respectively. The complement clause-type which recorded the lowest incidence of choice was the CP headed by the Complementiser, for with a 5.53% incident of occurrence. Further findings from the study indicate that semantic features of relevant embedding verb heads were not taken into consideration in the choice of complementisers which introduce the respective complement clauses, hence the that-clause was chosen to complement verbs like prefer. In addition, the dispreferred choice of the for-clause is explicable in terms of the fact that the respondents studied regard ‘for’ as a preposition, and not a complementiser.Keywords: complement, complement clause complement selection, complementisers, government-binding
Procedia PDF Downloads 18828127 Assessment of Five Photoplethysmographic Methods for Estimating Heart Rate Variability
Authors: Akshay B. Pawar, Rohit Y. Parasnis
Abstract:
Heart Rate Variability (HRV) is a widely used indicator of the regulation between the autonomic nervous system (ANS) and the cardiovascular system. Besides being non-invasive, it also has the potential to predict mortality in cases involving critical injuries. The gold standard method for determining HRV is based on the analysis of RR interval time series extracted from ECG signals. However, because it is much more convenient to obtain photoplethysmogramic (PPG) signals as compared to ECG signals (which require the attachment of several electrodes to the body), many researchers have used pulse cycle intervals instead of RR intervals to estimate HRV. They have also compared this method with the gold standard technique. Though most of their observations indicate a strong correlation between the two methods, recent studies show that in healthy subjects, except for a few parameters, the pulse-based method cannot be a surrogate for the standard RR interval- based method. Moreover, the former tends to overestimate short-term variability in heart rate. This calls for improvements in or alternatives to the pulse-cycle interval method. In this study, besides the systolic peak-peak interval method (PP method) that has been studied several times, four recent PPG-based techniques, namely the first derivative peak-peak interval method (P1D method), the second derivative peak-peak interval method (P2D method), the valley-valley interval method (VV method) and the tangent-intersection interval method (TI method) were compared with the gold standard technique. ECG and PPG signals were obtained from 10 young and healthy adults (consisting of both males and females) seated in the armchair position. In order to de-noise these signals and eliminate baseline drift, they were passed through certain digital filters. After filtering, the following HRV parameters were computed from PPG using each of the five methods and also from ECG using the gold standard method: time domain parameters (SDNN, pNN50 and RMSSD), frequency domain parameters (Very low-frequency power (VLF), Low-frequency power (LF), High-frequency power (HF) and Total power or “TP”). Besides, Poincaré plots were also plotted and their SD1/SD2 ratios determined. The resulting sets of parameters were compared with those yielded by the standard method using measures of statistical correlation (correlation coefficient) as well as statistical agreement (Bland-Altman plots). From the viewpoint of correlation, our results show that the best PPG-based methods for the determination of most parameters and Poincaré plots are the P2D method (shows more than 93% correlation with the standard method) and the PP method (mean correlation: 88%) whereas the TI, VV and P1D methods perform poorly (<70% correlation in most cases). However, our evaluation of statistical agreement using Bland-Altman plots shows that none of the five techniques agrees satisfactorily well with the gold standard method as far as time-domain parameters are concerned. In conclusion, excellent statistical correlation implies that certain PPG-based methods provide a good amount of information on the pattern of heart rate variation, whereas poor statistical agreement implies that PPG cannot completely replace ECG in the determination of HRV.Keywords: photoplethysmography, heart rate variability, correlation coefficient, Bland-Altman plot
Procedia PDF Downloads 32328126 Lithium-Ion Battery State of Charge Estimation Using One State Hysteresis Model with Nonlinear Estimation Strategies
Authors: Mohammed Farag, Mina Attari, S. Andrew Gadsden, Saeid R. Habibi
Abstract:
Battery state of charge (SOC) estimation is an important parameter as it measures the total amount of electrical energy stored at a current time. The SOC percentage acts as a fuel gauge if it is compared with a conventional vehicle. Estimating the SOC is, therefore, essential for monitoring the amount of useful life remaining in the battery system. This paper looks at the implementation of three nonlinear estimation strategies for Li-Ion battery SOC estimation. One of the most common behavioral battery models is the one state hysteresis (OSH) model. The extended Kalman filter (EKF), the smooth variable structure filter (SVSF), and the time-varying smoothing boundary layer SVSF are applied on this model, and the results are compared.Keywords: state of charge estimation, battery modeling, one-state hysteresis, filtering and estimation
Procedia PDF Downloads 44428125 ScRNA-Seq RNA Sequencing-Based Program-Polygenic Risk Scores Associated with Pancreatic Cancer Risks in the UK Biobank Cohort
Authors: Yelin Zhao, Xinxiu Li, Martin Smelik, Oleg Sysoev, Firoj Mahmud, Dina Mansour Aly, Mikael Benson
Abstract:
Background: Early diagnosis of pancreatic cancer is clinically challenging due to vague, or no symptoms, and lack of biomarkers. Polygenic risk score (PRS) scores may provide a valuable tool to assess increased or decreased risk of PC. This study aimed to develop such PRS by filtering genetic variants identified by GWAS using transcriptional programs identified by single-cell RNA sequencing (scRNA-seq). Methods: ScRNA-seq data from 24 pancreatic ductal adenocarcinoma (PDAC) tumor samples and 11 normal pancreases were analyzed to identify differentially expressed genes (DEGs) in in tumor and microenvironment cell types compared to healthy tissues. Pathway analysis showed that the DEGs were enriched for hundreds of significant pathways. These were clustered into 40 “programs” based on gene similarity, using the Jaccard index. Published genetic variants associated with PDAC were mapped to each program to generate program PRSs (pPRSs). These pPRSs, along with five previously published PRSs (PGS000083, PGS000725, PGS000663, PGS000159, and PGS002264), were evaluated in a European-origin population from the UK Biobank, consisting of 1,310 PDAC participants and 407,473 non-pancreatic cancer participants. Stepwise Cox regression analysis was performed to determine associations between pPRSs with the development of PC, with adjustments of sex and principal components of genetic ancestry. Results: The PDAC genetic variants were mapped to 23 programs and were used to generate pPRSs for these programs. Four distinct pPRSs (P1, P6, P11, and P16) and two published PRSs (PGS000663 and PGS002264) were significantly associated with an increased risk of developing PC. Among these, P6 exhibited the greatest hazard ratio (adjusted HR[95% CI] = 1.67[1.14-2.45], p = 0.008). In contrast, P10 and P4 were associated with lower risk of developing PC (adjusted HR[95% CI] = 0.58[0.42-0.81], p = 0.001, and adjusted HR[95% CI] = 0.75[0.59-0.96], p = 0.019). By comparison, two of the five published PRS exhibited an association with PDAC onset with HR (PGS000663: adjusted HR[95% CI] = 1.24[1.14-1.35], p < 0.001 and PGS002264: adjusted HR[95% CI] = 1.14[1.07-1.22], p < 0.001). Conclusion: Compared to published PRSs, scRNA-seq-based pPRSs may be used not only to assess increased but also decreased risk of PDAC.Keywords: cox regression, pancreatic cancer, polygenic risk score, scRNA-seq, UK biobank
Procedia PDF Downloads 10128124 Survey on Securing the Optimized Link State Routing (OLSR) Protocol in Mobile Ad-hoc Network
Authors: Kimaya Subhash Gaikwad, S. B. Waykar
Abstract:
The mobile ad-hoc network (MANET) is collection of various types of nodes. In MANET various protocols are used for communication. In OLSR protocol, a node is selected as multipoint relay (MPR) node which broadcast the messages. As the MANET is open kind of network any malicious node can easily enter into the network and affect the performance of the network. The performance of network mainly depends on the components which are taking part into the communication. If the proper nodes are not selected for the communication then the probability of network being attacked is more. Therefore, it is important to select the more reliable and secure components in the network. MANET does not have any filtering so that only selected nodes can be used for communication. The openness of the MANET makes it easier to attack the communication. The most of the attack are on the Quality of service (QoS) of the network. This paper gives the overview of the various attacks that are possible on OLSR protocol and some solutions. The papers focus mainly on the OLSR protocol.Keywords: communication, MANET, OLSR, QoS
Procedia PDF Downloads 45028123 GC-MS Analysis of Bioactive Compounds in the Ethanolic Extract of Nest Material of Mud Wasp, Sceliphron caementarium
Authors: P. Susheela, Mary Rosaline, R. Radha
Abstract:
This research was designed to determine the bioactive compounds present in the nest samples of the mud dauber wasp, Sceliophron caementarium. Insects and insect-based products have been used for the treatment of various ailments from a very long time. It has been found that all over the world including the western societies and the indigenous populations, the usage of insect-based medicine plays an important role in various healing practices and magic rituals. Studies on the therapeutic usage of insects are negligible when compared to plants, the. In the present scenario, it is important to explore bioactive compounds from natural sources rather than depending on synthetic drugs that have adverse effects on human body. Keeping this in view, an attempt was made to analyze and identify bioactive components from the nest sample of the mud dauber wasp, Sceliophron caementarium. The nests of the mud dauber wasp, Sceliophron caementarium were collected from Coimbatore, Tamil Nadu, India. The nest sample was extracted with ethanol for 6-8 hours using Soxhlet apparatus. The final residue was obtained by filtering the extract through Whatman filter paper No.41. The GCMS analysis of the nest sample was performed using Perkin Elmer Elite - 5 capillary column. The resultant compounds were compared with the database of National Institute Standard and Technology (NIST), WILEY8, FAME. The GC-MS analysis of the concentrated ethanol extract revealed the presence of eight constituents like Methylene chloride, Eicosanoic acid, 1, 1’:3’, 1’’-Terphenyl, 5'-Phenyl, Di-N-Decylsulfone, 1, 2-Bis (Trimethylsilyl) Benzene, Androstane-11, 17-Dione, 3-[(Trimethylsilyl) Oxy]-, 17-[O-(Phenylmethyl) O. Most of the identified compounds were reported as having biological activities viz. anti-inflammatory, antibacterial and antifungal properties that can be of pharmaceutical importance and further study of these isolated compounds may prove their medicinal importance in future.Keywords: Sceliophron caementarium, Gas chromatography-mass spectrometry, ethanol extract, bioactive compounds
Procedia PDF Downloads 29528122 Data-Mining Approach to Analyzing Industrial Process Information for Real-Time Monitoring
Authors: Seung-Lock Seo
Abstract:
This work presents a data-mining empirical monitoring scheme for industrial processes with partially unbalanced data. Measurement data of good operations are relatively easy to gather, but in unusual special events or faults it is generally difficult to collect process information or almost impossible to analyze some noisy data of industrial processes. At this time some noise filtering techniques can be used to enhance process monitoring performance in a real-time basis. In addition, pre-processing of raw process data is helpful to eliminate unwanted variation of industrial process data. In this work, the performance of various monitoring schemes was tested and demonstrated for discrete batch process data. It showed that the monitoring performance was improved significantly in terms of monitoring success rate of given process faults.Keywords: data mining, process data, monitoring, safety, industrial processes
Procedia PDF Downloads 401