Search results for: target specificity
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3157

Search results for: target specificity

2977 Systematic and Simple Guidance for Feed Forward Design in Model Predictive Control

Authors: Shukri Dughman, Anthony Rossiter

Abstract:

This paper builds on earlier work which demonstrated that Model Predictive Control (MPC) may give a poor choice of default feed forward compensator. By first demonstrating the impact of future information of target changes on the performance, this paper proposes a pragmatic method for identifying the amount of future information on the target that can be utilised effectively in both finite and infinite horizon algorithms. Numerical illustrations in MATLAB give evidence of the efficacy of the proposal.

Keywords: model predictive control, tracking control, advance knowledge, feed forward

Procedia PDF Downloads 511
2976 Effectiveness of the Lacey Assessment of Preterm Infants to Predict Neuromotor Outcomes of Premature Babies at 12 Months Corrected Age

Authors: Thanooja Naushad, Meena Natarajan, Tushar Vasant Kulkarni

Abstract:

Background: The Lacey Assessment of Preterm Infants (LAPI) is used in clinical practice to identify premature babies at risk of neuromotor impairments, especially cerebral palsy. This study attempted to find the validity of the Lacey assessment of preterm infants to predict neuromotor outcomes of premature babies at 12 months corrected age and to compare its predictive ability with the brain ultrasound. Methods: This prospective cohort study included 89 preterm infants (45 females and 44 males) born below 35 weeks gestation who were admitted to the neonatal intensive care unit of a government hospital in Dubai. Initial assessment was done using the Lacey assessment after the babies reached 33 weeks postmenstrual age. Follow up assessment on neuromotor outcomes was done at 12 months (± 1 week) corrected age using two standardized outcome measures, i.e., infant neurological international battery and Alberta infant motor scale. Brain ultrasound data were collected retrospectively. Data were statistically analyzed, and the diagnostic accuracy of the Lacey assessment of preterm infants (LAPI) was calculated -when used alone and in combination with the brain ultrasound. Results: On comparison with brain ultrasound, the Lacey assessment showed superior specificity (96% vs. 77%), higher positive predictive value (57% vs. 22%), and higher positive likelihood ratio (18 vs. 3) to predict neuromotor outcomes at one year of age. The sensitivity of Lacey assessment was lower than brain ultrasound (66% vs. 83%), whereas specificity was similar (97% vs. 98%). A combination of Lacey assessment and brain ultrasound results showed higher sensitivity (80%), positive (66%), and negative (98%) predictive values, positive likelihood ratio (24), and test accuracy (95%) than Lacey assessment alone in predicting neurological outcomes. The negative predictive value of the Lacey assessment was similar to that of its combination with brain ultrasound (96%). Conclusion: Results of this study suggest that the Lacey assessment of preterm infants can be used as a supplementary assessment tool for premature babies in the neonatal intensive care unit. Due to its high specificity, Lacey assessment can be used to identify those babies at low risk of abnormal neuromotor outcomes at a later age. When used along with the findings of the brain ultrasound, Lacey assessment has better sensitivity to identify preterm babies at particular risk. These findings have applications in identifying premature babies who may benefit from early intervention services.

Keywords: brain ultrasound, lacey assessment of preterm infants, neuromotor outcomes, preterm

Procedia PDF Downloads 119
2975 Place, Female and Latino Identities in Kali Fajardo-Anstine’s Short Story Collection Sabrina and Corina

Authors: Jaroslav Kusnir

Abstract:

In her short story collection, Sabrina & Corina, Kali Fajardo-Anstine depicts mostly Latina characters of indigenous background living and travelling in the American West and the Southwest. In all the stories, place and the environment plays an important role in the construction of cultural identity of these characters that is influenced by their indigenous background, a specificity of the American West, its culture and environment, as well as a contemporary (modern) American culture, position of women and gender roles in a Latino community in the USA. This paper will analyze Fajardo-Anstine´s depiction of a specificity of place, especially of the American West and its role in a construction of Latino/a cultural identity in a modern American society as manifested especially in Fajardo-Anstine´s stories Any Further West and Sabrina & Corina. At the same time, the paper will point out Fajardo-Anstine´s construction of cultural identity of female characters and their gender roles in both Latino and a contemporary American societies. The research results show that the formation of Latina cultural identity is closely connected with both place, that is the American West and the Soutwest as well as with Latina and contemporary American cultures.

Keywords: American culture, american west, cultural identity, female identity, latina identity, place

Procedia PDF Downloads 64
2974 A Topological Approach for Motion Track Discrimination

Authors: Tegan H. Emerson, Colin C. Olson, George Stantchev, Jason A. Edelberg, Michael Wilson

Abstract:

Detecting small targets at range is difficult because there is not enough spatial information present in an image sub-region containing the target to use correlation-based methods to differentiate it from dynamic confusers present in the scene. Moreover, this lack of spatial information also disqualifies the use of most state-of-the-art deep learning image-based classifiers. Here, we use characteristics of target tracks extracted from video sequences as data from which to derive distinguishing topological features that help robustly differentiate targets of interest from confusers. In particular, we calculate persistent homology from time-delayed embeddings of dynamic statistics calculated from motion tracks extracted from a wide field-of-view video stream. In short, we use topological methods to extract features related to target motion dynamics that are useful for classification and disambiguation and show that small targets can be detected at range with high probability.

Keywords: motion tracks, persistence images, time-delay embedding, topological data analysis

Procedia PDF Downloads 94
2973 Lies of Police Interrogators in the Ultimatum Game

Authors: Eitan Elaad

Abstract:

The present study's purpose was to examine lyingand pretend fairness by police interrogators in sharing situations. Forty police officers and 40 laypeople from the community, all males, self-assessed their lie-telling ability, rated the frequency of their lies, evaluated the acceptability of lying, and indicated using rational and intuitive thinking while lying. Next, according to the ultimatum game procedure, participants were asked to share 100 points with a virtual target, either a male police interrogator or a male layman. Participantsallocated points to the target person bearing in mind that the other person must accept their offer. Participants' goal was to retain as many points as possible, and to this end, they could tell the target person that fewer than 100 points were available for distribution. The difference between the available 100 points and the sum of points designated for sharing defines lying. The ratio of offered and designated points defines pretend fairness. Results indicate that those police officers lied more than laypeople. Similar results emergedeven when the target person was a police interrogator. However, police interrogators presented higher pretend fairness than laypeople. The higher pretend fairness may be in line with interrogation tactics of persuasion used in the criminal interrogation. Higher-lying frequency reported by police interrogators compared with laypeople support the present results. Finally, lie acceptability predicted lying in the ultimatum game. Specifically, participants who rated lying as more acceptable tended to lie more than low acceptability raters.

Keywords: lying, police interrogators, lie acceptability, ultimatum game, pretend fairness

Procedia PDF Downloads 134
2972 Ultrasensitive Detection and Discrimination of Cancer-Related Single Nucleotide Polymorphisms Using Poly-Enzyme Polymer Bead Amplification

Authors: Lorico D. S. Lapitan Jr., Yihan Xu, Yuan Guo, Dejian Zhou

Abstract:

The ability of ultrasensitive detection of specific genes and discrimination of single nucleotide polymorphisms is important for clinical diagnosis and biomedical research. Herein, we report the development of a new ultrasensitive approach for label-free DNA detection using magnetic nanoparticle (MNP) assisted rapid target capture/separation in combination with signal amplification using poly-enzyme tagged polymer nanobead. The sensor uses an MNP linked capture DNA and a biotin modified signal DNA to sandwich bind the target followed by ligation to provide high single-nucleotide polymorphism discrimination. Only the presence of a perfect match target DNA yields a covalent linkage between the capture and signal DNAs for subsequent conjugation of a neutravidin-modified horseradish peroxidase (HRP) enzyme through the strong biotin-nuetravidin interaction. This converts each captured DNA target into an HRP which can convert millions of copies of a non-fluorescent substrate (amplex red) to a highly fluorescent product (resorufin), for great signal amplification. The use of polymer nanobead each tagged with thousands of copies of HRPs as the signal amplifier greatly improves the signal amplification power, leading to greatly improved sensitivity. We show our biosensing approach can specifically detect an unlabeled DNA target down to 10 aM with a wide dynamic range of 5 orders of magnitude (from 0.001 fM to 100.0 fM). Furthermore, our approach has a high discrimination between a perfectly matched gene and its cancer-related single-base mismatch targets (SNPs): It can positively detect the perfect match DNA target even in the presence of 100 fold excess of co-existing SNPs. This sensing approach also works robustly in clinical relevant media (e.g. 10% human serum) and gives almost the same SNP discrimination ratio as that in clean buffers. Therefore, this ultrasensitive SNP biosensor appears to be well-suited for potential diagnostic applications of genetic diseases.

Keywords: DNA detection, polymer beads, signal amplification, single nucleotide polymorphisms

Procedia PDF Downloads 234
2971 O.MG- It’s a Cyber-Enabled Fraud

Authors: Damola O. Lawal, David W. Gresty, Diane E. Gan, Louise Hewitt

Abstract:

This paper investigates the feasibility of using a programmable USB such as the O.MG Cable to perform a file tampering attack. Here, the O.MG Cable, an apparently harmless mobile device charger, is used in an unauthorized way to alter the content of a file (accounts record-January_Contributions.xlsx). The aim is to determine if a forensics analyst can reliably determine who has altered the target file; the O.MG Cable or the user of the machine. This work highlights some of the traces of the O.MG Cable left behind on the target computer itself, such as the Product ID (PID) and Vendor ID (ID). Also discussed is the O.MG Cable’s behavior during the experiments. We determine if a forensics analyst could identify if any evidence has been left behind by the programmable device on the target file once it has been removed from the computer to establish if the analyst would be able to link the traces left by the O.MG Cable to the file tampering. It was discovered that the forensic analyst might mistake the actions of the O.MG Cable for the computer users. Experiments carried out in this work could further the discussion as to whether an innocent user could be punished for the unauthorized changes made by a programmable device.

Keywords: O.MG cable, programmable USB, file tampering attack, digital evidence credibility, miscarriage of justice, cyber fraud

Procedia PDF Downloads 136
2970 Optimized Real Ground Motion Scaling for Vulnerability Assessment of Building Considering the Spectral Uncertainty and Shape

Authors: Chen Bo, Wen Zengping

Abstract:

Based on the results of previous studies, we focus on the research of real ground motion selection and scaling method for structural performance-based seismic evaluation using nonlinear dynamic analysis. The input of earthquake ground motion should be determined appropriately to make them compatible with the site-specific hazard level considered. Thus, an optimized selection and scaling method are established including the use of not only Monte Carlo simulation method to create the stochastic simulation spectrum considering the multivariate lognormal distribution of target spectrum, but also a spectral shape parameter. Its applications in structural fragility analysis are demonstrated through case studies. Compared to the previous scheme with no consideration of the uncertainty of target spectrum, the method shown here can make sure that the selected records are in good agreement with the median value, standard deviation and spectral correction of the target spectrum, and greatly reveal the uncertainty feature of site-specific hazard level. Meanwhile, it can help improve computational efficiency and matching accuracy. Given the important infection of target spectrum’s uncertainty on structural seismic fragility analysis, this work can provide the reasonable and reliable basis for structural seismic evaluation under scenario earthquake environment.

Keywords: ground motion selection, scaling method, seismic fragility analysis, spectral shape

Procedia PDF Downloads 270
2969 Comparative Connectionism: Study of the Biological Constraints of Learning Through the Manipulation of Various Architectures in a Neural Network Model under the Biological Principle of the Correlation Between Structure and Function

Authors: Giselle Maggie-Fer Castañeda Lozano

Abstract:

The main objective of this research was to explore the role of neural network architectures in simulating behavioral phenomena as a potential explanation for selective associations, specifically related to biological constraints on learning. Biological constraints on learning refer to the limitations observed in conditioning procedures, where learning is expected to occur. The study involved simulations of five different experiments exploring various phenomena and sources of biological constraints in learning. These simulations included the interaction between response and reinforcer, stimulus and reinforcer, specificity of stimulus-reinforcer associations, species differences, neuroanatomical constraints, and learning in uncontrolled conditions. The overall results demonstrated that by manipulating neural network architectures, conditions can be created to model and explain diverse biological constraints frequently reported in comparative psychology literature as learning typicities. Additionally, the simulations offer predictive content worthy of experimental testing in the pursuit of new discoveries regarding the specificity of learning. The implications and limitations of these findings are discussed. Finally, it is suggested that this research could inaugurate a line of inquiry involving the use of neural networks to study biological factors in behavior, fostering the development of more ethical and precise research practices.

Keywords: comparative psychology, connectionism, conditioning, experimental analysis of behavior, neural networks

Procedia PDF Downloads 43
2968 Seamless Mobility in Heterogeneous Mobile Networks

Authors: Mohab Magdy Mostafa Mohamed

Abstract:

The objective of this paper is to introduce a vertical handover (VHO) algorithm between wireless LANs (WLANs) and LTE mobile networks. The proposed algorithm is based on the fuzzy control theory and takes into consideration power level, subscriber velocity, and target cell load instead of only power level in traditional algorithms. Simulation results show that network performance in terms of number of handovers and handover occurrence distance is improved.

Keywords: vertical handover, fuzzy control theory, power level, speed, target cell load

Procedia PDF Downloads 326
2967 Sensitivity and Specificity of Some Serological Tests Used for Diagnosis of Bovine Brucellosis in Egypt on Bacteriological and Molecular Basis

Authors: Hosein I. Hosein, Ragab Azzam, Ahmed M. S. Menshawy, Sherin Rouby, Khaled Hendy, Ayman Mahrous, Hany Hussien

Abstract:

Brucellosis is a highly contagious bacterial zoonotic disease of a worldwide spread and has different names; Infectious or enzootic abortion and Bang's disease in animals; and Mediterranean or Malta fever, Undulant Fever and Rock fever in humans. It is caused by the different species of genus Brucella which is a Gram-negative, aerobic, non-spore forming, facultative intracellular bacterium. Brucella affects a wide range of mammals including bovines, small ruminants, pigs, equines, rodents, marine mammals as well as human resulting in serious economic losses in animal populations. In human, Brucella causes a severe illness representing a great public health problem. The disease was reported in Egypt for the first time in 1939; since then the disease remained endemic at high levels among cattle, buffalo, sheep and goat and is still representing a public health hazard. The annual economic losses due to brucellosis were estimated to be about 60 million Egyptian pounds yearly, but actual estimates are still missing despite almost 30 years of implementation of the Egyptian control programme. Despite being the gold standard, bacterial isolation has been reported to show poor sensitivity for samples with low-level of Brucella and is impractical for regular screening of large populations. Thus, serological tests still remain the corner stone for routine diagnosis of brucellosis, especially in developing countries. In the present study, a total of 1533 cows (256 from Beni-Suef Governorate, 445 from Al-Fayoum Governorate and 832 from Damietta Governorate), were employed for estimation of relative sensitivity, relative specificity, positive predictive value and negative predictive value of buffered acidified plate antigen test (BPAT), rose bengal test (RBT) and complement fixation test (CFT). The overall seroprevalence of brucellosis revealed (19.63%). Relative sensitivity, relative specificity, positive predictive value and negative predictive value of BPAT,RBT and CFT were estimated as, (96.27 %, 96.76 %, 87.65 % and 99.10 %), (93.42 %, 96.27 %, 90.16 % and 98.35%) and (89.30 %, 98.60 %, 94.35 %and 97.24 %) respectively. BPAT showed the highest sensitivity among the three employed serological tests. RBT was less specific than BPAT. CFT showed the least sensitivity 89.30 % among the three employed serological tests but showed the highest specificity. Different tissues specimens of 22 seropositive cows (spleen, retropharyngeal udder, and supra-mammary lymph nodes) were subjected for bacteriological studies for isolation and identification of Brucella organisms. Brucella melitensis biovar 3 could be recovered from 12 (54.55%) cows. Bacteriological examinations failed to classify 10 cases (45.45%) and were culture negative. Bruce-ladder PCR was carried out for molecular identification of the 12 Brucella isolates at the species level. Three fragments of 587 bp, 1071 bp and 1682 bp sizes were amplified indicating Brucella melitensis. The results indicated the importance of using several procedures to overcome the problem of escaping of some infected animals from diagnosis.Bruce-ladder PCR is an important tool for diagnosis and epidemiologic studies, providing relevant information for identification of Brucella spp.

Keywords: brucellosis, relative sensitivity, relative specificity, Bruce-ladder, Egypt

Procedia PDF Downloads 326
2966 Business and Psychological Principles Integrated into Automated Capital Investment Systems through Mathematical Algorithms

Authors: Cristian Pauna

Abstract:

With few steps away from the 2020, investments in financial markets is a common activity nowadays. In the electronic trading environment, the automated investment software has become a major part in the business intelligence system of any modern financial company. The investment decisions are assisted and/or made automatically by computers using mathematical algorithms today. The complexity of these algorithms requires computer assistance in the investment process. This paper will present several investment strategies that can be automated with algorithmic trading for Deutscher Aktienindex DAX30. It was found that, based on several price action mathematical models used for high-frequency trading some investment strategies can be optimized and improved for automated investments with good results. This paper will present the way to automate these investment decisions. Automated signals will be built using all of these strategies. Three major types of investment strategies were found in this study. The types are separated by the target length and by the exit strategy used. The exit decisions will be also automated and the paper will present the specificity for each investment type. A comparative study will be also included in this paper in order to reveal the differences between strategies. Based on these results, the profit and the capital exposure will be compared and analyzed in order to qualify the investment methodologies presented and to compare them with any other investment system. As conclusion, some major investment strategies will be revealed and compared in order to be considered for inclusion in any automated investment system.

Keywords: Algorithmic trading, automated investment systems, limit conditions, trading principles, trading strategies

Procedia PDF Downloads 169
2965 Development of Visual Working Memory Precision: A Cross-Sectional Study of Simultaneously Delayed Responses Paradigm

Authors: Yao Fu, Xingli Zhang, Jiannong Shi

Abstract:

Visual working memory (VWM) capacity is the ability to maintain and manipulate short-term information which is not currently available. It is well known for its significance to form the basis of numerous cognitive abilities and its limitation in holding information. VWM span, the most popular measurable indicator, is found to reach the adult level (3-4 items) around 12-13 years’ old, while less is known about the precision development of the VWM capacity. By using simultaneously delayed responses paradigm, the present study investigates the development of VWM precision among 6-18-year-old children and young adults, besides its possible relationships with fluid intelligence and span. Results showed that precision and span both increased with age, and precision reached the maximum in 16-17 age-range. Moreover, when remembering 3 simultaneously presented items, the probability of remembering target item correlated with fluid intelligence and the probability of wrap errors (misbinding target and non-target items) correlated with age. When remembering more items, children had worse performance than adults due to their wrap errors. Compared to span, VWM precision was effective predictor of intelligence even after controlling for age. These results suggest that unlike VWM span, precision developed in a slow, yet longer fashion. Moreover, decreasing probability of wrap errors might be the main reason for the development of precision. Last, precision correlated more closely with intelligence than span in childhood and adolescence, which might be caused by the probability of remembering target item.

Keywords: fluid intelligence, precision, visual working memory, wrap errors

Procedia PDF Downloads 255
2964 Automatic Reporting System for Transcriptome Indel Identification and Annotation Based on Snapshot of Next-Generation Sequencing Reads Alignment

Authors: Shuo Mu, Guangzhi Jiang, Jinsa Chen

Abstract:

The analysis of Indel for RNA sequencing of clinical samples is easily affected by sequencing experiment errors and software selection. In order to improve the efficiency and accuracy of analysis, we developed an automatic reporting system for Indel recognition and annotation based on image snapshot of transcriptome reads alignment. This system includes sequence local-assembly and realignment, target point snapshot, and image-based recognition processes. We integrated high-confidence Indel dataset from several known databases as a training set to improve the accuracy of image processing and added a bioinformatical processing module to annotate and filter Indel artifacts. Subsequently, the system will automatically generate data, including data quality levels and images results report. Sanger sequencing verification of the reference Indel mutation of cell line NA12878 showed that the process can achieve 83% sensitivity and 96% specificity. Analysis of the collected clinical samples showed that the interpretation accuracy of the process was equivalent to that of manual inspection, and the processing efficiency showed a significant improvement. This work shows the feasibility of accurate Indel analysis of clinical next-generation sequencing (NGS) transcriptome. This result may be useful for RNA study for clinical samples with microsatellite instability in immunotherapy in the future.

Keywords: automatic reporting, indel, next-generation sequencing, NGS, transcriptome

Procedia PDF Downloads 157
2963 Classification of Echo Signals Based on Deep Learning

Authors: Aisulu Tileukulova, Zhexebay Dauren

Abstract:

Radar plays an important role because it is widely used in civil and military fields. Target detection is one of the most important radar applications. The accuracy of detecting inconspicuous aerial objects in radar facilities is lower against the background of noise. Convolutional neural networks can be used to improve the recognition of this type of aerial object. The purpose of this work is to develop an algorithm for recognizing aerial objects using convolutional neural networks, as well as training a neural network. In this paper, the structure of a convolutional neural network (CNN) consists of different types of layers: 8 convolutional layers and 3 layers of a fully connected perceptron. ReLU is used as an activation function in convolutional layers, while the last layer uses softmax. It is necessary to form a data set for training a neural network in order to detect a target. We built a Confusion Matrix of the CNN model to measure the effectiveness of our model. The results showed that the accuracy when testing the model was 95.7%. Classification of echo signals using CNN shows high accuracy and significantly speeds up the process of predicting the target.

Keywords: radar, neural network, convolutional neural network, echo signals

Procedia PDF Downloads 323
2962 Contactless and Multiple Space Debris Removal by Micro to Nanno Satellites

Authors: Junichiro Kawaguchi

Abstract:

Space debris problems have emerged and threatened the use of low earth orbit around the Earth owing to a large number of spacecraft. In debris removal, a number of research and patents have been proposed and published so far. They assume servicing spacecraft, robots to be built for accessing the target debris objects. The robots should be sophisticated enough automatically to access the debris articulating the attitude and the translation motion with respect to the debris. This paper presents the idea of using the torpedo-like third unsophisticated and disposable body, in addition to the first body of the servicing robot and the second body of the target debris. The third body is launched from the first body from a distance farer than the size of the second body. This paper presents the method and the system, so that the third body is launched from the first body. The third body carries both a net and an inflatable or extendible drag deceleration device and is built small and light. This method enables even a micro to nano satellite to perform contactless and multiple debris removal even via a single flight.

Keywords: ballute, debris removal, echo satellite, gossamer, gun-net, inflatable space structure, small satellite, un-cooperated target

Procedia PDF Downloads 97
2961 Amplitude and Latency of P300 Component from Auditory Stimulus in Different Types of Personality: An Event Related Potential Study

Authors: Nasir Yusoff, Ahmad Adamu Adamu, Tahamina Begum, Faruque Reza

Abstract:

The P300 from Event related potential (ERP) explains the psycho-physiological phenomenon in human body. The present study aims to identify the differences of amplitude and latency of P300 component from auditory stimuli, between ambiversion and extraversion types of personality. Ambivert (N=20) and extravert (N=20) undergoing ERP recording at the Hospital Universiti Sains Malaysia (HUSM) laboratory. Electroencephalogram data was recorded with oddball paradigm, counting auditory standard and target tones, from nine electrode sites (Fz, Cz, Pz, T3, T4, T5, T6, P3 and P4) by using the 128 HydroCel Geodesic Sensor Net. The P300 latency of the target tones at all electrodes were insignificant. Similarly, the P300 latency of the standard tones were also insignificant except at Fz and T3 electrode. Likewise, the P300 amplitude of the target and standard tone in all electrode sites were insignificant. Extravert and ambivert indicate similar characteristic in cognition processing from auditory task.

Keywords: amplitude, event related potential, p300 component, latency

Procedia PDF Downloads 347
2960 Antibody Reactivity of Synthetic Peptides Belonging to Proteins Encoded by Genes Located in Mycobacterium tuberculosis-Specific Genomic Regions of Differences

Authors: Abu Salim Mustafa

Abstract:

The comparisons of mycobacterial genomes have identified several Mycobacterium tuberculosis-specific genomic regions that are absent in other mycobacteria and are known as regions of differences. Due to M. tuberculosis-specificity, the peptides encoded by these regions could be useful in the specific diagnosis of tuberculosis. To explore this possibility, overlapping synthetic peptides corresponding to 39 proteins predicted to be encoded by genes present in regions of differences were tested for antibody-reactivity with sera from tuberculosis patients and healthy subjects. The results identified four immunodominant peptides corresponding to four different proteins, with three of the peptides showing significantly stronger antibody reactivity and rate of positivity with sera from tuberculosis patients than healthy subjects. The fourth peptide was recognized equally well by the sera of tuberculosis patients as well as healthy subjects. Predication of antibody epitopes by bioinformatics analyses using ABCpred server predicted multiple linear epitopes in each peptide. Furthermore, peptide sequence analysis for sequence identity using BLAST suggested M. tuberculosis-specificity for the three peptides that had preferential reactivity with sera from tuberculosis patients, but the peptide with equal reactivity with sera of TB patients and healthy subjects showed significant identity with sequences present in nob-tuberculous mycobacteria. The three identified M. tuberculosis-specific immunodominant peptides may be useful in the serological diagnosis of tuberculosis.

Keywords: genomic regions of differences, Mycobacterium tuberculossis, peptides, serodiagnosis

Procedia PDF Downloads 165
2959 Personalized Social Resource Recommender Systems on Interest-Based Social Networks

Authors: C. L. Huang, J. J. Sia

Abstract:

The interest-based social networks, also known as social bookmark sharing systems, are useful platforms for people to conveniently read and collect internet resources. These platforms also providing function of social networks, and users can share and explore internet resources from the social networks. Providing personalized internet resources to users is an important issue on these platforms. This study uses two types of relationship on the social networks—following and follower and proposes a collaborative recommender system, consisting of two main steps. First, this study calculates the relationship strength between the target user and the target user's followings and followers to find top-N similar neighbors. Second, from the top-N similar neighbors, the articles (internet resources) that may interest the target user are recommended to the target user. In this system, users can efficiently obtain recent, related and diverse internet resources (knowledge) from the interest-based social network. This study collected the experimental dataset from Diigo, which is a famous bookmark sharing system. The experimental results show that the proposed recommendation model is more accurate than two traditional baseline recommendation models but slightly lower than the cosine model in accuracy. However, in the metrics of the diversity and executing time, our proposed model outperforms the cosine model.

Keywords: recommender systems, social networks, tagging, bookmark sharing systems, collaborative recommender systems, knowledge management

Procedia PDF Downloads 149
2958 Easymodel: Web-based Bioinformatics Software for Protein Modeling Based on Modeller

Authors: Alireza Dantism

Abstract:

Presently, describing the function of a protein sequence is one of the most common problems in biology. Usually, this problem can be facilitated by studying the three-dimensional structure of proteins. In the absence of a protein structure, comparative modeling often provides a useful three-dimensional model of the protein that is dependent on at least one known protein structure. Comparative modeling predicts the three-dimensional structure of a given protein sequence (target) mainly based on its alignment with one or more proteins of known structure (templates). Comparative modeling consists of four main steps 1. Similarity between the target sequence and at least one known template structure 2. Alignment of target sequence and template(s) 3. Build a model based on alignment with the selected template(s). 4. Prediction of model errors 5. Optimization of the built model There are many computer programs and web servers that automate the comparative modeling process. One of the most important advantages of these servers is that it makes comparative modeling available to both experts and non-experts, and they can easily do their own modeling without the need for programming knowledge, but some other experts prefer using programming knowledge and do their modeling manually because by doing this they can maximize the accuracy of their modeling. In this study, a web-based tool has been designed to predict the tertiary structure of proteins using PHP and Python programming languages. This tool is called EasyModel. EasyModel can receive, according to the user's inputs, the desired unknown sequence (which we know as the target) in this study, the protein sequence file (template), etc., which also has a percentage of similarity with the primary sequence, and its third structure Predict the unknown sequence and present the results in the form of graphs and constructed protein files.

Keywords: structural bioinformatics, protein tertiary structure prediction, modeling, comparative modeling, modeller

Procedia PDF Downloads 71
2957 Modeling Search-And-Rescue Operations by Autonomous Mobile Robots at Sea

Authors: B. Kriheli, E. Levner, T. C. E. Cheng, C. T. Ng

Abstract:

During the last decades, research interest in planning, scheduling, and control of emergency response operations, especially people rescue and evacuation from the dangerous zone of marine accidents, has increased dramatically. Until the survivors (called ‘targets’) are found and saved, it may cause loss or damage whose extent depends on the location of the targets and the search duration. The problem is to efficiently search for and detect/rescue the targets as soon as possible with the help of intelligent mobile robots so as to maximize the number of saved people and/or minimize the search cost under restrictions on the amount of saved people within the allowable response time. We consider a special situation when the autonomous mobile robots (AMR), e.g., unmanned aerial vehicles and remote-controlled robo-ships have no operator on board as they are guided and completely controlled by on-board sensors and computer programs. We construct a mathematical model for the search process in an uncertain environment and provide a new fast algorithm for scheduling the activities of the autonomous robots during the search-and rescue missions after an accident at sea. We presume that in the unknown environments, the AMR’s search-and-rescue activity is subject to two types of error: (i) a 'false-negative' detection error where a target object is not discovered (‘overlooked') by the AMR’s sensors in spite that the AMR is in a close neighborhood of the latter and (ii) a 'false-positive' detection error, also known as ‘a false alarm’, in which a clean place or area is wrongly classified by the AMR’s sensors as a correct target. As the general resource-constrained discrete search problem is NP-hard, we restrict our study to finding local-optimal strategies. A specificity of the considered operational research problem in comparison with the traditional Kadane-De Groot-Stone search models is that in our model the probability of the successful search outcome depends not only on cost/time/probability parameters assigned to each individual location but, as well, on parameters characterizing the entire history of (unsuccessful) search before selecting any next location. We provide a fast approximation algorithm for finding the AMR route adopting a greedy search strategy in which, in each step, the on-board computer computes a current search effectiveness value for each location in the zone and sequentially searches for a location with the highest search effectiveness value. Extensive experiments with random and real-life data provide strong evidence in favor of the suggested operations research model and corresponding algorithm.

Keywords: disaster management, intelligent robots, scheduling algorithm, search-and-rescue at sea

Procedia PDF Downloads 153
2956 Role of Activated Partial Thromboplastin Time (APTT) to Assess the Need of Platelet Transfusion in Dengue

Authors: Kalyan Koganti

Abstract:

Background: In India, platelet transfusions are given to large no. of patients suffering from dengue due to the fear of bleeding especially when the platelet counts are low. Though many patients do not bleed when the platelet count falls to less than 20,000, certain patients bleed even if the platelet counts are more than 20,000 without any comorbid condition (like gastrointestinal ulcer) in the past. This fear has led to huge amounts of unnecessary platelet transfusions which cause significant economic burden to low and middle-income countries like India and also sometimes these transfusions end with transfusion-related adverse reactions. Objective: To identify the role of Activated Partial Thromboplastin Time (APTT) in comparison with thrombocytoenia as an indicator to assess the real need of platelet transfusions. Method: A prospective study was conducted at a hospital in South India which included 176 admitted cases of dengue confirmed by immunochromatography. APTT was performed in all these patients along with platelet count. Cut off values of > 60 seconds for APTT and < 20,000 for platelet count were considered to assess the bleeding manifestations. Results: Among the total 176 patients, 56 patients had bleeding manifestations like malena, hematuria, bleeding gums etc. APTT > 60 seconds had a sensitivity and specificity of 93% and 90% respectively in identifying bleeding manifestations where as platelet count of < 20,000 had a sensitivity and specificity of 64% and 73% respectively. Conclusion: Elevated APTT levels can be considered as an indicator to assess the need of platelet transfusion in dengue. As there is a significant variation among patients who bleed with respect to platelet count, APTT can be considered to avoid unnecessary transfusions.

Keywords: activated partial thromboplastin time, dengue, platelet transfusion, thrombocytopenia

Procedia PDF Downloads 199
2955 An Analysis of L1 Effects on the Learning of EFL: A Case Study of Undergraduate EFL Learners at Universities in Pakistan

Authors: Nadir Ali Mugheri, Shaukat Ali Lohar

Abstract:

In a multilingual society like Pakistan, code switching is commonly observed in different contexts. Mostly people use L1 (Native Languages) and L2 for common communications and L3 (i.e. English, Urdu, Sindhi) in formal contexts and for academic writings. Such a frequent code switching does affect EFL learners' acquisition of grammar and lexis of the target language which in the long run result in different types of errors in their writings. The current study is to investigate and identify common elements of L1 and L2 (spoken by students of the Universities in Pakistan) which create hindrances for EFL learners. Case study method was used for this research. Formal writings of 400 EFL learners (as participants from various Universities of the country) were observed. Among 400 participants, 200 were female and 200 were male EFL learners having different academic backgrounds. Errors found were categorized into different types according to grammatical items, the difference in meanings, structure of sentences and identifiers of tenses of L1 or L2 in comparison with those of the target language. The findings showed that EFL learners in Pakistani varsities have serious problems in their writings and they committed serious errors related to the grammar and meanings of the target language. After analysis of the committed errors, the results were found in the affirmation of the hypothesis that L1 or L2 does affect EFL learners. The research suggests in the end to adopt natural ways in pedagogy like task-based learning or communicative methods using contextualized material so as to avoid impediments of L1 or L2 in acquisition the target language.

Keywords: multilingualism, L2 acquisition, code switching, language acquisition, communicative language teaching

Procedia PDF Downloads 267
2954 Improved Acoustic Source Sensing and Localization Based On Robot Locomotion

Authors: V. Ramu Reddy, Parijat Deshpande, Ranjan Dasgupta

Abstract:

This paper presents different methodology for an acoustic source sensing and localization in an unknown environment. The developed methodology includes an acoustic based sensing and localization system, a converging target localization based on the recursive direction of arrival (DOA) error minimization, and a regressive obstacle avoidance function. Our method is able to augment the existing proven localization techniques and improve results incrementally by utilizing robot locomotion and is capable of converging to a position estimate with greater accuracy using fewer measurements. The results also evinced the DOA error minimization at each iteration, improvement in time for reaching the destination and the efficiency of this target localization method as gradually converging to the real target position. Initially, the system is tested using Kinect mounted on turntable with DOA markings which serve as a ground truth and then our approach is validated using a FireBird VI (FBVI) mobile robot on which Kinect is used to obtain bearing information.

Keywords: acoustic source localization, acoustic sensing, recursive direction of arrival, robot locomotion

Procedia PDF Downloads 473
2953 A New Learning Automata-Based Algorithm to the Priority-Based Target Coverage Problem in Directional Sensor Networks

Authors: Shaharuddin Salleh, Sara Marouf, Hosein Mohammadi

Abstract:

Directional sensor networks (DSNs) have recently attracted a great deal of attention due to their extensive applications in a wide range of situations. One of the most important problems associated with DSNs is covering a set of targets in a given area and, at the same time, maximizing the network lifetime. This is due to limitation in sensing angle and battery power of the directional sensors. This problem gets more complicated by the possibility that targets may have different coverage requirements. In the present study, this problem is referred to as priority-based target coverage (PTC). As sensors are often densely deployed, organizing the sensors into several cover sets and then activating these cover sets successively is a promising solution to this problem. In this paper, we propose a learning automata-based algorithm to organize the directional sensors into several cover sets in such a way that each cover set could satisfy coverage requirements of all the targets. Several experiments are conducted to evaluate the performance of the proposed algorithm. The results demonstrated that the algorithms were able to contribute to solving the problem.

Keywords: directional sensor networks, target coverage problem, cover set formation, learning automata

Procedia PDF Downloads 388
2952 Genome-Wide Mining of Potential Guide RNAs for Streptococcus pyogenes and Neisseria meningitides CRISPR-Cas Systems for Genome Engineering

Authors: Farahnaz Sadat Golestan Hashemi, Mohd Razi Ismail, Mohd Y. Rafii

Abstract:

Clustered regularly interspaced short palindromic repeats (CRISPR) and CRISPR-associated protein (Cas) system can facilitate targeted genome editing in organisms. Dual or single guide RNA (gRNA) can program the Cas9 nuclease to cut target DNA in particular areas; thus, introducing concise mutations either via error-prone non-homologous end-joining repairing or via incorporating foreign DNAs by homologous recombination between donor DNA and target area. In spite of high demand of such promising technology, developing a well-organized procedure in order for reliable mining of potential target sites for gRNAs in large genomic data is still challenging. Hence, we aimed to perform high-throughput detection of target sites by specific PAMs for not only common Streptococcus pyogenes (SpCas9) but also for Neisseria meningitides (NmCas9) CRISPR-Cas systems. Previous research confirmed the successful application of such RNA-guided Cas9 orthologs for effective gene targeting and subsequently genome manipulation. However, Cas9 orthologs need their particular PAM sequence for DNA cleavage activity. Activity levels are based on the sequence of the protospacer and specific combinations of favorable PAM bases. Therefore, based on the specific length and sequence of PAM followed by a constant length of the target site for the two orthogonals of Cas9 protein, we created a reliable procedure to explore possible gRNA sequences. To mine CRISPR target sites, four different searching modes of sgRNA binding to target DNA strand were applied. These searching modes are as follows i) coding strand searching, ii) anti-coding strand searching, iii) both strand searching, and iv) paired-gRNA searching. Finally, a complete list of all potential gRNAs along with their locations, strands, and PAMs sequence orientation can be provided for both SpCas9 as well as another potential Cas9 ortholog (NmCas9). The artificial design of potential gRNAs in a genome of interest can accelerate functional genomic studies. Consequently, the application of such novel genome editing tool (CRISPR/Cas technology) will enhance by presenting increased versatility and efficiency.

Keywords: CRISPR/Cas9 genome editing, gRNA mining, SpCas9, NmCas9

Procedia PDF Downloads 234
2951 Impedimetric Phage-Based Sensor for the Rapid Detection of Staphylococcus aureus from Nasal Swab

Authors: Z. Yousefniayejahr, S. Bolognini, A. Bonini, C. Campobasso, N. Poma, F. Vivaldi, M. Di Luca, A. Tavanti, F. Di Francesco

Abstract:

Pathogenic bacteria represent a threat to healthcare systems and the food industry because their rapid detection remains challenging. Electrochemical biosensors are gaining prominence as a novel technology for the detection of pathogens due to intrinsic features such as low cost, rapid response time, and portability, which make them a valuable alternative to traditional methodologies. These sensors use biorecognition elements that are crucial for the identification of specific bacteria. In this context, bacteriophages are promising tools for their inherent high selectivity towards bacterial hosts, which is of fundamental importance when detecting bacterial pathogens in complex biological samples. In this study, we present the development of a low-cost and portable sensor based on the Zeno phage for the rapid detection of Staphylococcus aureus. Screen-printed gold electrodes functionalized with the Zeno phage were used, and electrochemical impedance spectroscopy was applied to evaluate the change of the charge transfer resistance (Rct) as a result of the interaction with S. aureus MRSA ATCC 43300. The phage-based biosensor showed a linear range from 101 to 104 CFU/mL with a 20-minute response time and a limit of detection (LOD) of 1.2 CFU/mL under physiological conditions. The biosensor’s ability to recognize various strains of staphylococci was also successfully demonstrated in the presence of clinical isolates collected from different geographic areas. Assays using S. epidermidis were also carried out to verify the species-specificity of the phage sensor. We only observed a remarkable change of the Rct in the presence of the target S. aureus bacteria, while no substantial binding to S. epidermidis occurred. This confirmed that the Zeno phage sensor only targets S. aureus species within the genus Staphylococcus. In addition, the biosensor's specificity with respect to other bacterial species, including gram-positive bacteria like Enterococcus faecium and the gram-negative bacterium Pseudomonas aeruginosa, was evaluated, and a non-significant impedimetric signal was observed. Notably, the biosensor successfully identified S. aureus bacterial cells in a complex matrix such as a nasal swab, opening the possibility of its use in a real-case scenario. We diluted different concentrations of S. aureus from 108 to 100 CFU/mL with a ratio of 1:10 in the nasal swap matrices collected from healthy donors. Three different sensors were applied to measure various concentrations of bacteria. Our sensor indicated high selectivity to detect S. aureus in biological matrices compared to time-consuming traditional methods, such as enzyme-linked immunosorbent assay (ELISA), polymerase chain reaction (PCR), and radioimmunoassay (RIA), etc. With the aim to study the possibility to use this biosensor to address the challenge associated to pathogen detection, ongoing research is focused on the assessment of the biosensor’s analytical performances in different biological samples and the discovery of new phage bioreceptors.

Keywords: electrochemical impedance spectroscopy, bacteriophage, biosensor, Staphylococcus aureus

Procedia PDF Downloads 47
2950 Production of Neutrons by High Intensity Picosecond Laser Interacting with Thick Solid Target at XingGuangIII

Authors: Xi Yuan, Xuebin Zhu, Bojun Li

Abstract:

This work describes the experiment to produce high-intensity pulsed neutron beams on XingGuangIII laser facility. The high-intensity laser is utilized to drive protons and deuterons, which hit a thick solid target to produce neutrons. The pulse duration of the laser used in the experiment is about 0.8 ps, and the laser energy is around 100 J. Protons and deuterons are accelerated from a 10-μm-thick deuterated polyethylene (CD₂) foil and diagnosed by a Thomson parabola ion-spectrometer. The energy spectrum of neutrons generated via ⁷Li(d,n) and ⁷Li(p,n) reaction when proton and deuteron beams hit a 5-mm-thick LiF target is measured by a scintillator-based time-of-flight spectrometer. Results from the neuron measurements show that the maximum neutron energy is about 12.5 MeV and the neutron yield is up to 2×10⁹/pulse. The high-intensity pulsed neutron beams demonstrated in this work can provide a valuable neutron source for material research, fast neutron induced fission research, and so on.

Keywords: picosecond laser driven, fast neutron, time-of-flight spectrometry, XinggungIII

Procedia PDF Downloads 144
2949 Streptavidin-Biotin Attachment on Modified Silicon Nanowires

Authors: Shalini Singh, Sanjay K. Srivastava, Govind, Mukhtar. A. Khan, P. K. Singh

Abstract:

Nanotechnology is revolutionizing the development of biosensors. Nanomaterials and nanofabrication technologies are increasingly being used to design novel biosensors. Sensitivity and other attributes of biosensors can be improved by using nanomaterials with unique chemical, physical, and mechanical properties in their construction. Silicon is a promising biomaterial that is non-toxic and biodegradable and can be exploited in chemical and biological sensing. Present study demonstrated the streptavidin–biotin interaction on silicon surfaces with different topographies such as flat and nanostructured silicon (nanowires) surfaces. Silicon nanowires with wide range of surface to volume ratio were prepared by electrochemical etching of silicon wafer. The large specific surface of silicon nanowires can be chemically modified to link different molecular probes (DNA strands, enzymes, proteins and so on), which recognize the target analytes, in order to enhance the selectivity and specificity of the sensor device. The interaction of streptavidin with biotin was carried out on 3-aminopropyltriethoxysilane (APTS) functionalized silicon surfaces. Fourier Transform Infrared Spectroscopy (FTIR) and X-ray Photoelectron Spectroscopy (XPS) studies have been performed to characterize the surface characteristics to ensure the protein attachment. Silicon nanowires showed the enhance protein attachment, as compared to flat silicon surface due to its large surface area and good molecular penetration to its surface. The methodology developed herein could be generalized to a wide range of protein-ligand interactions, since it is relatively easy to conjugate biotin with diverse biomolecules such as antibodies, enzymes, peptides, and nucleotides.

Keywords: FTIR, silicon nanowires, streptavidin-biotin, XPS

Procedia PDF Downloads 391
2948 Development of Nondestructive Imaging Analysis Method Using Muonic X-Ray with a Double-Sided Silicon Strip Detector

Authors: I-Huan Chiu, Kazuhiko Ninomiya, Shin’ichiro Takeda, Meito Kajino, Miho Katsuragawa, Shunsaku Nagasawa, Atsushi Shinohara, Tadayuki Takahashi, Ryota Tomaru, Shin Watanabe, Goro Yabu

Abstract:

In recent years, a nondestructive elemental analysis method based on muonic X-ray measurements has been developed and applied for various samples. Muonic X-rays are emitted after the formation of a muonic atom, which occurs when a negatively charged muon is captured in a muon atomic orbit around the nucleus. Because muonic X-rays have higher energy than electronic X-rays due to the muon mass, they can be measured without being absorbed by a material. Thus, estimating the two-dimensional (2D) elemental distribution of a sample became possible using an X-ray imaging detector. In this work, we report a non-destructive imaging experiment using muonic X-rays at Japan Proton Accelerator Research Complex. The irradiated target consisted of polypropylene material, and a double-sided silicon strip detector, which was developed as an imaging detector for astronomical observation, was employed. A peak corresponding to muonic X-rays from the carbon atoms in the target was clearly observed in the energy spectrum at an energy of 14 keV, and 2D visualizations were successfully reconstructed to reveal the projection image from the target. This result demonstrates the potential of the non-destructive elemental imaging method that is based on muonic X-ray measurement. To obtain a higher position resolution for imaging a smaller target, a new detector system will be developed to improve the statistical analysis in further research.

Keywords: DSSD, muon, muonic X-ray, imaging, non-destructive analysis

Procedia PDF Downloads 188