Search results for: intent detection
1921 Deleterious SNP’s Detection Using Machine Learning
Authors: Hamza Zidoum
Abstract:
This paper investigates the impact of human genetic variation on the function of human proteins using machine-learning algorithms. Single-Nucleotide Polymorphism represents the most common form of human genome variation. We focus on the single amino-acid polymorphism located in the coding region as they can affect the protein function leading to pathologic phenotypic change. We use several supervised Machine Learning methods to identify structural properties correlated with increased risk of the missense mutation being damaging. SVM associated with Principal Component Analysis give the best performance.Keywords: single-nucleotide polymorphism, machine learning, feature selection, SVM
Procedia PDF Downloads 3781920 Determination of Four Anions in the Ground Layer of Tomb Murals by Ion Chromatography
Authors: Liping Qiu, Xiaofeng Zhang
Abstract:
The ion chromatography method for the rapid determination of four anions (F⁻、Cl⁻、SO₄²⁻、NO₃⁻) in burial ground poles was optimized. The L₉(₃⁴) orthogonal test was used to determine the optimal parameters of sample pretreatment: accurately weigh 2.000g of sample, add 10mL of ultrapure water, and extract for 40min under the conditions of shaking temperature 40℃ and shaking speed 180 r·min-1. The eluent was 25 mmol/L KOH solution, the analytical column was Ion Pac® AS11-SH (250 mm × 4.0 mm), and the purified filtrate was measured by a conductivity detector. Under this method, the detection limit of each ion is 0.066~0.078mg/kg, the relative standard deviation is 0.86%~2.44% (n=7), and the recovery rate is 94.6~101.9.Keywords: ion chromatography, tomb, anion (F⁻, Cl⁻, SO₄²⁻, NO₃⁻), environmental protection
Procedia PDF Downloads 1021919 Genetic Diversity of Norovirus Strains in Outpatient Children from Rural Communities of Vhembe District, South Africa, 2014-2015
Authors: Jean Pierre Kabue, Emma Meader, Afsatou Ndama Traore, Paul R. Hunter, Natasha Potgieter
Abstract:
Norovirus is now considered the most common cause of outbreaks of nonbacterial gastroenteritis. Limited data are available for Norovirus strains in Africa, especially in rural and peri-urban areas. Despite the excessive burden of diarrhea disease in developing countries, Norovirus infections have been to date mostly reported in developed countries. There is a need to investigate intensively the role of viral agents associated with diarrhea in different settings in Africa continent. To determine the prevalence and genetic diversity of Norovirus strains circulating in the rural communities in the Limpopo Province, South Africa and investigate the genetic relationship between Norovirus strains, a cross-sectional study was performed on human stools collected from rural communities. Between July 2014 and April 2015, outpatient children under 5 years of age from rural communities of Vhembe District, South Africa, were recorded for the study. A total of 303 stool specimens were collected from those with diarrhea (n=253) and without (n=50) diarrhea. NoVs were identified using real-time one-step RT-PCR. Partial Sequence analyses were performed to genotype the strains. Phylogenetic analyses were performed to compare identified NoVs genotypes to the worldwide circulating strains. Norovirus detection rate was 41.1% (104/253) in children with diarrhea. There was no significant difference (OR=1.24; 95% CI 0.66-2.33) in Norovirus detection between symptomatic and asymptomatic children. Comparison of the median CT values for NoV in children with diarrhea and without diarrhea revealed significant statistical difference of estimated GII viral load from both groups, with a much higher viral burden in children with diarrhea. To our knowledge, this is the first study reporting on the differences in estimated viral load of GII and GI NoV positive cases and controls. GII.Pe (n=9) were the predominant genotypes followed by GII.Pe/GII.4 Sydney 2012 (n=8) suspected recombinant and GII.4 Sydney 2012 variants(n=7). Two unassigned GII.4 variants and an unusual RdRp genotype GII.P15 were found. With note, the rare GIIP15 identified in this study has a common ancestor with GIIP15 strain from Japan previously reported as GII/untypeable recombinant strain implicated in a gastroenteritis outbreak. To our knowledge, this is the first report of this unusual genotype in the African continent. Though not confirmed predictive of diarrhea disease in this study, the high detection rate of NoV is an indication of subsequent exposure of children from rural communities to enteric pathogens due to poor sanitation and hygiene practices. The results reveal that the difference between asymptomatic and symptomatic children with NoV may possibly be related to the NoV genogroups involved. The findings emphasize NoV genetic diversity and predominance of GII.Pe/GII.4 Sydney 2012, indicative of increased NoV activity. An uncommon GII.P15 and two unassigned GII.4 variants were also identified from rural settings of the Vhembe District/South Africa. NoV surveillance is required to help to inform investigations into NoV evolution, and to support vaccine development programmes in Africa.Keywords: asymptomatic, common, outpatients, norovirus genetic diversity, sporadic gastroenteritis, South African rural communities, symptomatic
Procedia PDF Downloads 1951918 qPCR Method for Detection of Halal Food Adulteration
Authors: Gabriela Borilova, Monika Petrakova, Petr Kralik
Abstract:
Nowadays, European producers are increasingly interested in the production of halal meat products. Halal meat has been increasingly appearing in the EU's market network and meat products from European producers are being exported to Islamic countries. Halal criteria are mainly related to the origin of muscle used in production, and also to the way products are obtained and processed. Although the EU has legislatively addressed the question of food authenticity, the circumstances of previous years when products with undeclared horse or poultry meat content appeared on EU markets raised the question of the effectiveness of control mechanisms. Replacement of expensive or not-available types of meat for low-priced meat has been on a global scale for a long time. Likewise, halal products may be contaminated (falsified) by pork or food components obtained from pigs. These components include collagen, offal, pork fat, mechanically separated pork, emulsifier, blood, dried blood, dried blood plasma, gelatin, and others. These substances can influence sensory properties of the meat products - color, aroma, flavor, consistency and texture or they are added for preservation and stabilization. Food manufacturers sometimes access these substances mainly due to their dense availability and low prices. However, the use of these substances is not always declared on the product packaging. Verification of the presence of declared ingredients, including the detection of undeclared ingredients, are among the basic control procedures for determining the authenticity of food. Molecular biology methods, based on DNA analysis, offer rapid and sensitive testing. The PCR method and its modification can be successfully used to identify animal species in single- and multi-ingredient raw and processed foods and qPCR is the first choice for food analysis. Like all PCR-based methods, it is simple to implement and its greatest advantage is the absence of post-PCR visualization by electrophoresis. qPCR allows detection of trace amounts of nucleic acids, and by comparing an unknown sample with a calibration curve, it can also provide information on the absolute quantity of individual components in the sample. Our study addresses a problem that is related to the fact that the molecular biological approach of most of the work associated with the identification and quantification of animal species is based on the construction of specific primers amplifying the selected section of the mitochondrial genome. In addition, the sections amplified in conventional PCR are relatively long (hundreds of bp) and unsuitable for use in qPCR, because in DNA fragmentation, amplification of long target sequences is quite limited. Our study focuses on finding a suitable genomic DNA target and optimizing qPCR to reduce variability and distortion of results, which is necessary for the correct interpretation of quantification results. In halal products, the impact of falsification of meat products by the addition of components derived from pigs is all the greater that it is not just about the economic aspect but above all about the religious and social aspect. This work was supported by the Ministry of Agriculture of the Czech Republic (QJ1530107).Keywords: food fraud, halal food, pork, qPCR
Procedia PDF Downloads 2471917 Enhanced Multi-Scale Feature Extraction Using a DCNN by Proposing Dynamic Soft Margin SoftMax for Face Emotion Detection
Authors: Armin Nabaei, M. Omair Ahmad, M. N. S. Swamy
Abstract:
Many facial expression and emotion recognition methods in the traditional approaches of using LDA, PCA, and EBGM have been proposed. In recent years deep learning models have provided a unique platform addressing by automatically extracting the features for the detection of facial expression and emotions. However, deep networks require large training datasets to extract automatic features effectively. In this work, we propose an efficient emotion detection algorithm using face images when only small datasets are available for training. We design a deep network whose feature extraction capability is enhanced by utilizing several parallel modules between the input and output of the network, each focusing on the extraction of different types of coarse features with fined grained details to break the symmetry of produced information. In fact, we leverage long range dependencies, which is one of the main drawback of CNNs. We develop this work by introducing a Dynamic Soft-Margin SoftMax.The conventional SoftMax suffers from reaching to gold labels very soon, which take the model to over-fitting. Because it’s not able to determine adequately discriminant feature vectors for some variant class labels. We reduced the risk of over-fitting by using a dynamic shape of input tensor instead of static in SoftMax layer with specifying a desired Soft- Margin. In fact, it acts as a controller to how hard the model should work to push dissimilar embedding vectors apart. For the proposed Categorical Loss, by the objective of compacting the same class labels and separating different class labels in the normalized log domain.We select penalty for those predictions with high divergence from ground-truth labels.So, we shorten correct feature vectors and enlarge false prediction tensors, it means we assign more weights for those classes with conjunction to each other (namely, “hard labels to learn”). By doing this work, we constrain the model to generate more discriminate feature vectors for variant class labels. Finally, for the proposed optimizer, our focus is on solving weak convergence of Adam optimizer for a non-convex problem. Our noteworthy optimizer is working by an alternative updating gradient procedure with an exponential weighted moving average function for faster convergence and exploiting a weight decay method to help drastically reducing the learning rate near optima to reach the dominant local minimum. We demonstrate the superiority of our proposed work by surpassing the first rank of three widely used Facial Expression Recognition datasets with 93.30% on FER-2013, and 16% improvement compare to the first rank after 10 years, reaching to 90.73% on RAF-DB, and 100% k-fold average accuracy for CK+ dataset, and shown to provide a top performance to that provided by other networks, which require much larger training datasets.Keywords: computer vision, facial expression recognition, machine learning, algorithms, depp learning, neural networks
Procedia PDF Downloads 741916 Evaluation of Antimicrobial Susceptibility Profile of Urinary Tract Infections in Massoud Medical Laboratory: 2018-2021
Authors: Ali Ghorbanipour
Abstract:
The aim of this study is to investigate the drug resistance pattern and the value of the MIC (minimum inhibitory concentration)method to reduce the impact of infectious diseases and the slow development of resistance. Method: The study was conducted on clinical specimens collected between 2018 to 2021. identification of isolates and antibiotic susceptibility testing were performed using conventional biochemical tests. Antibiotic resistance was determined using kibry-Bauer disk diffusion and MIC by E-test methods comparative with microdilution plate elisa method. Results were interpreted according to CLSI. Results: Out of 249600 different clinical specimens, 18720 different pathogenic bacteria by overall detection ratio 7.7% were detected. Among pathogen bacterial were Gram negative bacteria (70%,n=13000) and Gram positive bacteria(30%,n=5720).Medically relevant gram-negative bacteria include a multitude of species such as E.coli , Klebsiella .spp , Pseudomonas .aeroginosa , Acinetobacter .spp , Enterobacterspp ,and gram positive bacteria Staphylococcus.spp , Enterococcus .spp , Streptococcus .spp was isolated . Conclusion: Our results highlighted that the resistance ratio among Gram Negative bacteria and Gram positive bacteria with different infection is high it suggest constant screening and follow-up programs for the detection of antibiotic resistance and the value of MIC drug susceptibility reporting that provide a new way to the usage of resistant antibiotic in combination with other antibiotics or accurate weight of antibiotics that inhibit or kill bacteria. Evaluation of wrong medication in the expansion of resistance and side effects of over usage antibiotics are goals. Ali ghorbanipour presently working as a supervision at the microbiology department of Massoud medical laboratory. Iran. Earlier, he worked as head department of pulmonary infection in firoozgarhospital, Iran. He received master degree in 2012 from Fergusson College. His research prime objective is a biologic wound dressing .to his credit, he has Published10 articles in various international congresses by presenting posters.Keywords: antimicrobial profile, MIC & MBC Method, microplate antimicrobial assay, E-test
Procedia PDF Downloads 1331915 A Comprehensive Survey of Artificial Intelligence and Machine Learning Approaches across Distinct Phases of Wildland Fire Management
Authors: Ursula Das, Manavjit Singh Dhindsa, Kshirasagar Naik, Marzia Zaman, Richard Purcell, Srinivas Sampalli, Abdul Mutakabbir, Chung-Horng Lung, Thambirajah Ravichandran
Abstract:
Wildland fires, also known as forest fires or wildfires, are exhibiting an alarming surge in frequency in recent times, further adding to its perennial global concern. Forest fires often lead to devastating consequences ranging from loss of healthy forest foliage and wildlife to substantial economic losses and the tragic loss of human lives. Despite the existence of substantial literature on the detection of active forest fires, numerous potential research avenues in forest fire management, such as preventative measures and ancillary effects of forest fires, remain largely underexplored. This paper undertakes a systematic review of these underexplored areas in forest fire research, meticulously categorizing them into distinct phases, namely pre-fire, during-fire, and post-fire stages. The pre-fire phase encompasses the assessment of fire risk, analysis of fuel properties, and other activities aimed at preventing or reducing the risk of forest fires. The during-fire phase includes activities aimed at reducing the impact of active forest fires, such as the detection and localization of active fires, optimization of wildfire suppression methods, and prediction of the behavior of active fires. The post-fire phase involves analyzing the impact of forest fires on various aspects, such as the extent of damage in forest areas, post-fire regeneration of forests, impact on wildlife, economic losses, and health impacts from byproducts produced during burning. A comprehensive understanding of the three stages is imperative for effective forest fire management and mitigation of the impact of forest fires on both ecological systems and human well-being. Artificial intelligence and machine learning (AI/ML) methods have garnered much attention in the cyber-physical systems domain in recent times leading to their adoption in decision-making in diverse applications including disaster management. This paper explores the current state of AI/ML applications for managing the activities in the aforementioned phases of forest fire. While conventional machine learning and deep learning methods have been extensively explored for the prevention, detection, and management of forest fires, a systematic classification of these methods into distinct AI research domains is conspicuously absent. This paper gives a comprehensive overview of the state of forest fire research across more recent and prominent AI/ML disciplines, including big data, classical machine learning, computer vision, explainable AI, generative AI, natural language processing, optimization algorithms, and time series forecasting. By providing a detailed overview of the potential areas of research and identifying the diverse ways AI/ML can be employed in forest fire research, this paper aims to serve as a roadmap for future investigations in this domain.Keywords: artificial intelligence, computer vision, deep learning, during-fire activities, forest fire management, machine learning, pre-fire activities, post-fire activities
Procedia PDF Downloads 721914 Developing a Maturity Model of Digital Twin Application for Infrastructure Asset Management
Authors: Qingqing Feng, S. Thomas Ng, Frank J. Xu, Jiduo Xing
Abstract:
Faced with unprecedented challenges including aging assets, lack of maintenance budget, overtaxed and inefficient usage, and outcry for better service quality from the society, today’s infrastructure systems has become the main focus of many metropolises to pursue sustainable urban development and improve resilience. Digital twin, being one of the most innovative enabling technologies nowadays, may open up new ways for tackling various infrastructure asset management (IAM) problems. Digital twin application for IAM, as its name indicated, represents an evolving digital model of intended infrastructure that possesses functions including real-time monitoring; what-if events simulation; and scheduling, maintenance, and management optimization based on technologies like IoT, big data and AI. Up to now, there are already vast quantities of global initiatives of digital twin applications like 'Virtual Singapore' and 'Digital Built Britain'. With digital twin technology permeating the IAM field progressively, it is necessary to consider the maturity of the application and how those institutional or industrial digital twin application processes will evolve in future. In order to deal with the gap of lacking such kind of benchmark, a draft maturity model is developed for digital twin application in the IAM field. Firstly, an overview of current smart cities maturity models is given, based on which the draft Maturity Model of Digital Twin Application for Infrastructure Asset Management (MM-DTIAM) is developed for multi-stakeholders to evaluate and derive informed decision. The process of development follows a systematic approach with four major procedures, namely scoping, designing, populating and testing. Through in-depth literature review, interview and focus group meeting, the key domain areas are populated, defined and iteratively tuned. Finally, the case study of several digital twin projects is conducted for self-verification. The findings of the research reveal that: (i) the developed maturity model outlines five maturing levels leading to an optimised digital twin application from the aspects of strategic intent, data, technology, governance, and stakeholders’ engagement; (ii) based on the case study, levels 1 to 3 are already partially implemented in some initiatives while level 4 is on the way; and (iii) more practices are still needed to refine the draft to be mutually exclusive and collectively exhaustive in key domain areas.Keywords: digital twin, infrastructure asset management, maturity model, smart city
Procedia PDF Downloads 1571913 Culturally Relevant Pedagogy: A Cross-Cultural Comparison
Authors: Medha Talpade, Salil Talpade
Abstract:
The intent of this quantitative project was to compare the values and perceptions of students from a predominantly white college (PWI) to those from a historically black college (HBCU) about culturally relevant teaching and learning practices in the academic realm. The reason for interrelating student culture with teaching practices is to enable a pedagogical response to the low retention rates of African American students and first generation Caucasian students in high schools, colleges, and their low rates of social mobility and educational achievement. Culturally relevant pedagogy, according to related research, is deemed rewarding to students, teachers, the local and national community. Critical race theory (CRT) is the main framework used in this project to explain the ubiquity of a culturally relevant pedagogy. The purpose of this quantitative study was to test the critical race theory that relates the presence of the factors associated with culturally relevant teaching strategies with perceived relevance. The culturally relevant teaching strategies were identified based on the recommendations and findings of past research. Participants in this study included approximately 145 students from a HBCU and 55 students from the PWI. A survey consisting of 37 items related to culturally relevant pedagogy was administered. The themes used to construct the items were: Use of culturally-specific examples in class whenever possible; use of culturally-specific presentational models, use of relational reinforcers, and active engagement. All the items had a likert-type response scale. Participants reported their degree of agreement (5-point scale ranging from strongly disagree to strongly agree) and importance (3-point scale ranging from not at all important to very important) with each survey item. A new variable, Relevance was formed based on the multiplicative function of importance and presence of a teaching and learning strategy. A set of six demographic questions were included in the survey. A consent form based on NIH and APA ethical standards was distributed prior to survey administration to the volunteers. Results of a Factor Analyses on the data from the PWI and the HBCU, and a ANOVA indicated significant differences on ‘Relevance’ related to specific themes. Results of this study are expected to inform educational practices and improve teaching and learning outcomes.Keywords: culturally relevant pedagogy, college students, cross-cultural, applied psychology
Procedia PDF Downloads 4321912 Comprehensive Machine Learning-Based Glucose Sensing from Near-Infrared Spectra
Authors: Bitewulign Mekonnen
Abstract:
Context: This scientific paper focuses on the use of near-infrared (NIR) spectroscopy to determine glucose concentration in aqueous solutions accurately and rapidly. The study compares six different machine learning methods for predicting glucose concentration and also explores the development of a deep learning model for classifying NIR spectra. The objective is to optimize the detection model and improve the accuracy of glucose prediction. This research is important because it provides a comprehensive analysis of various machine-learning techniques for estimating aqueous glucose concentrations. Research Aim: The aim of this study is to compare and evaluate different machine-learning methods for predicting glucose concentration from NIR spectra. Additionally, the study aims to develop and assess a deep-learning model for classifying NIR spectra. Methodology: The research methodology involves the use of machine learning and deep learning techniques. Six machine learning regression models, including support vector machine regression, partial least squares regression, extra tree regression, random forest regression, extreme gradient boosting, and principal component analysis-neural network, are employed to predict glucose concentration. The NIR spectra data is randomly divided into train and test sets, and the process is repeated ten times to increase generalization ability. In addition, a convolutional neural network is developed for classifying NIR spectra. Findings: The study reveals that the SVMR, ETR, and PCA-NN models exhibit excellent performance in predicting glucose concentration, with correlation coefficients (R) > 0.99 and determination coefficients (R²)> 0.985. The deep learning model achieves high macro-averaging scores for precision, recall, and F1-measure. These findings demonstrate the effectiveness of machine learning and deep learning methods in optimizing the detection model and improving glucose prediction accuracy. Theoretical Importance: This research contributes to the field by providing a comprehensive analysis of various machine-learning techniques for estimating glucose concentrations from NIR spectra. It also explores the use of deep learning for the classification of indistinguishable NIR spectra. The findings highlight the potential of machine learning and deep learning in enhancing the prediction accuracy of glucose-relevant features. Data Collection and Analysis Procedures: The NIR spectra and corresponding references for glucose concentration are measured in increments of 20 mg/dl. The data is randomly divided into train and test sets, and the models are evaluated using regression analysis and classification metrics. The performance of each model is assessed based on correlation coefficients, determination coefficients, precision, recall, and F1-measure. Question Addressed: The study addresses the question of whether machine learning and deep learning methods can optimize the detection model and improve the accuracy of glucose prediction from NIR spectra. Conclusion: The research demonstrates that machine learning and deep learning methods can effectively predict glucose concentration from NIR spectra. The SVMR, ETR, and PCA-NN models exhibit superior performance, while the deep learning model achieves high classification scores. These findings suggest that machine learning and deep learning techniques can be used to improve the prediction accuracy of glucose-relevant features. Further research is needed to explore their clinical utility in analyzing complex matrices, such as blood glucose levels.Keywords: machine learning, signal processing, near-infrared spectroscopy, support vector machine, neural network
Procedia PDF Downloads 941911 Detection and Molecular Identification of Bacteria Forming Polyhydroxyalkanoate and Polyhydroxybutyrate Isolated from Soil in Saudi Arabia
Authors: Ali Bahkali, Rayan Yousef Booq, Mohammad Khiyami
Abstract:
Soil samples were collected from five different regions in the Kingdom of Saudi Arabia. Microbiological methods included dilution methods and pour plates to isolate and purify bacteria soil. The ability of isolates to develop biopolymer was investigated on petri dishes containing elements and substance concentrations stimulating developing biopolymer. Fluorescent stains, Nile red and Nile blue were used to stain the bacterial cells developing biopolymers. In addition, Sudan black was used to detect biopolymers in bacterial cells. The isolates which developed biopolymers were identified based on their gene sequence of 1 6sRNA and their ability to grow and synthesize PHAs on mineral medium supplemented with 1% dates molasses as the only carbon source under nitrogen limitation. During the study 293 bacterial isolates were isolated and detected. Through the initial survey on the petri dishes, 84 isolates showed the ability to develop biopolymers. These bacterial colonies developed a pink color due to accumulation of the biopolymers in the cells. Twenty-three isolates were able to grow on dates molasses, three strains of which showed the ability to accumulate biopolymers. These strains included Bacillus sp., Ralstonia sp. and Microbacterium sp. They were detected by Nile blue A stain with fluorescence microscopy (OLYMPUS IX 51). Among the isolated strains Ralstonia sp. was selected after its ability to grow on molasses dates in the presence of a limited nitrogen source was detected. The optimum conditions for formation of biopolymers by isolated strains were investigated. Conditions studied included, best incubation duration (2 days), temperature (30°C) and pH (7-8). The maximum PHB production was raised by 1% (v1v) when using concentrations of dates molasses 1, 2, 3, 4 and 5% in MSM. The best inoculated with 1% old inoculum (1= OD). The ideal extraction method of PHA and PHB proved to be 0.4% sodium hypochlorite solution, producing a quantity of polymer 98.79% of the cell's dry weight. The maximum PHB production was 1.79 g/L recorded by Ralstonia sp. after 48 h, while it was 1.40 g/L produced by R.eutropha ATCC 17697 after 48 h.Keywords: bacteria forming polyhydroxyalkanoate, detection, molecular, Saudi Arabia
Procedia PDF Downloads 3471910 Snapchat’s Scanning Feature
Authors: Reham Banwair, Lana Alshehri, Sara Hadrawi
Abstract:
The purpose of this project is to identify user satisfaction with the AI functions on Snapchat, in order to generate improvement proposals that allow its development within the app. To achieve this, a qualitative analysis was carried out through interviews to people who usually use the application, revealing their satisfaction or dissatisfaction with the usefulness of the AI. In addition, the background of the company and its introduction in these algorithms were analyzed. Furthermore, the characteristics of the three main functions of AI were explained: identify songs, solve mathematical problems, and recognize plants. As a result, it was obtained that 50% still do not know the characteristics of AI, 50% still believe song recognition is not always correct, 41.7% believe that math problems are usually accurate and 91.7% believes the plant detection tool is working properly.Keywords: artificial intelligence, scanning, Snapchat, machine learning
Procedia PDF Downloads 1341909 Visitor Management in the National Parks: Recreational Carrying Capacity Assessment of Çıralı Coast, Turkey
Authors: Tendü H. Göktuğ, Gönül T. İçemer, Bülent Deniz
Abstract:
National parks, which are rich in natural and cultural resources values are protected in the context of the idea to develop sustainability, are among the most important recreated areas demanding with each passing day. Increasing recreational use or unplanned use forms negatively affect the resource values and visitor satisfaction. The intent of national parks management is to protect the natural and cultural resource values and to provide the visitors with a quality of recreational experience, as well. In this context, the current studies to improve the appropriate tourism and recreation planning and visitor management, approach have focused on recreational carrying capacity analysis. The aim of this study is to analyze recreational carrying capacity of Çıralı Coast in the Bey Mountains Coastal National Park to compare the analyze results with the current usage format and to develop alternative management strategies. In the first phase of the study, the annual and daily visitations, geographic, bio-physical, and managerial characteristics of the park and the type of recreational usage and the recreational areas were analyzed. In addition to these, ecological observations were carried out in order to determine recreational-based pressures on the ecosystems. On-site questionnaires were administrated to a sample of 284 respondents in the August 2015 - 2016 to collect data concerning the demographics and visit characteristics. The second phase of the study, the coastal area separated into four different usage zones and the methodology proposed by Cifuentes (1992) was used for capacity analyses. This method supplies the calculation of physical, real and effective carrying capacities by using environmental, ecological, climatic and managerial parameters in a formulation. Expected numbers which estimated three levels of carrying capacities were compared to current numbers of national parks’ visitors. In the study, it was determined that the current recreational uses in the north of the beach were caused by ecological pressures, and the current numbers in the south of beach much more than estimated numbers of visitors. Based on these results management strategies were defined and the appropriate management tools were developed in accordance with these strategies. The authors are grateful for the financial support of this project by The Scientific and Technological Research Council of Turkey (No: 114O344)Keywords: Çıralı Coast, national parks, recreational carrying capacity, visitor management
Procedia PDF Downloads 2741908 A Geosynchronous Orbit Synthetic Aperture Radar Simulator for Moving Ship Targets
Authors: Linjie Zhang, Baifen Ren, Xi Zhang, Genwang Liu
Abstract:
Ship detection is of great significance for both military and civilian applications. Synthetic aperture radar (SAR) with all-day, all-weather, ultra-long-range characteristics, has been used widely. In view of the low time resolution of low orbit SAR and the needs for high time resolution SAR data, GEO (Geosynchronous orbit) SAR is getting more and more attention. Since GEO SAR has short revisiting period and large coverage area, it is expected to be well utilized in marine ship targets monitoring. However, the height of the orbit increases the time of integration by almost two orders of magnitude. For moving marine vessels, the utility and efficacy of GEO SAR are still not sure. This paper attempts to find the feasibility of GEO SAR by giving a GEO SAR simulator of moving ships. This presented GEO SAR simulator is a kind of geometrical-based radar imaging simulator, which focus on geometrical quality rather than high radiometric. Inputs of this simulator are 3D ship model (.obj format, produced by most 3D design software, such as 3D Max), ship's velocity, and the parameters of satellite orbit and SAR platform. Its outputs are simulated GEO SAR raw signal data and SAR image. This simulating process is accomplished by the following four steps. (1) Reading 3D model, including the ship rotations (pitch, yaw, and roll) and velocity (speed and direction) parameters, extract information of those little primitives (triangles) which is visible from the SAR platform. (2) Computing the radar scattering from the ship with physical optics (PO) method. In this step, the vessel is sliced into many little rectangles primitives along the azimuth. The radiometric calculation of each primitive is carried out separately. Since this simulator only focuses on the complex structure of ships, only single-bounce reflection and double-bounce reflection are considered. (3) Generating the raw data with GEO SAR signal modeling. Since the normal ‘stop and go’ model is not available for GEO SAR, the range model should be reconsidered. (4) At last, generating GEO SAR image with improved Range Doppler method. Numerical simulation of fishing boat and cargo ship will be given. GEO SAR images of different posture, velocity, satellite orbit, and SAR platform will be simulated. By analyzing these simulated results, the effectiveness of GEO SAR for the detection of marine moving vessels is evaluated.Keywords: GEO SAR, radar, simulation, ship
Procedia PDF Downloads 1771907 Smart Defect Detection in XLPE Cables Using Convolutional Neural Networks
Authors: Tesfaye Mengistu
Abstract:
Power cables play a crucial role in the transmission and distribution of electrical energy. As the electricity generation, transmission, distribution, and storage systems become smarter, there is a growing emphasis on incorporating intelligent approaches to ensure the reliability of power cables. Various types of electrical cables are employed for transmitting and distributing electrical energy, with cross-linked polyethylene (XLPE) cables being widely utilized due to their exceptional electrical and mechanical properties. However, insulation defects can occur in XLPE cables due to subpar manufacturing techniques during production and cable joint installation. To address this issue, experts have proposed different methods for monitoring XLPE cables. Some suggest the use of interdigital capacitive (IDC) technology for online monitoring, while others propose employing continuous wave (CW) terahertz (THz) imaging systems to detect internal defects in XLPE plates used for power cable insulation. In this study, we have developed models that employ a custom dataset collected locally to classify the physical safety status of individual power cables. Our models aim to replace physical inspections with computer vision and image processing techniques to classify defective power cables from non-defective ones. The implementation of our project utilized the Python programming language along with the TensorFlow package and a convolutional neural network (CNN). The CNN-based algorithm was specifically chosen for power cable defect classification. The results of our project demonstrate the effectiveness of CNNs in accurately classifying power cable defects. We recommend the utilization of similar or additional datasets to further enhance and refine our models. Additionally, we believe that our models could be used to develop methodologies for detecting power cable defects from live video feeds. We firmly believe that our work makes a significant contribution to the field of power cable inspection and maintenance. Our models offer a more efficient and cost-effective approach to detecting power cable defects, thereby improving the reliability and safety of power grids.Keywords: artificial intelligence, computer vision, defect detection, convolutional neural net
Procedia PDF Downloads 1121906 Comparison of Monte Carlo Simulations and Experimental Results for the Measurement of Complex DNA Damage Induced by Ionizing Radiations of Different Quality
Authors: Ifigeneia V. Mavragani, Zacharenia Nikitaki, George Kalantzis, George Iliakis, Alexandros G. Georgakilas
Abstract:
Complex DNA damage consisting of a combination of DNA lesions, such as Double Strand Breaks (DSBs) and non-DSB base lesions occurring in a small volume is considered as one of the most important biological endpoints regarding ionizing radiation (IR) exposure. Strong theoretical (Monte Carlo simulations) and experimental evidence suggests an increment of the complexity of DNA damage and therefore repair resistance with increasing linear energy transfer (LET). Experimental detection of complex (clustered) DNA damage is often associated with technical deficiencies limiting its measurement, especially in cellular or tissue systems. Our groups have recently made significant improvements towards the identification of key parameters relating to the efficient detection of complex DSBs and non-DSBs in human cellular systems exposed to IR of varying quality (γ-, X-rays 0.3-1 keV/μm, α-particles 116 keV/μm and 36Ar ions 270 keV/μm). The induction and processing of DSB and non-DSB-oxidative clusters were measured using adaptations of immunofluorescence (γH2AX or 53PB1 foci staining as DSB probes and human repair enzymes OGG1 or APE1 as probes for oxidized purines and abasic sites respectively). In the current study, Relative Biological Effectiveness (RBE) values for DSB and non-DSB induction have been measured in different human normal (FEP18-11-T1) and cancerous cell lines (MCF7, HepG2, A549, MO59K/J). The experimental results are compared to simulation data obtained using a validated microdosimetric fast Monte Carlo DNA Damage Simulation code (MCDS). Moreover, this simulation approach is implemented in two realistic clinical cases, i.e. prostate cancer treatment using X-rays generated by a linear accelerator and a pediatric osteosarcoma case using a 200.6 MeV proton pencil beam. RBE values for complex DNA damage induction are calculated for the tumor areas. These results reveal a disparity between theory and experiment and underline the necessity for implementing highly precise and more efficient experimental and simulation approaches.Keywords: complex DNA damage, DNA damage simulation, protons, radiotherapy
Procedia PDF Downloads 3251905 Encoded Nanospheres for the Fast Ratiometric Detection of Cystic Fibrosis
Authors: Iván Castelló, Georgiana Stoica, Emilio Palomares, Fernando Bravo
Abstract:
We present herein two colour encoded silica nanospheres (2nanoSi) for the fluorescence quantitative ratiometric determination of trypsin in humans. The system proved to be a faster (minutes) method, with two times higher sensitivity than the state-of-the-art biomarkers based sensors for cystic fibrosis (CF), allowing the quantification of trypsin concentrations in a wide range (0-350 mg/L). Furthermore, as trypsin is directly related to the development of cystic fibrosis, different human genotypes, i.e. healthy homozygotic (> 80 mg/L), CF homozygotic (< 50 mg/L), and heterozygotic (> 50 mg/L), respectively, can be determined using our 2nanoSi nanospheres.Keywords: cystic fibrosis, trypsin, quantum dots, biomarker, homozygote, heterozygote
Procedia PDF Downloads 4841904 Comparative Analysis of Feature Extraction and Classification Techniques
Authors: R. L. Ujjwal, Abhishek Jain
Abstract:
In the field of computer vision, most facial variations such as identity, expression, emotions and gender have been extensively studied. Automatic age estimation has been rarely explored. With age progression of a human, the features of the face changes. This paper is providing a new comparable study of different type of algorithm to feature extraction [Hybrid features using HAAR cascade & HOG features] & classification [KNN & SVM] training dataset. By using these algorithms we are trying to find out one of the best classification algorithms. Same thing we have done on the feature selection part, we extract the feature by using HAAR cascade and HOG. This work will be done in context of age group classification model.Keywords: computer vision, age group, face detection
Procedia PDF Downloads 3681903 Best-Performing Color Space for Land-Sea Segmentation Using Wavelet Transform Color-Texture Features and Fusion of over Segmentation
Authors: Seynabou Toure, Oumar Diop, Kidiyo Kpalma, Amadou S. Maiga
Abstract:
Color and texture are the two most determinant elements for perception and recognition of the objects in an image. For this reason, color and texture analysis find a large field of application, for example in image classification and segmentation. But, the pioneering work in texture analysis was conducted on grayscale images, thus discarding color information. Many grey-level texture descriptors have been proposed and successfully used in numerous domains for image classification: face recognition, industrial inspections, food science medical imaging among others. Taking into account color in the definition of these descriptors makes it possible to better characterize images. Color texture is thus the subject of recent work, and the analysis of color texture images is increasingly attracting interest in the scientific community. In optical remote sensing systems, sensors measure separately different parts of the electromagnetic spectrum; the visible ones and even those that are invisible to the human eye. The amounts of light reflected by the earth in spectral bands are then transformed into grayscale images. The primary natural colors Red (R) Green (G) and Blue (B) are then used in mixtures of different spectral bands in order to produce RGB images. Thus, good color texture discrimination can be achieved using RGB under controlled illumination conditions. Some previous works investigate the effect of using different color space for color texture classification. However, the selection of the best performing color space in land-sea segmentation is an open question. Its resolution may bring considerable improvements in certain applications like coastline detection, where the detection result is strongly dependent on the performance of the land-sea segmentation. The aim of this paper is to present the results of a study conducted on different color spaces in order to show the best-performing color space for land-sea segmentation. In this sense, an experimental analysis is carried out using five different color spaces (RGB, XYZ, Lab, HSV, YCbCr). For each color space, the Haar wavelet decomposition is used to extract different color texture features. These color texture features are then used for Fusion of Over Segmentation (FOOS) based classification; this allows segmentation of the land part from the sea one. By analyzing the different results of this study, the HSV color space is found as the best classification performance while using color and texture features; which is perfectly coherent with the results presented in the literature.Keywords: classification, coastline, color, sea-land segmentation
Procedia PDF Downloads 2471902 Optical Imaging Based Detection of Solder Paste in Printed Circuit Board Jet-Printing Inspection
Authors: D. Heinemann, S. Schramm, S. Knabner, D. Baumgarten
Abstract:
Purpose: Applying solder paste to printed circuit boards (PCB) with stencils has been the method of choice over the past years. A new method uses a jet printer to deposit tiny droplets of solder paste through an ejector mechanism onto the board. This allows for more flexible PCB layouts with smaller components. Due to the viscosity of the solder paste, air blisters can be trapped in the cartridge. This can lead to missing solder joints or deviations in the applied solder volume. Therefore, a built-in and real-time inspection of the printing process is needed to minimize uncertainties and increase the efficiency of the process by immediate correction. The objective of the current study is the design of an optimal imaging system and the development of an automatic algorithm for the detection of applied solder joints from optical from the captured images. Methods: In a first approach, a camera module connected to a microcomputer and LED strips are employed to capture images of the printed circuit board under four different illuminations (white, red, green and blue). Subsequently, an improved system including a ring light, an objective lens, and a monochromatic camera was set up to acquire higher quality images. The obtained images can be divided into three main components: the PCB itself (i.e., the background), the reflections induced by unsoldered positions or screw holes and the solder joints. Non-uniform illumination is corrected by estimating the background using a morphological opening and subtraction from the input image. Image sharpening is applied in order to prevent error pixels in the subsequent segmentation. The intensity thresholds which divide the main components are obtained from the multimodal histogram using three probability density functions. Determining the intersections delivers proper thresholds for the segmentation. Remaining edge gradients produces small error areas which are removed by another morphological opening. For quantitative analysis of the segmentation results, the dice coefficient is used. Results: The obtained PCB images show a significant gradient in all RGB channels, resulting from ambient light. Using different lightings and color channels 12 images of a single PCB are available. A visual inspection and the investigation of 27 specific points show the best differentiation between those points using a red lighting and a green color channel. Estimating two thresholds from analyzing the multimodal histogram of the corrected images and using them for segmentation precisely extracts the solder joints. The comparison of the results to manually segmented images yield high sensitivity and specificity values. Analyzing the overall result delivers a Dice coefficient of 0.89 which varies for single object segmentations between 0.96 for a good segmented solder joints and 0.25 for single negative outliers. Conclusion: Our results demonstrate that the presented optical imaging system and the developed algorithm can robustly detect solder joints on printed circuit boards. Future work will comprise a modified lighting system which allows for more precise segmentation results using structure analysis.Keywords: printed circuit board jet-printing, inspection, segmentation, solder paste detection
Procedia PDF Downloads 3361901 Hybrid MIMO-OFDM Detection Scheme for High Performance
Authors: Young-Min Ko, Dong-Hyun Ha, Chang-Bin Ha, Hyoung-Kyu Song
Abstract:
In recent years, a multi-antenna system is actively used to improve the performance of the communication. A MIMO-OFDM system can provide multiplexing gain or diversity gain. These gains are obtained in proportion to the increase of the number of antennas. In order to provide the optimal gain of the MIMO-OFDM system, various transmission and reception schemes are presented. This paper aims to propose a hybrid scheme that base station provides both diversity gain and multiplexing gain at the same time.Keywords: DFE, diversity gain, hybrid, MIMO, multiplexing gain.
Procedia PDF Downloads 6851900 The Development Status of Terahertz Wave and Its Prospect in Wireless Communication
Authors: Yiquan Liao, Quanhong Jiang
Abstract:
Since terahertz was observed by German scientists, we have obtained terahertz through different generation technologies of broadband and narrowband. Then, with the development of semiconductor and other technologies, the imaging technology of terahertz has become increasingly perfect. From the earliest application of nondestructive testing in aviation to the present application of information transmission and human safety detection, the role of terahertz will shine in various fields. The weapons produced by terahertz were epoch-making, which is a crushing deterrent against technologically backward countries. At the same time, terahertz technology in the fields of imaging, medical and livelihood, communication and communication are for the well-being of the country and the people.Keywords: terahertz, imaging, communication, medical treatment
Procedia PDF Downloads 991899 Tip-Enhanced Raman Spectroscopy with Plasmonic Lens Focused Longitudinal Electric Field Excitation
Authors: Mingqian Zhang
Abstract:
Tip-enhanced Raman spectroscopy (TERS) is a scanning probe technique for individual objects and structured surfaces investigation that provides a wealth of enhanced spectral information with nanoscale spatial resolution and high detection sensitivity. It has become a powerful and promising chemical and physical information detection method in the nanometer scale. The TERS technique uses a sharp metallic tip regulated in the near-field of a sample surface, which is illuminated with a certain incident beam meeting the excitation conditions of the wave-vector matching. The local electric field, and, consequently, the Raman scattering, from the sample in the vicinity of the tip apex are both greatly tip-enhanced owning to the excitation of localized surface plasmons and the lightning-rod effect. Typically, a TERS setup is composed of a scanning probe microscope, excitation and collection optical configurations, and a Raman spectroscope. In the illumination configuration, an objective lens or a parabolic mirror is always used as the most important component, in order to focus the incident beam on the tip apex for excitation. In this research, a novel TERS setup was built up by introducing a plasmonic lens to the excitation optics as a focusing device. A plasmonic lens with symmetry breaking semi-annular slits corrugated on gold film was designed for the purpose of generating concentrated sub-wavelength light spots with strong longitudinal electric field. Compared to conventional far-field optical components, the designed plasmonic lens not only focuses an incident beam to a sub-wavelength light spot, but also realizes a strong z-component that dominants the electric field illumination, which is ideal for the excitation of tip-enhancement. Therefore, using a PL in the illumination configuration of TERS contributes to improve the detection sensitivity by both reducing the far-field background and effectively exciting the localized electric field enhancement. The FDTD method was employed to investigate the optical near-field distribution resulting from the light-nanostructure interaction. And the optical field distribution was characterized using an scattering-type scanning near-field optical microscope to demonstrate the focusing performance of the lens. The experimental result is in agreement with the theoretically calculated one. It verifies the focusing performance of the plasmonic lens. The optical field distribution shows a bright elliptic spot in the lens center and several arc-like side-lobes on both sides. After the focusing performance was experimentally verified, the designed plasmonic lens was used as a focusing component in the excitation configuration of TERS setup to concentrate incident energy and generate a longitudinal optical field. A collimated linearly polarized laser beam, with along x-axis polarization, was incident from the bottom glass side on the plasmonic lens. The incident light focused by the plasmonic lens interacted with the silver-coated tip apex and enhanced the Raman signal of the sample locally. The scattered Raman signal was gathered by a parabolic mirror and detected with a Raman spectroscopy. Then, the plasmonic lens based setup was employed to investigate carbon nanotubes and TERS experiment was performed. Experimental results indicate that the Raman signal is considerably enhanced which proves that the novel TERS configuration is feasible and promising.Keywords: longitudinal electric field, plasmonics, raman spectroscopy, tip-enhancement
Procedia PDF Downloads 3731898 Engineering the Topological Insulator Structures for Terahertz Detectors
Authors: M. Marchewka
Abstract:
The article is devoted to the possible optical transitions in double quantum wells system based on HgTe/HgCd(Mn)Te heterostructures. Such structures can find applications as detectors and sources of radiation in the terahertz range. The Double Quantum Wells (DQW) systems consist of two QWs separated by the transparent for electrons barrier. Such systems look promising from the point of view of the additional degrees of freedom. In the case of the topological insulator in about 6.4nm wide HgTe QW or strained 3D HgTe films at the interfaces, the topologically protected surface states appear at the interfaces/surfaces. Electrons in those edge states move along the interfaces/surfaces without backscattering due to time-reversal symmetry. Combination of the topological properties, which was already verified by the experimental way, together with the very well know properties of the DQWs, can be very interesting from the applications point of view, especially in the THz area. It is important that at the present stage, the technology makes it possible to create high-quality structures of this type, and intensive experimental and theoretical studies of their properties are already underway. The idea presented in this paper is based on the eight-band KP model, including the additional terms related to the structural inversion asymmetry, interfaces inversion asymmetry, the influence of the magnetically content, and the uniaxial strain describe the full pictures of the possible real structure. All of this term, together with the external electric field, can be sources of breaking symmetry in investigated materials. Using the 8 band KP model, we investigated the electronic shape structure with and without magnetic field from the application point of view as a THz detector in a small magnetic field (below 2T). We believe that such structures are the way to get the tunable topological insulators and the multilayer topological insulator. Using the one-dimensional electrons at the topologically protected interface states as fast and collision-free signal carriers as charge and signal carriers, the detection of the optical signal should be fast, which is very important in the high-resolution detection of signals in the THz range. The proposed engineering of the investigated structures is now one of the important steps on the way to get the proper structures with predicted properties.Keywords: topological insulator, THz spectroscopy, KP model, II-VI compounds
Procedia PDF Downloads 1221897 Riesz Mixture Model for Brain Tumor Detection
Authors: Mouna Zitouni, Mariem Tounsi
Abstract:
This research introduces an application of the Riesz mixture model for medical image segmentation for accurate diagnosis and treatment of brain tumors. We propose a pixel classification technique based on the Riesz distribution, derived from an extended Bartlett decomposition. To our knowledge, this is the first study addressing this approach. The Expectation-Maximization algorithm is implemented for parameter estimation. A comparative analysis, using both synthetic and real brain images, demonstrates the superiority of the Riesz model over a recent method based on the Wishart distribution.Keywords: EM algorithm, segmentation, Riesz probability distribution, Wishart probability distribution
Procedia PDF Downloads 171896 The Role of Accounting and Auditing in Anti-Corruption Strategies: The Case of ECOWAS
Authors: Edna Gnomblerou
Abstract:
Given the current scale of corruption epidemic in West African economies, governments are seeking for immediate and effective measures to reduce the likelihood of the plague within the region. Generally, accountants and auditors are expected to help organizations in detecting illegal practices. However, their role in the fight against corruption is sometimes limited due to the collusive nature of corruption. The Denmark anti-corruption model shows that the implementation of additional controls over public accounts and independent efficient audits improve transparency and increase the probability of detection. This study is aimed at reviewing the existing anti-corruption policies of the Economic Commission of West African States (ECOWAS) as to observe the role attributed to accounting, auditing and other managerial practices in their anti-corruption drive. It further discusses the usefulness of accounting and auditing in helping anti-corruption commissions in controlling misconduct and increasing the perception to detect irregularities within public administration. The purpose of this initiative is to identify and assess the relevance of accounting and auditing in curbing corruption. To meet this purpose, the study was designed to answer the questions of whether accounting and auditing processes were included in the reviewed anti-corruption strategies, and if yes, whether they were effective in the detection process. A descriptive research method was adopted in examining the role of accounting and auditing in West African anti-corruption strategies. The analysis reveals that proper recognition of accounting standards and implementation of financial audits are viewed as strategic mechanisms in tackling corruption. Additionally, codes of conduct, whistle-blowing and information disclosure to the public are among the most common managerial practices used throughout anti-corruption policies to effectively and efficiently address the problem. These observations imply that sound anti-corruption strategies cannot ignore the values of including accounting and auditing processes. On one hand, this suggests that governments should employ all resources possible to improve accounting and auditing practices in the management of public sector organizations. On the other hand, governments must ensure that accounting and auditing practices are not limited to the private sector, but when properly implemented constitute crucial mechanisms to control and reduce corrupt incentives in public sector.Keywords: accounting, anti-corruption strategy, auditing, ECOWAS
Procedia PDF Downloads 2551895 Risk of Occupational Exposure to Cytotoxic Drugs: The Role of Handling Procedures of Hospital Workers
Authors: J. Silva, P. Arezes, R. Schierl, N. Costa
Abstract:
In order to study environmental contamination by cytostatic drugs in Portugal hospitals, sampling campaigns were conducted in three hospitals in 2015 (112 samples). Platinum containing drugs and fluorouracil were chosen because both were administered in high amounts. The detection limit was 0.01 pg/cm² for platinum and 0.1 pg/cm² for fluorouracil. The results show that spills occur mainly on the patient`s chair, while the most referenced occurrence is due to an inadequately closed wrapper. Day hospitals facilities were detected as having the largest number of contaminated samples and with higher levels of contamination.Keywords: cytostatic, contamination, hospital, procedures, handling
Procedia PDF Downloads 2951894 Forensic Detection of Errors Permitted by the Witnesses in Their Testimony
Authors: Lev Bertovsky
Abstract:
The purpose of this study was to determine the reasons for the formation of false testimony from witnesses and make recommendations on the recognition of such cases. During the studies, which were based on the achievements of professionals in the field of psychology, as well as personal investigative practice, the stages of perception of the information were studied, as well as the process of its reclaim from the memory and transmission to the communicator upon request. Based on the principles of the human brain, kinds of conscientious witness mistakes were systematized. Proposals were formulated for the optimization of investigative actions in cases where the witnesses make an honest mistake with respect to the effects previously observed by them.Keywords: criminology, eyewitness testimony, honest mistake, information, investigator, investigation, questioning
Procedia PDF Downloads 1851893 Clinico-Microbiological Study of S. aureus from Various Clinical Samples with Reference to Methicillin Resistant S. aureus (MRSA)
Authors: T. G. Pathrikar, A. D. Urhekar, M. P. Bansal
Abstract:
To find out S. aureus from patient samples on the basis of coagulase test. We have evaluated slide coagulase (n=46 positive), tube coagulase (n=48 positive) and DNase test (n=44, positive) , We have isolated and identified MRSA from various clinical samples and specimens by disc diffusion method determined the incidence of MRSA 50% in patients. Found out the in vitro antimicrobial susceptibility pattern of MRSA isolates and also the MIC of MRSA of oxacillin by E-Test.Keywords: cefoxitin disc diffusion MRSA detection, e – test, S. aureus devastating pathogen, tube coagulase confirmation
Procedia PDF Downloads 4911892 Stretchable and Flexible Thermoelectric Polymer Composites for Self-Powered Volatile Organic Compound Vapors Detection
Authors: Petr Slobodian, Pavel Riha, Jiri Matyas, Robert Olejnik, Nuri Karakurt
Abstract:
Thermoelectric devices generate an electrical current when there is a temperature gradient between the hot and cold junctions of two dissimilar conductive materials typically n-type and p-type semiconductors. Consequently, also the polymeric semiconductors composed of polymeric matrix filled by different forms of carbon nanotubes with proper structural hierarchy can have thermoelectric properties which temperature difference transfer into electricity. In spite of lower thermoelectric efficiency of polymeric thermoelectrics in terms of the figure of merit, the properties as stretchability, flexibility, lightweight, low thermal conductivity, easy processing, and low manufacturing cost are advantages in many technological and ecological applications. Polyethylene-octene copolymer based highly elastic composites filled with multi-walled carbon nanotubes (MWCTs) were prepared by sonication of nanotube dispersion in a copolymer solution followed by their precipitation pouring into non-solvent. The electronic properties of MWCNTs were moderated by different treatment techniques such as chemical oxidation, decoration by Ag clusters or addition of low molecular dopants. In this concept, for example, the amounts of oxygenated functional groups attached on MWCNT surface by HNO₃ oxidation increase p-type charge carriers. p-type of charge carriers can be further increased by doping with molecules of triphenylphosphine. For partial altering p-type MWCNTs into less p-type ones, Ag nanoparticles were deposited on MWCNT surface and then doped with 7,7,8,8-tetracyanoquino-dimethane. Both types of MWCNTs with the highest difference in generated thermoelectric power were combined to manufacture polymeric based thermoelectric module generating thermoelectric voltage when the temperature difference is applied between hot and cold ends of the module. Moreover, it was found that the generated voltage by the thermoelectric module at constant temperature gradient was significantly affected when exposed to vapors of different volatile organic compounds representing then a self-powered thermoelectric sensor for chemical vapor detection.Keywords: carbon nanotubes, polymer composites, thermoelectric materials, self-powered gas sensor
Procedia PDF Downloads 153