Search results for: aerial imaging and detection
3069 Directional Search for Dark Matter Using Nuclear Emulsion
Authors: Ali Murat Guler
Abstract:
A variety of experiments have been developed over the past decades, aiming at the detection of Weakly Interactive Massive Particles (WIMPs) via their scattering in an instrumented medium. The sensitivity of these experiments has improved with a tremendous speed, thanks to a constant development of detectors and analysis methods. Detectors capable of reconstructing the direction of the nuclear recoil induced by the WIMP scattering are opening a new frontier to possibly extend Dark Matter searches beyond the neutrino background. Measurement of WIMP’s direction will allow us to detect the galactic origin of dark matter and, therefore to have a clear signal-background separation. The NEWSdm experiment, based on nuclear emulsions, is intended to measure the direction of WIMP-induced nuclear coils with a solid-state detector, thus with high sensitivity. We discuss the discovery potential of a directional experiment based on the use of a solid target made of newly developed nuclear emulsions and novel read-out systems achieving nanometric resolution. We also report results of a technical test conducted in Gran Sasso.Keywords: dark matter, direct detection, nuclear emulsion, WIMPS
Procedia PDF Downloads 2723068 Computer Aided Diagnostic System for Detection and Classification of a Brain Tumor through MRI Using Level Set Based Segmentation Technique and ANN Classifier
Authors: Atanu K Samanta, Asim Ali Khan
Abstract:
Due to the acquisition of huge amounts of brain tumor magnetic resonance images (MRI) in clinics, it is very difficult for radiologists to manually interpret and segment these images within a reasonable span of time. Computer-aided diagnosis (CAD) systems can enhance the diagnostic capabilities of radiologists and reduce the time required for accurate diagnosis. An intelligent computer-aided technique for automatic detection of a brain tumor through MRI is presented in this paper. The technique uses the following computational methods; the Level Set for segmentation of a brain tumor from other brain parts, extraction of features from this segmented tumor portion using gray level co-occurrence Matrix (GLCM), and the Artificial Neural Network (ANN) to classify brain tumor images according to their respective types. The entire work is carried out on 50 images having five types of brain tumor. The overall classification accuracy using this method is found to be 98% which is significantly good.Keywords: brain tumor, computer-aided diagnostic (CAD) system, gray-level co-occurrence matrix (GLCM), tumor segmentation, level set method
Procedia PDF Downloads 5123067 Cut-Off of CMV Cobas® Taqman® (CAP/CTM Roche®) for Introduction of Ganciclovir Pre-Emptive Therapy in Allogeneic Hematopoietic Stem Cell Transplant Recipients
Authors: B. B. S. Pereira, M. O. Souza, L. P. Zanetti, L. C. S. Oliveira, J. R. P. Moreno, M. P. Souza, V. R. Colturato, C. M. Machado
Abstract:
Background: The introduction of prophylactic or preemptive therapies has effectively decreased the CMV mortality rates after hematopoietic stem cell transplantation (HSCT). CMV antigenemia (pp65) or quantitative PCR are methods currently approved for CMV surveillance in pre-emptive strategies. Commercial assays are preferred as cut-off levels defined by in-house assays may vary among different protocols and in general show low reproducibility. Moreover, comparison of published data among different centers is only possible if international standards of quantification are included in the assays. Recently, the World Health Organization (WHO) established the first international standard for CMV detection. The real time PCR COBAS Ampliprep/ CobasTaqMan (CAP/CTM) (Roche®) was developed using the WHO standard for CMV quantification. However, the cut-off for the introduction of antiviral has not been determined yet. Methods: We conducted a retrospective study to determine: 1) the sensitivity and specificity of the new CMV CAP/CTM test in comparison with pp65 antigenemia to detect episodes of CMV infection/reactivation, and 2) the cut-off of viral load for introduction of ganciclovir (GCV). Pp65 antigenemia was performed and the corresponding plasma samples were stored at -20°C for further CMV detection by CAP/CTM. Comparison of tests was performed by kappa index. The appearance of positive antigenemia was considered the state variable to determine the cut-off of CMV viral load by ROC curve. Statistical analysis was performed using SPSS software version 19 (SPSS, Chicago, IL, USA.). Results: Thirty-eight patients were included and followed from August 2014 through May 2015. The antigenemia test detected 53 episodes of CMV infection in 34 patients (89.5%), while CAP/CTM detected 37 episodes in 33 patients (86.8%). AG and PCR results were compared in 431 samples and Kappa index was 30.9%. The median time for first AG detection was 42 (28-140) days, while CAP/CTM detected at a median of 7 days earlier (34 days, ranging from 7 to 110 days). The optimum cut-off value of CMV DNA was 34.25 IU/mL to detect positive antigenemia with 88.2% of sensibility, 100% of specificity and AUC of 0.91. This cut-off value is below the limit of detection and quantification of the equipment which is 56 IU/mL. According to CMV recurrence definition, 16 episodes of CMV recurrence were detected by antigenemia (47.1%) and 4 (12.1%) by CAP/CTM. The duration of viremia as detected by antigenemia was shorter (60.5% of the episodes lasted ≤ 7 days) in comparison to CAP/CTM (57.9% of the episodes lasting 15 days or more). This data suggests that the use of antigenemia to define the duration of GCV therapy might prompt early interruption of antiviral, which may favor CMV reactivation. The CAP/CTM PCR could possibly provide a safer information concerning the duration of GCV therapy. As prolonged treatment may increase the risk of toxicity, this hypothesis should be confirmed in prospective trials. Conclusions: Even though CAP/CTM by ROCHE showed great qualitative correlation with the antigenemia technique, the fully automated CAP/CTM did not demonstrate increased sensitivity. The cut-off value below the limit of detection and quantification may result in delayed introduction of pre-emptive therapy.Keywords: antigenemia, CMV COBAS/TAQMAN, cytomegalovirus, antiviral cut-off
Procedia PDF Downloads 1913066 Monitoring the Effect of Doxorubicin Liposomal in VX2 Tumor Using Magnetic Resonance Imaging
Authors: Ren-Jy Ben, Jo-Chi Jao, Chiu-Ya Liao, Ya-Ru Tsai, Lain-Chyr Hwang, Po-Chou Chen
Abstract:
Cancer is still one of the serious diseases threatening the lives of human beings. How to have an early diagnosis and effective treatment for tumors is a very important issue. The animal carcinoma model can provide a simulation tool for the study of pathogenesis, biological characteristics and therapeutic effects. Recently, drug delivery systems have been rapidly developed to effectively improve the therapeutic effects. Liposome plays an increasingly important role in clinical diagnosis and therapy for delivering a pharmaceutic or contrast agent to the targeted sites. Liposome can be absorbed and excreted by the human body, and is well known that no harm to the human body. This study aimed to compare the therapeutic effects between encapsulated (doxorubicin liposomal, LipoDox) and un-encapsulated (doxorubicin, Dox) anti-tumor drugs using Magnetic Resonance Imaging (MRI). Twenty-four New Zealand rabbits implanted with VX2 carcinoma at left thigh were classified into three groups: control group (untreated), Dox-treated group and LipoDox-treated group, 8 rabbits for each group. MRI scans were performed three days after tumor implantation. A 1.5T GE Signa HDxt whole body MRI scanner with a high resolution knee coil was used in this study. After a 3-plane localizer scan was performed, Three-Dimensional (3D) Fast Spin Echo (FSE) T2-Weighted Images (T2WI) was used for tumor volumetric quantification. And Two-Dimensional (2D) spoiled gradient recalled echo (SPGR) dynamic Contrast-enhanced (DCE) MRI was used for tumor perfusion evaluation. DCE-MRI was designed to acquire four baseline images, followed by contrast agent Gd-DOTA injection through the ear vein of rabbits. Afterwards, a series of 32 images were acquired to observe the signals change over time in the tumor and muscle. The MRI scanning was scheduled on a weekly basis for a period of four weeks to observe the tumor progression longitudinally. The Dox and LipoDox treatments were prescribed 3 times in the first week immediately after VX2 tumor implantation. ImageJ was used to quantitate tumor volume and time course signal enhancement on DCE images. The changes of tumor size showed that the growth of VX2 tumors was effectively inhibited for both LipoDox-treated and Dox-treated groups. Furthermore, the tumor volume of LipoDox-treated group was significantly lower than that of Dox-treated group, which implies that LipoDox has better therapeutic effect than Dox. The signal intensity of LipoDox-treated group is significantly lower than that of the other two groups, which implies that targeted therapeutic drug remained in the tumor tissue. This study provides a radiation-free and non-invasive MRI method for therapeutic monitoring of targeted liposome on an animal tumor model.Keywords: doxorubicin, dynamic contrast-enhanced MRI, lipodox, magnetic resonance imaging, VX2 tumor model
Procedia PDF Downloads 4573065 Volatile Organic Compounds Detection by Surface Acoustic Wave Sensors with Nanoparticles Embedded in Polymer Sensitive Layers
Authors: Cristian Viespe, Dana Miu
Abstract:
Surface acoustic wave (SAW) sensors with nanoparticles (NPs) of various dimensions and concentrations embedded in different types of polymer sensing films for detecting volatile organic compounds (VOCs) were studied. The sensors were ‘delay line’ type with a center frequency of 69.4 MHz on ST-X quartz substrates. NPs with different diameters of 7 nm or 13 nm were obtained by laser ablation with lasers having 5 ns or 10 ps pulse durations, respectively. The influence of NPs dimensions and concentrations on sensor properties such as frequency shift, sensitivity, noise and response time were investigated. To the best of our knowledge, the influence of NP dimensions on SAW sensor properties with has not been investigated. The frequency shift and sensitivity increased with increasing NP concentration in the polymer for a given NP dimension and with decreasing NP diameter for a given concentration. The best performances were obtained for the smallest NPs used. The SAW sensor with NPs of 7 nm had a limit of detection (LOD) of 65 ppm (almost five times better than the sensor with polymer alone), and a response time of about 9 s for ethanol.Keywords: surface acoustic wave sensor, nanoparticles, volatile organic compounds, laser ablation
Procedia PDF Downloads 1503064 Temperature-Based Detection of Initial Yielding Point in Loading of Tensile Specimens Made of Structural Steel
Authors: Aqsa Jamil, Tamura Hiroshi, Katsuchi Hiroshi, Wang Jiaqi
Abstract:
The yield point represents the upper limit of forces which can be applied to a specimen without causing any permanent deformation. After yielding, the behavior of the specimen suddenly changes, including the possibility of cracking or buckling. So, the accumulation of damage or type of fracture changes depending on this condition. As it is difficult to accurately detect yield points of the several stress concentration points in structural steel specimens, an effort has been made in this research work to develop a convenient technique using thermography (temperature-based detection) during tensile tests for the precise detection of yield point initiation. To verify the applicability of thermography camera, tests were conducted under different loading conditions and measuring the deformation by installing various strain gauges and monitoring the surface temperature with the help of a thermography camera. The yield point of specimens was estimated with the help of temperature dip, which occurs due to the thermoelastic effect during the plastic deformation. The scattering of the data has been checked by performing a repeatability analysis. The effects of temperature imperfection and light source have been checked by carrying out the tests at daytime as well as midnight and by calculating the signal to noise ratio (SNR) of the noised data from the infrared thermography camera, it can be concluded that the camera is independent of testing time and the presence of a visible light source. Furthermore, a fully coupled thermal-stress analysis has been performed by using Abaqus/Standard exact implementation technique to validate the temperature profiles obtained from the thermography camera and to check the feasibility of numerical simulation for the prediction of results extracted with the help of the thermographic technique.Keywords: signal to noise ratio, thermoelastic effect, thermography, yield point
Procedia PDF Downloads 1073063 An Algebraic Geometric Imaging Approach for Automatic Dairy Cow Body Condition Scoring System
Authors: Thi Thi Zin, Pyke Tin, Ikuo Kobayashi, Yoichiro Horii
Abstract:
Today dairy farm experts and farmers have well recognized the importance of dairy cow Body Condition Score (BCS) since these scores can be used to optimize milk production, managing feeding system and as an indicator for abnormality in health even can be utilized to manage for having healthy calving times and process. In tradition, BCS measures are done by animal experts or trained technicians based on visual observations focusing on pin bones, pin, thurl and hook area, tail heads shapes, hook angles and short and long ribs. Since the traditional technique is very manual and subjective, the results can lead to different scores as well as not cost effective. Thus this paper proposes an algebraic geometric imaging approach for an automatic dairy cow BCS system. The proposed system consists of three functional modules. In the first module, significant landmarks or anatomical points from the cow image region are automatically extracted by using image processing techniques. To be specific, there are 23 anatomical points in the regions of ribs, hook bones, pin bone, thurl and tail head. These points are extracted by using block region based vertical and horizontal histogram methods. According to animal experts, the body condition scores depend mainly on the shape structure these regions. Therefore the second module will investigate some algebraic and geometric properties of the extracted anatomical points. Specifically, the second order polynomial regression is employed to a subset of anatomical points to produce the regression coefficients which are to be utilized as a part of feature vector in scoring process. In addition, the angles at thurl, pin, tail head and hook bone area are computed to extend the feature vector. Finally, in the third module, the extracted feature vectors are trained by using Markov Classification process to assign BCS for individual cows. Then the assigned BCS are revised by using multiple regression method to produce the final BCS score for dairy cows. In order to confirm the validity of proposed method, a monitoring video camera is set up at the milk rotary parlor to take top view images of cows. The proposed method extracts the key anatomical points and the corresponding feature vectors for each individual cows. Then the multiple regression calculator and Markov Chain Classification process are utilized to produce the estimated body condition score for each cow. The experimental results tested on 100 dairy cows from self-collected dataset and public bench mark dataset show very promising with accuracy of 98%.Keywords: algebraic geometric imaging approach, body condition score, Markov classification, polynomial regression
Procedia PDF Downloads 1583062 A Convolutional Neural Network Based Vehicle Theft Detection, Location, and Reporting System
Authors: Michael Moeti, Khuliso Sigama, Thapelo Samuel Matlala
Abstract:
One of the principal challenges that the world is confronted with is insecurity. The crime rate is increasing exponentially, and protecting our physical assets especially in the motorist industry, is becoming impossible when applying our own strength. The need to develop technological solutions that detect and report theft without any human interference is inevitable. This is critical, especially for vehicle owners, to ensure theft detection and speedy identification towards recovery efforts in cases where a vehicle is missing or attempted theft is taking place. The vehicle theft detection system uses Convolutional Neural Network (CNN) to recognize the driver's face captured using an installed mobile phone device. The location identification function uses a Global Positioning System (GPS) to determine the real-time location of the vehicle. Upon identification of the location, Global System for Mobile Communications (GSM) technology is used to report or notify the vehicle owner about the whereabouts of the vehicle. The installed mobile app was implemented by making use of python as it is undoubtedly the best choice in machine learning. It allows easy access to machine learning algorithms through its widely developed library ecosystem. The graphical user interface was developed by making use of JAVA as it is better suited for mobile development. Google's online database (Firebase) was used as a means of storage for the application. The system integration test was performed using a simple percentage analysis. Sixty (60) vehicle owners participated in this study as a sample, and questionnaires were used in order to establish the acceptability of the system developed. The result indicates the efficiency of the proposed system, and consequently, the paper proposes the use of the system can effectively monitor the vehicle at any given place, even if it is driven outside its normal jurisdiction. More so, the system can be used as a database to detect, locate and report missing vehicles to different security agencies.Keywords: CNN, location identification, tracking, GPS, GSM
Procedia PDF Downloads 1663061 Surface Elevation Dynamics Assessment Using Digital Elevation Models, Light Detection and Ranging, GPS and Geospatial Information Science Analysis: Ecosystem Modelling Approach
Authors: Ali K. M. Al-Nasrawi, Uday A. Al-Hamdany, Sarah M. Hamylton, Brian G. Jones, Yasir M. Alyazichi
Abstract:
Surface elevation dynamics have always responded to disturbance regimes. Creating Digital Elevation Models (DEMs) to detect surface dynamics has led to the development of several methods, devices and data clouds. DEMs can provide accurate and quick results with cost efficiency, in comparison to the inherited geomatics survey techniques. Nowadays, remote sensing datasets have become a primary source to create DEMs, including LiDAR point clouds with GIS analytic tools. However, these data need to be tested for error detection and correction. This paper evaluates various DEMs from different data sources over time for Apple Orchard Island, a coastal site in southeastern Australia, in order to detect surface dynamics. Subsequently, 30 chosen locations were examined in the field to test the error of the DEMs surface detection using high resolution global positioning systems (GPSs). Results show significant surface elevation changes on Apple Orchard Island. Accretion occurred on most of the island while surface elevation loss due to erosion is limited to the northern and southern parts. Concurrently, the projected differential correction and validation method aimed to identify errors in the dataset. The resultant DEMs demonstrated a small error ratio (≤ 3%) from the gathered datasets when compared with the fieldwork survey using RTK-GPS. As modern modelling approaches need to become more effective and accurate, applying several tools to create different DEMs on a multi-temporal scale would allow easy predictions in time-cost-frames with more comprehensive coverage and greater accuracy. With a DEM technique for the eco-geomorphic context, such insights about the ecosystem dynamic detection, at such a coastal intertidal system, would be valuable to assess the accuracy of the predicted eco-geomorphic risk for the conservation management sustainability. Demonstrating this framework to evaluate the historical and current anthropogenic and environmental stressors on coastal surface elevation dynamism could be profitably applied worldwide.Keywords: DEMs, eco-geomorphic-dynamic processes, geospatial Information Science, remote sensing, surface elevation changes,
Procedia PDF Downloads 2673060 Nano-Sensors: Search for New Features
Authors: I. Filikhin, B. Vlahovic
Abstract:
We focus on a novel type of detection based on electron tunneling properties of double nanoscale structures in semiconductor materials. Semiconductor heterostructures as quantum wells (QWs), quantum dots (QDs), and quantum rings (QRs) may have energy level structure of several hundred of electron confinement states. The single electron spectra of the double quantum objects (DQW, DQD, and DQR) were studied in our previous works with relation to the electron localization and tunneling between the objects. The wave function of electron may be localized in one of the QDs or be delocalized when it is spread over the whole system. The localizing-delocalizing tunneling occurs when an electron transition between both states is possible. The tunneling properties of spectra differ strongly for “regular” and “chaotic” systems. We have shown that a small violation of the geometry drastically affects localization of electron. In particular, such violations lead to the elimination of the delocalized states of the system. The same symmetry violation effect happens if electrical or magnetic fields are applied. These phenomena could be used to propose a new type of detection based on the high sensitivity of charge transport between double nanostructures and small violations of the shapes. It may have significant technological implications.Keywords: double quantum dots, single electron levels, tunneling, electron localizations
Procedia PDF Downloads 5053059 Tibial Hemimelia Type VIIa: A Case Report
Authors: M. Medrano, M. D. M. S., L. Younes, M. D.
Abstract:
Introduction: Incidence of tibial hemimelia is 1:1,000,000. Due to pre-existing case studies and literature, there is now a better understanding of the genetics, etiology and pathoanatomy of tibial hemimelia, but an underlying cause is generally unknown. This presentation aims to discuss a rare, congenital lower limb deficiency observed in a patient in order to identify potential prenatal risk factors and future considerations for the patient’s well-being. Observation: A newborn female child, born full term via spontaneous vaginal delivery after induction of labor to unaffected and non-consanguineous parents. The prenatal course was notable for limited and disjointed prenatal care as well as maternal tobacco and marijuana use, anemia of pregnancy, and inadequate weight gain. Prenatal imaging showed lower extremity deformity with the inability to visualize tibia and bilateral clubfeet in the setting of Intrauterine Growth Restriction (IUGR). The patient presented with right equino varus deformity of the foot and right knee joint deformity. Radiological imaging showed the absence of the right tibia and varus angulation of the right foot with dislocation of the tibiotalar joint. Normal femur with lateral and mild anterior displacement of a wide fibula (Weber Type VIIa). Due to the absence of the patient’s tibia and knee extensor mechanism, the patient was not a candidate for reconstructive surgery and ultimately underwent successful right knee disarticulation. Discussion and Conclusion: By utilizing a retrospective chart review of this case, possible risk factors in prenatal care may be identified and add to existing knowledge on etiology. Hopefully, a cause can be clearly identified in the future and, thus, addressed in the prenatal period. In addition, we can investigate the patient’s well-being and adjustment post-operatively to support outpatient management of an uncommon anomaly.Keywords: Tibial hemimelia, prenatal care, pediatric orthopedics, congenital deformity
Procedia PDF Downloads 1593058 Pathological Disparities in Patients Diagnosed with Prostate Imaging Reporting and Data System 3 Lesions: A Retrospective Study in a High-Volume Academic Center
Authors: M. Reza Roshandel, Tannaz Aghaei Badr, Batoul Khoundabi, Sara C. Lewis, Soroush Rais-Bahrami, John Sfakianos, Reza Mehrazin, Ash K. Tewari
Abstract:
Introduction: Prostate biopsy is the most reliable diagnostic method for choosing the appropriate management of prostate cancer. However, discrepancies between Gleason grade groups (GG) of different biopsies remain a significant concern. This study aims to assess the association of the radiological factors with GG discrepancies in patients with index Prostate Imaging Reporting and Data System (PI-RADS) 3 lesions, using radical prostatectomy (RP) specimens as the most accurate and informative pathology. Methods: This single-institutional retrospective study was performed on a total of 2289 consecutive prostate cancer patients with combined targeted and systematic prostate biopsy followed by radical prostatectomy (RP). The database was explored for patients with the index PI-RADS 3 lesions version 2 and 2.1. Cancers with PI-RADS 4 or 5 scoring were excluded from the study. Patient characteristics and radiologic features were analyzed by multivariable logistic regression. Number-density of lesions was defined as the number of lesions per prostatic volume. Results: Of the 151 prostate cancer cases with PI-RADS 3 index lesions, 27% and 17% had upgrades and downgrades at RP, respectively. Analysis of grade changes showed no significant associations between discrepancies and the number or the number density of PI-RADS 3 lesions. Moreover, the study showed no significant association of the GG changes with race, age, location of the lesions, or prostate volume. Conclusions: This study demonstrated that in PI-RADS 3 cancerous nodules, the chance of the pathology changes in the final pathology of RP specimens was low. Furthermore, having multiple PI-RADS 3 nodules did not change the conclusion, as the possibility of grade changes in patients with multiple nodules was similar to those with solitary lesions.Keywords: prostate, adenocarcinoma, multiparametric MRI, Gleason score, robot-assisted surgery
Procedia PDF Downloads 1333057 Automatic Identification and Classification of Contaminated Biodegradable Plastics using Machine Learning Algorithms and Hyperspectral Imaging Technology
Authors: Nutcha Taneepanichskul, Helen C. Hailes, Mark Miodownik
Abstract:
Plastic waste has emerged as a critical global environmental challenge, primarily driven by the prevalent use of conventional plastics derived from petrochemical refining and manufacturing processes in modern packaging. While these plastics serve vital functions, their persistence in the environment post-disposal poses significant threats to ecosystems. Addressing this issue necessitates approaches, one of which involves the development of biodegradable plastics designed to degrade under controlled conditions, such as industrial composting facilities. It is imperative to note that compostable plastics are engineered for degradation within specific environments and are not suited for uncontrolled settings, including natural landscapes and aquatic ecosystems. The full benefits of compostable packaging are realized when subjected to industrial composting, preventing environmental contamination and waste stream pollution. Therefore, effective sorting technologies are essential to enhance composting rates for these materials and diminish the risk of contaminating recycling streams. In this study, it leverage hyperspectral imaging technology (HSI) coupled with advanced machine learning algorithms to accurately identify various types of plastics, encompassing conventional variants like Polyethylene terephthalate (PET), Polypropylene (PP), Low density polyethylene (LDPE), High density polyethylene (HDPE) and biodegradable alternatives such as Polybutylene adipate terephthalate (PBAT), Polylactic acid (PLA), and Polyhydroxyalkanoates (PHA). The dataset is partitioned into three subsets: a training dataset comprising uncontaminated conventional and biodegradable plastics, a validation dataset encompassing contaminated plastics of both types, and a testing dataset featuring real-world packaging items in both pristine and contaminated states. Five distinct machine learning algorithms, namely Partial Least Squares Discriminant Analysis (PLS-DA), Support Vector Machine (SVM), Convolutional Neural Network (CNN), Logistic Regression, and Decision Tree Algorithm, were developed and evaluated for their classification performance. Remarkably, the Logistic Regression and CNN model exhibited the most promising outcomes, achieving a perfect accuracy rate of 100% for the training and validation datasets. Notably, the testing dataset yielded an accuracy exceeding 80%. The successful implementation of this sorting technology within recycling and composting facilities holds the potential to significantly elevate recycling and composting rates. As a result, the envisioned circular economy for plastics can be established, thereby offering a viable solution to mitigate plastic pollution.Keywords: biodegradable plastics, sorting technology, hyperspectral imaging technology, machine learning algorithms
Procedia PDF Downloads 793056 Disaster Management Using Wireless Sensor Networks
Authors: Akila Murali, Prithika Manivel
Abstract:
Disasters are defined as a serious disruption of the functioning of a community or a society, which involves widespread human, material, economic or environmental impacts. The number of people suffering food crisis as a result of natural disasters has tripled in the last thirty years. The economic losses due to natural disasters have shown an increase with a factor of eight over the past four decades, caused by the increased vulnerability of the global society, and also due to an increase in the number of weather-related disasters. Efficient disaster detection and alerting systems could reduce the loss of life and properties. In the event of a disaster, another important issue is a good search and rescue system with high levels of precision, timeliness and safety for both the victims and the rescuers. Wireless Sensor Networks technology has the capability of quick capturing, processing, and transmission of critical data in real-time with high resolution. This paper studies the capacity of sensors and a Wireless Sensor Network to collect, collate and analyze valuable and worthwhile data, in an ordered manner to help with disaster management.Keywords: alerting systems, disaster detection, Ad Hoc network, WSN technology
Procedia PDF Downloads 4043055 Contextual Toxicity Detection with Data Augmentation
Authors: Julia Ive, Lucia Specia
Abstract:
Understanding and detecting toxicity is an important problem to support safer human interactions online. Our work focuses on the important problem of contextual toxicity detection, where automated classifiers are tasked with determining whether a short textual segment (usually a sentence) is toxic within its conversational context. We use “toxicity” as an umbrella term to denote a number of variants commonly named in the literature, including hate, abuse, offence, among others. Detecting toxicity in context is a non-trivial problem and has been addressed by very few previous studies. These previous studies have analysed the influence of conversational context in human perception of toxicity in controlled experiments and concluded that humans rarely change their judgements in the presence of context. They have also evaluated contextual detection models based on state-of-the-art Deep Learning and Natural Language Processing (NLP) techniques. Counterintuitively, they reached the general conclusion that computational models tend to suffer performance degradation in the presence of context. We challenge these empirical observations by devising better contextual predictive models that also rely on NLP data augmentation techniques to create larger and better data. In our study, we start by further analysing the human perception of toxicity in conversational data (i.e., tweets), in the absence versus presence of context, in this case, previous tweets in the same conversational thread. We observed that the conclusions of previous work on human perception are mainly due to data issues: The contextual data available does not provide sufficient evidence that context is indeed important (even for humans). The data problem is common in current toxicity datasets: cases labelled as toxic are either obviously toxic (i.e., overt toxicity with swear, racist, etc. words), and thus context does is not needed for a decision, or are ambiguous, vague or unclear even in the presence of context; in addition, the data contains labeling inconsistencies. To address this problem, we propose to automatically generate contextual samples where toxicity is not obvious (i.e., covert cases) without context or where different contexts can lead to different toxicity judgements for the same tweet. We generate toxic and non-toxic utterances conditioned on the context or on target tweets using a range of techniques for controlled text generation(e.g., Generative Adversarial Networks and steering techniques). On the contextual detection models, we posit that their poor performance is due to limitations on both of the data they are trained on (same problems stated above) and the architectures they use, which are not able to leverage context in effective ways. To improve on that, we propose text classification architectures that take the hierarchy of conversational utterances into account. In experiments benchmarking ours against previous models on existing and automatically generated data, we show that both data and architectural choices are very important. Our model achieves substantial performance improvements as compared to the baselines that are non-contextual or contextual but agnostic of the conversation structure.Keywords: contextual toxicity detection, data augmentation, hierarchical text classification models, natural language processing
Procedia PDF Downloads 1703054 Comparison of Techniques for Detection and Diagnosis of Eccentricity in the Air-Gap Fault in Induction Motors
Authors: Abrahão S. Fontes, Carlos A. V. Cardoso, Levi P. B. Oliveira
Abstract:
The induction motors are used worldwide in various industries. Several maintenance techniques are applied to increase the operating time and the lifespan of these motors. Among these, the predictive maintenance techniques such as Motor Current Signature Analysis (MCSA), Motor Square Current Signature Analysis (MSCSA), Park's Vector Approach (PVA) and Park's Vector Square Modulus (PVSM) are used to detect and diagnose faults in electric motors, characterized by patterns in the stator current frequency spectrum. In this article, these techniques are applied and compared on a real motor, which has the fault of eccentricity in the air-gap. It was used as a theoretical model of an electric induction motor without fault in order to assist comparison between the stator current frequency spectrum patterns with and without faults. Metrics were purposed and applied to evaluate the sensitivity of each technique fault detection. The results presented here show that the above techniques are suitable for the fault of eccentricity in the air gap, whose comparison between these showed the suitability of each one.Keywords: eccentricity in the air-gap, fault diagnosis, induction motors, predictive maintenance
Procedia PDF Downloads 3503053 Detection of Biomechanical Stress for the Prevention of Disability Derived from Musculoskeletal Disorders
Authors: Leydi Noemi Peraza Gómez, Jose Álvarez Nemegyei, Damaris Francis Estrella Castillo
Abstract:
In order to have an epidemiological tool to detect biomechanical stress (ERGO-Mex), which impose physical labor or recreational activities, a questionnaire is constructed in Spanish, validated and culturally adapted to the Mayan indigenous population of Yucatan. Through the seven steps proposed by Guillemin and Beaton the procedure was: initial translation, synthesis of the translations, feed back of the translation. After that review by a committee of experts, pre-test of the preliminary version, and presentation of the results to the committee of experts and members of the community. Finally the evaluation of its internal validity (Cronbach's α coefficient) and external (intraclass correlation coefficient). The results for the validation in Spanish indicated that 45% of the participants have biomechanical stress. The ERGO-Mex correlation was 0.69 (p <0.0001). Subjects with high biomechanical stress had a higher score than subjects with low biomechanical stress (17.4 ± 8.9 vs.9.8 ± 2.8, p = 0.003). The Cronbach's α coefficient was 0.92; and for validation in Cronbach's α maya it was 0.82 and CCI = 0.70 (95% CI: 0.58-0.79; p˂0.0001); ERGO-Mex is suitable for performing early detection of musculoskeletal diseases and helping to prevent disability.Keywords: biomechanical stress, disability, musculoskeletal disorders, prevention
Procedia PDF Downloads 1803052 Preliminary Evaluation of Maximum Intensity Projection SPECT Imaging for Whole Body Tc-99m Hydroxymethylene Diphosphonate Bone Scanning
Authors: Yasuyuki Takahashi, Hirotaka Shimada, Kyoko Saito
Abstract:
Bone scintigraphy is widely used as a screening tool for bone metastases. However, the 180 to 240 minutes (min) waiting time after the intravenous (i.v.) injection of the tracer is both long and tiresome. To solve this shortcoming, a bone scan with a shorter waiting time is needed. In this study, we applied the Maximum Intensity Projection (MIP) and triple energy window (TEW) scatter correction to a whole body bone SPECT (Merged SPECT) and investigated shortening the waiting time. Methods: In a preliminary phantom study, hot gels of 99mTc-HMDP were inserted into sets of rods with diameters ranging from 4 to 19 mm. Each rod set covered a sector of a cylindrical phantom. The activity concentration of all rods was 2.5 times that of the background in the cylindrical body of the phantom. In the human study, SPECT images were obtained from chest to abdomen at 30 to 180 min after 99mTc- hydroxymethylene diphosphonate (HMDP) injection of healthy volunteers. For both studies, MIP images were reconstructed. Planar whole body images of the patients were also obtained. These were acquired at 200 min. The image quality of the SPECT and the planar images was compared. Additionally, 36 patients with breast cancer were scanned in the same way. The delectability of uptake regions (metastases) was compared visually. Results: In the phantom study, a 4 mm size hot gel was difficult to depict on the conventional SPECT, but MIP images could recognize it clearly. For both the healthy volunteers and the clinical patients, the accumulation of 99mTc-HMDP in the SPECT was good as early as 90 min. All findings of both image sets were in agreement. Conclusion: In phantoms, images from MIP with TEW scatter correction could detect all rods down to those with a diameter of 4 mm. In patients, MIP reconstruction with TEW scatter correction could improve the detectability of hot lesions. In addition, the time between injection and imaging could be shortened from that conventionally used for whole body scans.Keywords: merged SPECT, MIP, TEW scatter correction, 99mTc-HMDP
Procedia PDF Downloads 4123051 Deep Learning Based Fall Detection Using Simplified Human Posture
Authors: Kripesh Adhikari, Hamid Bouchachia, Hammadi Nait-Charif
Abstract:
Falls are one of the major causes of injury and death among elderly people aged 65 and above. A support system to identify such kind of abnormal activities have become extremely important with the increase in ageing population. Pose estimation is a challenging task and to add more to this, it is even more challenging when pose estimations are performed on challenging poses that may occur during fall. Location of the body provides a clue where the person is at the time of fall. This paper presents a vision-based tracking strategy where available joints are grouped into three different feature points depending upon the section they are located in the body. The three feature points derived from different joints combinations represents the upper region or head region, mid-region or torso and lower region or leg region. Tracking is always challenging when a motion is involved. Hence the idea is to locate the regions in the body in every frame and consider it as the tracking strategy. Grouping these joints can be beneficial to achieve a stable region for tracking. The location of the body parts provides a crucial information to distinguish normal activities from falls.Keywords: fall detection, machine learning, deep learning, pose estimation, tracking
Procedia PDF Downloads 1893050 Deep Learning based Image Classifiers for Detection of CSSVD in Cacao Plants
Authors: Atuhurra Jesse, N'guessan Yves-Roland Douha, Pabitra Lenka
Abstract:
The detection of diseases within plants has attracted a lot of attention from computer vision enthusiasts. Despite the progress made to detect diseases in many plants, there remains a research gap to train image classifiers to detect the cacao swollen shoot virus disease or CSSVD for short, pertinent to cacao plants. This gap has mainly been due to the unavailability of high quality labeled training data. Moreover, institutions have been hesitant to share their data related to CSSVD. To fill these gaps, image classifiers to detect CSSVD-infected cacao plants are presented in this study. The classifiers are based on VGG16, ResNet50 and Vision Transformer (ViT). The image classifiers are evaluated on a recently released and publicly accessible KaraAgroAI Cocoa dataset. The best performing image classifier, based on ResNet50, achieves 95.39\% precision, 93.75\% recall, 94.34\% F1-score and 94\% accuracy on only 20 epochs. There is a +9.75\% improvement in recall when compared to previous works. These results indicate that the image classifiers learn to identify cacao plants infected with CSSVD.Keywords: CSSVD, image classification, ResNet50, vision transformer, KaraAgroAI cocoa dataset
Procedia PDF Downloads 1033049 A Framework for Automated Nuclear Waste Classification
Authors: Seonaid Hume, Gordon Dobie, Graeme West
Abstract:
Detecting and localizing radioactive sources is a necessity for safe and secure decommissioning of nuclear facilities. An important aspect for the management of the sort-and-segregation process is establishing the spatial distributions and quantities of the waste radionuclides, their type, corresponding activity, and ultimately classification for disposal. The data received from surveys directly informs decommissioning plans, on-site incident management strategies, the approach needed for a new cell, as well as protecting the workforce and the public. Manual classification of nuclear waste from a nuclear cell is time-consuming, expensive, and requires significant expertise to make the classification judgment call. Also, in-cell decommissioning is still in its relative infancy, and few techniques are well-developed. As with any repetitive and routine tasks, there is the opportunity to improve the task of classifying nuclear waste using autonomous systems. Hence, this paper proposes a new framework for the automatic classification of nuclear waste. This framework consists of five main stages; 3D spatial mapping and object detection, object classification, radiological mapping, source localisation based on gathered evidence and finally, waste classification. The first stage of the framework, 3D visual mapping, involves object detection from point cloud data. A review of related applications in other industries is provided, and recommendations for approaches for waste classification are made. Object detection focusses initially on cylindrical objects since pipework is significant in nuclear cells and indeed any industrial site. The approach can be extended to other commonly occurring primitives such as spheres and cubes. This is in preparation of stage two, characterizing the point cloud data and estimating the dimensions, material, degradation, and mass of the objects detected in order to feature match them to an inventory of possible items found in that nuclear cell. Many items in nuclear cells are one-offs, have limited or poor drawings available, or have been modified since installation, and have complex interiors, which often and inadvertently pose difficulties when accessing certain zones and identifying waste remotely. Hence, this may require expert input to feature match objects. The third stage, radiological mapping, is similar in order to facilitate the characterization of the nuclear cell in terms of radiation fields, including the type of radiation, activity, and location within the nuclear cell. The fourth stage of the framework takes the visual map for stage 1, the object characterization from stage 2, and radiation map from stage 3 and fuses them together, providing a more detailed scene of the nuclear cell by identifying the location of radioactive materials in three dimensions. The last stage involves combining the evidence from the fused data sets to reveal the classification of the waste in Bq/kg, thus enabling better decision making and monitoring for in-cell decommissioning. The presentation of the framework is supported by representative case study data drawn from an application in decommissioning from a UK nuclear facility. This framework utilises recent advancements of the detection and mapping capabilities of complex radiation fields in three dimensions to make the process of classifying nuclear waste faster, more reliable, cost-effective and safer.Keywords: nuclear decommissioning, radiation detection, object detection, waste classification
Procedia PDF Downloads 2003048 DISGAN: Efficient Generative Adversarial Network-Based Method for Cyber-Intrusion Detection
Authors: Hongyu Chen, Li Jiang
Abstract:
Ubiquitous anomalies endanger the security of our system con- stantly. They may bring irreversible damages to the system and cause leakage of privacy. Thus, it is of vital importance to promptly detect these anomalies. Traditional supervised methods such as Decision Trees and Support Vector Machine (SVM) are used to classify normality and abnormality. However, in some case, the abnormal status are largely rarer than normal status, which leads to decision bias of these methods. Generative adversarial network (GAN) has been proposed to handle the case. With its strong generative ability, it only needs to learn the distribution of normal status, and identify the abnormal status through the gap between it and the learned distribution. Nevertheless, existing GAN-based models are not suitable to process data with discrete values, leading to immense degradation of detection performance. To cope with the discrete features, in this paper, we propose an efficient GAN-based model with specifically-designed loss function. Experiment results show that our model outperforms state-of-the-art models on discrete dataset and remarkably reduce the overhead.Keywords: GAN, discrete feature, Wasserstein distance, multiple intermediate layers
Procedia PDF Downloads 1293047 Direct Integration of 3D Ultrasound Scans with Patient Educational Mobile Application
Authors: Zafar Iqbal, Eugene Chan, Fareed Ahmed, Mohamed Jama, Avez Rizvi
Abstract:
Advancements in Ultrasound Technology have enabled machines to capture 3D and 4D images with intricate features of the growing fetus. Sonographers can now capture clear 3D images and 4D videos of the fetus, especially of the face. Fetal faces are often seen on the ultrasound scan of the third trimester where anatomical features become more defined. Parents often want 3D/4D images and videos of their ultrasounds, and particularly image that capture the child’s face. Sidra Medicine developed a patient education mobile app called 10 Moons to improve care and provide useful information during the length of their pregnancy. In addition to general information, we built the ability to send ultrasound images directly from the modality to the mobile application, allowing expectant mothers to easily store and share images of their baby. 10 Moons represent the length of the pregnancy on a lunar calendar, which has both cultural and religious significance in the Middle East. During the third trimester scan, sonographers can capture 3D pictures of the fetus. Ultrasound machines are connected with a local 10 Moons Server with a Digital Imaging and Communications in Medicine (DICOM) application running on it. Sonographers are able to send images directly to the DICOM server by a preprogrammed button on the ultrasound modality. Mothers can also request which pictures they would like to be available on the app. An internally built DICOM application receives the image and saves the patient information from DICOM header (for verification purpose). The application also anonymizes the image by removing all the DICOM header information and subsequently converts it into a lossless JPEG. Finally, and the application passes the image to the mobile application server. On the 10 Moons mobile app – patients enter their Medical Record Number (MRN) and Date of Birth (DOB) to receive a One Time Password (OTP) for security reasons to view the images. Patients can also share the images anonymized images with friends and family. Furthermore, patients can also request 3D printed mementos of their child through 10 Moons. 10 Moons is unique patient education and information application where expected mothers can also see 3D ultrasound images of their children. Sidra Medicine staff has the added benefit of a full content management administrative backend where updates to content can be made. The app is available on secure infrastructure with both local and public interfaces. The application is also available in both English and Arabic languages to facilitate most of the patients in the region. Innovation is at the heart of modern healthcare management. With Innovation being one of Sidra Medicine’s core values, our 10 Moons application provides expectant mothers with unique educational content as well as the ability to store and share images of their child and purchase 3D printed mementos.Keywords: patient educational mobile application, ultrasound images, digital imaging and communications in medicine (DICOM), imaging informatics
Procedia PDF Downloads 1403046 Non-intrusive Hand Control of Drone Using an Inexpensive and Streamlined Convolutional Neural Network Approach
Authors: Evan Lowhorn, Rocio Alba-Flores
Abstract:
The purpose of this work is to develop a method for classifying hand signals and using the output in a drone control algorithm. To achieve this, methods based on Convolutional Neural Networks (CNN) were applied. CNN's are a subset of deep learning, which allows grid-like inputs to be processed and passed through a neural network to be trained for classification. This type of neural network allows for classification via imaging, which is less intrusive than previous methods using biosensors, such as EMG sensors. Classification CNN's operate purely from the pixel values in an image; therefore they can be used without additional exteroceptive sensors. A development bench was constructed using a desktop computer connected to a high-definition webcam mounted on a scissor arm. This allowed the camera to be pointed downwards at the desk to provide a constant solid background for the dataset and a clear detection area for the user. A MATLAB script was created to automate dataset image capture at the development bench and save the images to the desktop. This allowed the user to create their own dataset of 12,000 images within three hours. These images were evenly distributed among seven classes. The defined classes include forward, backward, left, right, idle, and land. The drone has a popular flip function which was also included as an additional class. To simplify control, the corresponding hand signals chosen were the numerical hand signs for one through five for movements, a fist for land, and the universal “ok” sign for the flip command. Transfer learning with PyTorch (Python) was performed using a pre-trained 18-layer residual learning network (ResNet-18) to retrain the network for custom classification. An algorithm was created to interpret the classification and send encoded messages to a Ryze Tello drone over its 2.4 GHz Wi-Fi connection. The drone’s movements were performed in half-meter distance increments at a constant speed. When combined with the drone control algorithm, the classification performed as desired with negligible latency when compared to the delay in the drone’s movement commands.Keywords: classification, computer vision, convolutional neural networks, drone control
Procedia PDF Downloads 2103045 Audio-Visual Co-Data Processing Pipeline
Authors: Rita Chattopadhyay, Vivek Anand Thoutam
Abstract:
Speech is the most acceptable means of communication where we can quickly exchange our feelings and thoughts. Quite often, people can communicate orally but cannot interact or work with computers or devices. It’s easy and quick to give speech commands than typing commands to computers. In the same way, it’s easy listening to audio played from a device than extract output from computers or devices. Especially with Robotics being an emerging market with applications in warehouses, the hospitality industry, consumer electronics, assistive technology, etc., speech-based human-machine interaction is emerging as a lucrative feature for robot manufacturers. Considering this factor, the objective of this paper is to design the “Audio-Visual Co-Data Processing Pipeline.” This pipeline is an integrated version of Automatic speech recognition, a Natural language model for text understanding, object detection, and text-to-speech modules. There are many Deep Learning models for each type of the modules mentioned above, but OpenVINO Model Zoo models are used because the OpenVINO toolkit covers both computer vision and non-computer vision workloads across Intel hardware and maximizes performance, and accelerates application development. A speech command is given as input that has information about target objects to be detected and start and end times to extract the required interval from the video. Speech is converted to text using the Automatic speech recognition QuartzNet model. The summary is extracted from text using a natural language model Generative Pre-Trained Transformer-3 (GPT-3). Based on the summary, essential frames from the video are extracted, and the You Only Look Once (YOLO) object detection model detects You Only Look Once (YOLO) objects on these extracted frames. Frame numbers that have target objects (specified objects in the speech command) are saved as text. Finally, this text (frame numbers) is converted to speech using text to speech model and will be played from the device. This project is developed for 80 You Only Look Once (YOLO) labels, and the user can extract frames based on only one or two target labels. This pipeline can be extended for more than two target labels easily by making appropriate changes in the object detection module. This project is developed for four different speech command formats by including sample examples in the prompt used by Generative Pre-Trained Transformer-3 (GPT-3) model. Based on user preference, one can come up with a new speech command format by including some examples of the respective format in the prompt used by the Generative Pre-Trained Transformer-3 (GPT-3) model. This pipeline can be used in many projects like human-machine interface, human-robot interaction, and surveillance through speech commands. All object detection projects can be upgraded using this pipeline so that one can give speech commands and output is played from the device.Keywords: OpenVINO, automatic speech recognition, natural language processing, object detection, text to speech
Procedia PDF Downloads 803044 Paper-Like and Battery Free Sensor Patches for Wound Monitoring
Authors: Xiaodi Su, Xin Ting Zheng, Laura Sutarlie, Nur Asinah binte Mohamed Salleh, Yong Yu
Abstract:
Wound healing is a dynamic process with multiple phases. Rapid profiling and quantitative characterization of inflammation and infection remain challenging. We have developed paper-like battery-free multiplexed sensors for holistic wound assessment via quantitative detection of multiple inflammation and infection markers. In one of the designs, the sensor patch consists of a wax-printed paper panel with five colorimetric sensor channels arranged in a pattern resembling a five-petaled flower (denoted as a ‘Petal’ sensor). The five sensors are for temperature, pH, trimethylamine, uric acid, and moisture. The sensor patch is sandwiched between a top transparent silicone layer and a bottom adhesive wound contact layer. In the second design, a palm-like-shaped paper strip is fabricated by a paper-cutter printer (denoted as ‘Palm’ sensor). This sensor strip carries five sensor regions connected by a stem sampling entrance that enables rapid colorimetric detection of multiple bacteria metabolites (aldehyde, lactate, moisture, trimethylamine, tryptophan) from wound exudate. For both the “\’ Petal’ and ‘Palm’ sensors, color images can be captured by a mobile phone. According to the color changes, one can quantify the concentration of the biomarkers and then determine wound healing status and identify/quantify bacterial species in infected wounds. The ‘Petal’ and ‘Palm’ sensors are validated with in-situ animal and ex-situ skin wound models, respectively. These sensors have the potential for integration with wound dressing to allow early warning of adverse events without frequent removal of the plasters. Such in-situ and early detection of non-healing condition can trigger immediate clinical intervention to facilitate wound care management.Keywords: wound infection, colorimetric sensor, paper fluidic sensor, wound care
Procedia PDF Downloads 813043 Linear Quadratic Gaussian/Loop Transfer Recover Control Flight Control on a Nonlinear Model
Authors: T. Sanches, K. Bousson
Abstract:
As part of the development of a 4D autopilot system for unmanned aerial vehicles (UAVs), i.e. a time-dependent robust trajectory generation and control algorithm, this work addresses the problem of optimal path control based on the flight sensors data output that may be unreliable due to noise on data acquisition and/or transmission under certain circumstances. Although several filtering methods, such as the Kalman-Bucy filter or the Linear Quadratic Gaussian/Loop Transfer Recover Control (LQG/LTR), are available, the utter complexity of the control system, together with the robustness and reliability required of such a system on a UAV for airworthiness certifiable autonomous flight, required the development of a proper robust filter for a nonlinear system, as a way of further mitigate errors propagation to the control system and improve its ,performance. As such, a nonlinear algorithm based upon the LQG/LTR, is validated through computational simulation testing, is proposed on this paper.Keywords: autonomous flight, LQG/LTR, nonlinear state estimator, robust flight control
Procedia PDF Downloads 1383042 Closing the Gap: Efficient Voxelization with Equidistant Scanlines and Gap Detection
Authors: S. Delgado, C. Cerrada, R. S. Gómez
Abstract:
This research introduces an approach to voxelizing the surfaces of triangular meshes with efficiency and accuracy. Our method leverages parallel equidistant scan-lines and introduces a Gap Detection technique to address the limitations of existing approaches. We present a comprehensive study showcasing the method's effectiveness, scalability, and versatility in different scenarios. Voxelization is a fundamental process in computer graphics and simulations, playing a pivotal role in applications ranging from scientific visualization to virtual reality. Our algorithm focuses on enhancing the voxelization process, especially for complex models and high resolutions. One of the major challenges in voxelization in the Graphics Processing Unit (GPU) is the high cost of discovering the same voxels multiple times. These repeated voxels incur in costly memory operations with no useful information. Our scan-line-based method ensures that each voxel is detected exactly once when processing the triangle, enhancing performance without compromising the quality of the voxelization. The heart of our approach lies in the use of parallel, equidistant scan-lines to traverse the interiors of triangles. This minimizes redundant memory operations and avoids revisiting the same voxels, resulting in a significant performance boost. Moreover, our method's computational efficiency is complemented by its simplicity and portability. Written as a single compute shader in Graphics Library Shader Language (GLSL), it is highly adaptable to various rendering pipelines and hardware configurations. To validate our method, we conducted extensive experiments on a diverse set of models from the Stanford repository. Our results demonstrate not only the algorithm's efficiency, but also its ability to produce 26 tunnel free accurate voxelizations. The Gap Detection technique successfully identifies and addresses gaps, ensuring consistent and visually pleasing voxelized surfaces. Furthermore, we introduce the Slope Consistency Value metric, quantifying the alignment of each triangle with its primary axis. This metric provides insights into the impact of triangle orientation on scan-line based voxelization methods. It also aids in understanding how the Gap Detection technique effectively improves results by targeting specific areas where simple scan-line-based methods might fail. Our research contributes to the field of voxelization by offering a robust and efficient approach that overcomes the limitations of existing methods. The Gap Detection technique fills a critical gap in the voxelization process. By addressing these gaps, our algorithm enhances the visual quality and accuracy of voxelized models, making it valuable for a wide range of applications. In conclusion, "Closing the Gap: Efficient Voxelization with Equidistant Scan-lines and Gap Detection" presents an effective solution to the challenges of voxelization. Our research combines computational efficiency, accuracy, and innovative techniques to elevate the quality of voxelized surfaces. With its adaptable nature and valuable innovations, this technique could have a positive influence on computer graphics and visualization.Keywords: voxelization, GPU acceleration, computer graphics, compute shaders
Procedia PDF Downloads 723041 Solvent Dependent Triazole-Appended Glucofuranose-Based Fluorometric Sensor for Detection of Au³⁺ Ions
Authors: Samiul Islam Hazarika, Domngam Boje, Ananta Kumar Atta
Abstract:
It is well familiar that solvents play a significant role in modern chemistry. Solvents can change the reactivity and physicochemical properties of molecules in a solution. Keeping this in mind, we have designed and synthesized a mono-triazolyl-linked pyrenyl-appended xylofuranose derivative for the detection of metal ions with changing solvent systems. The incorporation of a sugar backbone in the sensor increases the water solubility and biocompatibility. The experimental study revealed that the xylofuranose-based fluorescence probe did not exhibit any specific selectivity towards metal ions in acetonitrile (CH₃CN) solvent. Whereas, we revealed that triazole-linked pyrenyl-appended xylofuranose-based fluorescent sensor would exhibit high selectivity and sensitivity towards Au³⁺ ions in CH₃CN-H₂O (1/1, v/v) system. This observation might be explained by the viscosity and polarity differences of CH₃CN and CH₃CN-H₂O solvent systems. The formation of the sensor-Au³⁺ complex was also established by high-resolution mass spectrometry (HRMS) data of the complex.Keywords: triazole, furanose, fluorometric, solvent dependent
Procedia PDF Downloads 1153040 Functional Nanomaterials for Environmental Applications
Authors: S. A. M. Sabrina, Gouget Lammel, Anne Chantal, Chazalviel, Jean Noël, Ozanam François, Etcheberry Arnaud, Tighlit Fatma Zohra, B. Samia, Gabouze Noureddine
Abstract:
The elaboration and characterization of hybrid nano materials give rise to considerable interest due to the new properties that arising. They are considered as an important category of new materials having innovative characteristics by combining the specific intrinsic properties of inorganic compounds (semiconductors) with the grafted organic species. This open the way to improved properties and spectacular applications in various and important fields, especially in the environment. In this work, nano materials based-semiconductors were elaborated by chemical route. The obtained surfaces were grafted with organic functional groups. The functionalization process was optimized in order to confer to the hybrid nano material a good stability as well as the right properties required for the subsequent applications. Different characterization techniques were used to investigate the resulting nano structures, such as SEM, UV-Visible, FTIR, Contact angle and electro chemical measurements. Finally, applications were envisaged in environmental area. The elaborated nano structures were tested for the detection and the elimination of pollutants.Keywords: hybrid materials, porous silicon, peptide, metal detection
Procedia PDF Downloads 499