Search results for: damaged detection
2544 Cognitive Radio in Aeronautic: Comparison of Some Spectrum Sensing Technics
Authors: Abdelkhalek Bouchikhi, Elyes Benmokhtar, Sebastien Saletzki
Abstract:
The aeronautical field is experiencing issues with RF spectrum congestion due to the constant increase in the number of flights, aircrafts and telecom systems on board. In addition, these systems are bulky in size, weight and energy consumption. The cognitive radio helps particularly solving the spectrum congestion issue by its capacity to detect idle frequency channels then, allowing an opportunistic exploitation of the RF spectrum. The present work aims to propose a new use case for aeronautical spectrum sharing and to study the performances of three different detection techniques: energy detector, matched filter and cyclostationary detector within the aeronautical use case. The spectrum in the proposed cognitive radio is allocated dynamically where each cognitive radio follows a cognitive cycle. The spectrum sensing is a crucial step. The goal of the sensing is gathering data about the surrounding environment. Cognitive radio can use different sensors: antennas, cameras, accelerometer, thermometer, etc. In IEEE 802.22 standard, for example, a primary user (PU) has always the priority to communicate. When a frequency channel witch used by the primary user is idle, the secondary user (SU) is allowed to transmit in this channel. The Distance Measuring Equipment (DME) is composed of a UHF transmitter/receiver (interrogator) in the aircraft and a UHF receiver/transmitter on the ground. While the future cognitive radio will be used jointly to alleviate the spectrum congestion issue in the aeronautical field. LDACS, for example, is a good candidate; it provides two isolated data-links: ground-to-air and air-to-ground data-links. The first contribution of the present work is a strategy allowing sharing the L-band. The adopted spectrum sharing strategy is as follow: the DME will play the role of PU which is the licensed user and the LDACS1 systems will be the SUs. The SUs could use the L-band channels opportunely as long as they do not causing harmful interference signals which affect the QoS of the DME system. Although the spectrum sensing is a key step, it helps detecting holes by determining whether the primary signal is present or not in a given frequency channel. A missing detection on primary user presence creates interference between PU and SU and will affect seriously the QoS of the legacy radio. In this study, first brief definitions, concepts and the state of the art of cognitive radio will be presented. Then, a study of three communication channel detection algorithms in a cognitive radio context is carried out. The study is made from the point of view of functions, material requirements and signal detection capability in the aeronautical field. Then, we presented a modeling of the detection problem by three different methods (energy, adapted filter, and cyclostationary) as well as an algorithmic description of these detectors is done. Then, we study and compare the performance of the algorithms. Simulations were carried out using MATLAB software. We analyzed the results based on ROCs curves for SNR between -10dB and 20dB. The three detectors have been tested with a synthetics and real world signals.Keywords: aeronautic, communication, navigation, surveillance systems, cognitive radio, spectrum sensing, software defined radio
Procedia PDF Downloads 1752543 A Study on the Application of Machine Learning and Deep Learning Techniques for Skin Cancer Detection
Authors: Hritwik Ghosh, Irfan Sadiq Rahat, Sachi Nandan Mohanty, J. V. R. Ravindra
Abstract:
In the rapidly evolving landscape of medical diagnostics, the early detection and accurate classification of skin cancer remain paramount for effective treatment outcomes. This research delves into the transformative potential of Artificial Intelligence (AI), specifically Deep Learning (DL), as a tool for discerning and categorizing various skin conditions. Utilizing a diverse dataset of 3,000 images representing nine distinct skin conditions, we confront the inherent challenge of class imbalance. This imbalance, where conditions like melanomas are over-represented, is addressed by incorporating class weights during the model training phase, ensuring an equitable representation of all conditions in the learning process. Our pioneering approach introduces a hybrid model, amalgamating the strengths of two renowned Convolutional Neural Networks (CNNs), VGG16 and ResNet50. These networks, pre-trained on the ImageNet dataset, are adept at extracting intricate features from images. By synergizing these models, our research aims to capture a holistic set of features, thereby bolstering classification performance. Preliminary findings underscore the hybrid model's superiority over individual models, showcasing its prowess in feature extraction and classification. Moreover, the research emphasizes the significance of rigorous data pre-processing, including image resizing, color normalization, and segmentation, in ensuring data quality and model reliability. In essence, this study illuminates the promising role of AI and DL in revolutionizing skin cancer diagnostics, offering insights into its potential applications in broader medical domains.Keywords: artificial intelligence, machine learning, deep learning, skin cancer, dermatology, convolutional neural networks, image classification, computer vision, healthcare technology, cancer detection, medical imaging
Procedia PDF Downloads 862542 Performance and Damage Detection of Composite Structural Insulated Panels Subjected to Shock Wave Loading
Authors: Anupoju Rajeev, Joanne Mathew, Amit Shelke
Abstract:
In the current study, a new type of Composite Structural Insulated Panels (CSIPs) is developed and investigated its performance against shock loading which can replace the conventional wooden structural materials. The CSIPs is made of Fibre Cement Board (FCB)/aluminum as the facesheet and the expanded polystyrene foam as the core material. As tornadoes are very often in the western countries, it is suggestable to monitor the health of the CSIPs during its lifetime. So, the composite structure is installed with three smart sensors located randomly at definite locations. Each smart sensor is fabricated with an embedded half stainless phononic crystal sensor attached to both ends of the nylon shaft that can resist the shock and impact on facesheet as well as polystyrene foam core and safeguards the system. In addition to the granular crystal sensors, the accelerometers are used in the horizontal spanning and vertical spanning with a definite offset distance. To estimate the health and damage of the CSIP panel using granular crystal sensor, shock wave loading experiments are conducted. During the experiments, the time of flight response from the granular sensors is measured. The main objective of conducting shock wave loading experiments on the CSIP panels is to study the effect and the sustaining capacity of the CSIP panels in the extreme hazardous situations like tornados and hurricanes which are very common in western countries. The effects have been replicated using a shock tube, an instrument that can be used to create the same wind and pressure intensity of tornado for the experimental study. Numerous experiments have been conducted to investigate the flexural strength of the CSIP. Furthermore, the study includes the damage detection using three smart sensors embedded in the CSIPs during the shock wave loading.Keywords: composite structural insulated panels, damage detection, flexural strength, sandwich structures, shock wave loading
Procedia PDF Downloads 1462541 Assessing P0.1 and Occlusion Pressures in Brain-Injured Patients on Pressure Support Ventilation: A Study Protocol
Authors: S. B. R. Slagmulder
Abstract:
Monitoring inspiratory effort and dynamic lung stress in patients on pressure support ventilation in the ICU is important for protecting against self inflicted lung injury (P-SILI) and diaphragm dysfunction. Strategies to address the detrimental effects of respiratory drive and effort can lead to improved patient outcomes. Two non-invasive estimation methods, occlusion pressure (Pocc) and P0.1, have been proposed for achieving lung and diaphragm protective ventilation. However, their relationship and interpretation in neuro ICU patients is not well understood. P0.1 is the airway pressure measured during a 100-millisecond occlusion of the inspiratory port. It reflects the neural drive from the respiratory centers to the diaphragm and respiratory muscles, indicating the patient's respiratory drive during the initiation of each breath. Occlusion pressure, measured during a brief inspiratory pause against a closed airway, provides information about the inspiratory muscles' strength and the system's total resistance and compliance. Research Objective: Understanding the relationship between Pocc and P0.1 in brain-injured patients can provide insights into the interpretation of these values in pressure support ventilation. This knowledge can contribute to determining extubation readiness and optimizing ventilation strategies to improve patient outcomes. The central goal is to asses a study protocol for determining the relationship between Pocc and P0.1 in brain-injured patients on pressure support ventilation and their ability to predict successful extubation. Additionally, comparing these values between brain-damaged and non-brain-damaged patients may provide valuable insights. Key Areas of Inquiry: 1. How do Pocc and P0.1 values correlate within brain injury patients undergoing pressure support ventilation? 2. To what extent can Pocc and P0.1 values serve as predictive indicators for successful extubation in patients with brain injuries? 3. What differentiates the Pocc and P0.1 values between patients with brain injuries and those without? Methodology: P0.1 and occlusion pressures are standard measurements for pressure support ventilation patients, taken by attending doctors as per protocol. We utilize electronic patient records for existing data. Unpaired T-test will be conducted to compare P0.1 and Pocc values between both study groups. Associations between P0.1 and Pocc and other study variables, such as extubation, will be explored with simple regression and correlation analysis. Depending on how the data evolve, subgroup analysis will be performed for patients with and without extubation failure. Results: While it is anticipated that neuro patients may exhibit high respiratory drive, the linkage between such elevation, quantified by P0.1, and successful extubation remains unknown The analysis will focus on determining the ability of these values to predict successful extubation and their potential impact on ventilation strategies. Conclusion: Further research is pending to fully understand the potential of these indices and their impact on mechanical ventilation in different patient populations and clinical scenarios. Understanding these relationships can aid in determining extubation readiness and tailoring ventilation strategies to improve patient outcomes in this specific patient population. Additionally, it is vital to account for the influence of sedatives, neurological scores, and BMI on respiratory drive and occlusion pressures to ensure a comprehensive analysis.Keywords: brain damage, diaphragm dysfunction, occlusion pressure, p0.1, respiratory drive
Procedia PDF Downloads 682540 Feature Engineering Based Detection of Buffer Overflow Vulnerability in Source Code Using Deep Neural Networks
Authors: Mst Shapna Akter, Hossain Shahriar
Abstract:
One of the most important challenges in the field of software code audit is the presence of vulnerabilities in software source code. Every year, more and more software flaws are found, either internally in proprietary code or revealed publicly. These flaws are highly likely exploited and lead to system compromise, data leakage, or denial of service. C and C++ open-source code are now available in order to create a largescale, machine-learning system for function-level vulnerability identification. We assembled a sizable dataset of millions of opensource functions that point to potential exploits. We developed an efficient and scalable vulnerability detection method based on deep neural network models that learn features extracted from the source codes. The source code is first converted into a minimal intermediate representation to remove the pointless components and shorten the dependency. Moreover, we keep the semantic and syntactic information using state-of-the-art word embedding algorithms such as glove and fastText. The embedded vectors are subsequently fed into deep learning networks such as LSTM, BilSTM, LSTM-Autoencoder, word2vec, BERT, and GPT-2 to classify the possible vulnerabilities. Furthermore, we proposed a neural network model which can overcome issues associated with traditional neural networks. Evaluation metrics such as f1 score, precision, recall, accuracy, and total execution time have been used to measure the performance. We made a comparative analysis between results derived from features containing a minimal text representation and semantic and syntactic information. We found that all of the deep learning models provide comparatively higher accuracy when we use semantic and syntactic information as the features but require higher execution time as the word embedding the algorithm puts on a bit of complexity to the overall system.Keywords: cyber security, vulnerability detection, neural networks, feature extraction
Procedia PDF Downloads 892539 Development, Evaluation and Scale-Up of a Mental Health Care Plan (MHCP) in Nepal
Authors: Nagendra P. Luitel, Mark J. D. Jordans
Abstract:
Globally, there is a significant gap between the number of individuals in need of mental health care and those who actually receive treatment. The evidence is accumulating that mental health services can be delivered effectively by primary health care workers through community-based programs and task-sharing approaches. Changing the role of specialist mental health workers from service delivery to building clinical capacity of the primary health care (PHC) workers could help in reducing treatment gap in low and middle-income countries (LMICs). We developed a comprehensive mental health care plan in 2012 and evaluated its feasibility and effectiveness over the past three years. Initially, a mixed method formative study was conducted for the development of mental health care plan (MHCP). Routine monitoring and evaluation data, including client flow and reports of satisfaction, were obtained from beneficiaries (n=135) during the pilot-testing phase. Repeated community survey (N=2040); facility detection survey (N=4704) and the cohort study (N=576) were conducted for evaluation of the MHCP. The resulting MHCP consists of twelve packages divided over the community, health facility, and healthcare organization platforms. Detection of mental health problems increased significantly after introducing MHCP. Service implementation data support the real-life applicability of the MHCP, with reasonable treatment uptake. Currently, MHCP has been implemented in the entire Chitwan district where over 1400 people (438 people with depression, 406 people with psychosis, 181 people with epilepsy, 360 people with alcohol use disorder and 51 others) have received mental health services from trained health workers. Key barriers were identified and addressed, namely dissatisfaction with privacy, perceived burden among health workers, high drop-out rates and continue the supply of medicines. The results indicated that involvement of PHC workers in detection and management of mental health problems is an effective strategy to minimize treatment gap on mental health care in Nepal.Keywords: mental health, Nepal, primary care, treatment gap
Procedia PDF Downloads 2952538 A New DIDS Design Based on a Combination Feature Selection Approach
Authors: Adel Sabry Eesa, Adnan Mohsin Abdulazeez Brifcani, Zeynep Orman
Abstract:
Feature selection has been used in many fields such as classification, data mining and object recognition and proven to be effective for removing irrelevant and redundant features from the original data set. In this paper, a new design of distributed intrusion detection system using a combination feature selection model based on bees and decision tree. Bees algorithm is used as the search strategy to find the optimal subset of features, whereas decision tree is used as a judgment for the selected features. Both the produced features and the generated rules are used by Decision Making Mobile Agent to decide whether there is an attack or not in the networks. Decision Making Mobile Agent will migrate through the networks, moving from node to another, if it found that there is an attack on one of the nodes, it then alerts the user through User Interface Agent or takes some action through Action Mobile Agent. The KDD Cup 99 data set is used to test the effectiveness of the proposed system. The results show that even if only four features are used, the proposed system gives a better performance when it is compared with the obtained results using all 41 features.Keywords: distributed intrusion detection system, mobile agent, feature selection, bees algorithm, decision tree
Procedia PDF Downloads 4082537 Applications of Hyperspectral Remote Sensing: A Commercial Perspective
Authors: Tuba Zahra, Aakash Parekh
Abstract:
Hyperspectral remote sensing refers to imaging of objects or materials in narrow conspicuous spectral bands. Hyperspectral images (HSI) enable the extraction of spectral signatures for objects or materials observed. These images contain information about the reflectance of each pixel across the electromagnetic spectrum. It enables the acquisition of data simultaneously in hundreds of spectral bands with narrow bandwidths and can provide detailed contiguous spectral curves that traditional multispectral sensors cannot offer. The contiguous, narrow bandwidth of hyperspectral data facilitates the detailed surveying of Earth's surface features. This would otherwise not be possible with the relatively coarse bandwidths acquired by other types of imaging sensors. Hyperspectral imaging provides significantly higher spectral and spatial resolution. There are several use cases that represent the commercial applications of hyperspectral remote sensing. Each use case represents just one of the ways that hyperspectral satellite imagery can support operational efficiency in the respective vertical. There are some use cases that are specific to VNIR bands, while others are specific to SWIR bands. This paper discusses the different commercially viable use cases that are significant for HSI application areas, such as agriculture, mining, oil and gas, defense, environment, and climate, to name a few. Theoretically, there is n number of use cases for each of the application areas, but an attempt has been made to streamline the use cases depending upon economic feasibility and commercial viability and present a review of literature from this perspective. Some of the specific use cases with respect to agriculture are crop species (sub variety) detection, soil health mapping, pre-symptomatic crop disease detection, invasive species detection, crop condition optimization, yield estimation, and supply chain monitoring at scale. Similarly, each of the industry verticals has a specific commercially viable use case that is discussed in the paper in detail.Keywords: agriculture, mining, oil and gas, defense, environment and climate, hyperspectral, VNIR, SWIR
Procedia PDF Downloads 792536 Detection of PCD-Related Transcription Factors for Improving Salt Tolerance in Plant
Authors: A. Bahieldin, A. Atef, S. Edris, N. O. Gadalla, S. M. Hassan, M. A. Al-Kordy, A. M. Ramadan, A. S. M. Al- Hajar, F. M. El-Domyati
Abstract:
The idea of this work is based on a natural exciting phenomenon suggesting that suppression of genes related to the program cell death (or PCD) mechanism might help the plant cells to efficiently tolerate abiotic stresses. The scope of this work was the detection of PCD-related transcription factors (TFs) that might also be related to salt stress tolerance in plant. Two model plants, e.g., tobacco and Arabidopsis, were utilized in order to investigate this phenomenon. Occurrence of PCD was first proven by Evans blue staining and DNA laddering after tobacco leaf discs were treated with oxalic acid (OA) treatment (20 mM) for 24 h. A number of 31 TFs up regulated after 2 h and co-expressed with genes harboring PCD-related domains were detected via RNA-Seq analysis and annotation. These TFs were knocked down via virus induced gene silencing (VIGS), an RNA interference (RNAi) approach, and tested for their influence on triggering PCD machinery. Then, Arabidopsis SALK knocked out T-DNA insertion mutants in selected TFs analogs to those in tobacco were tested under salt stress (up to 250 mM NaCl) in order to detect the influence of different TFs on conferring salt tolerance in Arabidopsis. Involvement of a number of candidate abiotic-stress related TFs was investigated.Keywords: VIGS, PCD, RNA-Seq, transcription factors
Procedia PDF Downloads 2742535 Vibratinal Spectroscopic Identification of Beta-Carotene in Usnic Acid and PAHs as a Potential Martian Analogue
Authors: A. I. Alajtal, H. G. M. Edwards, M. A. Elbagermi
Abstract:
Raman spectroscopy is currently a part of the instrumentation suite of the ESA ExoMars mission for the remote detection of life signatures in the Martian surface and subsurface. Terrestrial analogues of Martian sites have been identified and the biogeological modifications incurred as a result of extremophilic activity have been studied. Analytical instrumentation protocols for the unequivocal detection of biomarkers in suitable geological matrices are critical for future unmanned explorations, including the forthcoming ESA ExoMars mission to search for life on Mars scheduled for 2018 and Raman spectroscopy is currently a part of the Pasteur instrumentation suite of this mission. Here, Raman spectroscopy using 785nm excitation was evaluated for determining various concentrations of beta-carotene in admixture with polyaromatic hydrocarbons and usnic acid have been investigated by Raman microspectrometry to determine the lowest levels detectable in simulation of their potential identification remotely in geobiological conditions in Martian scenarios. Information from this study will be important for the development of a miniaturized Raman instrument for targetting Martian sites where the biosignatures of relict or extant life could remain in the geological record.Keywords: raman spectroscopy, mars-analog, beta-carotene, PAHs
Procedia PDF Downloads 3382534 Poly (L-Lysine)-Coated Liquid Crystal Droplets for Sensitive Detection of DNA and Its Applications in Controlled Release of Drug Molecules
Authors: Indu Verma, Santanu Kumar Pal
Abstract:
Interactions between DNA and adsorbed Poly (L-lysine) (PLL) on liquid crystal (LC) droplets were investigated using polarizing optical microcopy (POM) and epi-fluorescence microscopy. Earlier, we demonstrated that adsorption of PLL to the LC/aqueous interface resulted in homeotropic orientation of the LC and thus exhibited a radial configuration of the LC confined within the droplets. Subsequent adsorption of DNA (single stranded DNA/double stranded DNA) at PLL coated LC droplets was found to trigger a LC reorientation within the droplets leading to pre-radial/bipolar configuration of those droplets. To our surprise, subsequent exposure of complementary ssDNA (c-ssDNA) to ssDNA/ adsorbed PLL modified LC droplets did not cause the LC reorientation. This is likely due to the formation of polyplexes (DNA-PLL complex) as confirmed by fluorescence microscopy and atomic force microscopy. In addition, dsDNA adsorbed PLL droplets have been found to be effectively used to displace (controlled release) propidium iodide (a model drug) encapsulated within dsDNA over time. These observations suggest the potential for a label free droplet based LC detection system that can respond to DNA and may provide a simple method to develop DNA-based drug nano-carriers.Keywords: DNA biosensor, drug delivery, interfaces, liquid crystal droplets
Procedia PDF Downloads 2982533 Comparison of Real-Time PCR and FTIR with Chemometrics Technique in Analysing Halal Supplement Capsules
Authors: Mohd Sukri Hassan, Ahlam Inayatullah Badrul Munir, M. Husaini A. Rahman
Abstract:
Halal authentication and verification in supplement capsules are highly required as the gelatine available in the market can be from halal or non-halal sources. It is an obligation for Muslim to consume and use the halal consumer goods. At present, real-time polymerase chain reaction (RT-PCR) is the most common technique being used for the detection of porcine and bovine DNA in gelatine due to high sensitivity of the technique and higher stability of DNA compared to protein. In this study, twenty samples of supplements capsules from different products with different Halal logos were analyzed for porcine and bovine DNA using RT-PCR. Standard bovine and porcine gelatine from eurofins at a range of concentration from 10-1 to 10-5 ng/µl were used to determine the linearity range, limit of detection and specificity on RT-PCR (SYBR Green method). RT-PCR detected porcine (two samples), bovine (four samples) and mixture of porcine and bovine (six samples). The samples were also tested using FT-IR technique where normalized peak of IR spectra were pre-processed using Savitsky Golay method before Principal Components Analysis (PCA) was performed on the database. Scores plot of PCA shows three clusters of samples; bovine, porcine and mixture (bovine and porcine). The RT-PCR and FT-IR with chemometrics technique were found to give same results for porcine gelatine samples which can be used for Halal authentication.Keywords: halal, real-time PCR, gelatine, chemometrics
Procedia PDF Downloads 2412532 Development of a Device for Detecting Fluids in the Esophagus
Authors: F. J. Puertas, M. Castro, A. Tebar, P. J. Fito, R. Gadea, J. M. Monzó, R. J. Colom
Abstract:
There is a great diversity of diseases that affect the integrity of the walls of the esophagus, generally of a digestive nature. Among them, gastroesophageal reflux is a common disease in the general population, affecting the patient's quality of life; however, there are still unmet diagnostic and therapeutic issues. The consequences of untreated or asymptomatic acid reflux on the esophageal mucosa are not only pain, heartburn, and acid regurgitation but also an increased risk of esophageal cancer. Currently, the diagnostic methods to detect problems in the esophageal tract are invasive and annoying, as 24-hour impedance-pH monitoring forces the patient to be uncomfortable for hours to be able to make a correct diagnosis. In this work, the development of a sensor able to measure in depth is proposed, allowing the detection of liquids circulating in the esophageal tract. The multisensor detection system is based on radiofrequency photospectrometry. At an experimental level, consumers representative of the population in terms of sex and age have been used, placing the sensors between the trachea and the diaphragm analyzing the measurements in vacuum, water, orange juice and saline medium. The results obtained have allowed us to detect the appearance of different liquid media in the esophagus, segregating them based on their ionic content.Keywords: bioimpedance, dielectric spectroscopy, gastroesophageal reflux, GERD
Procedia PDF Downloads 1022531 Enhancing Financial Security: Real-Time Anomaly Detection in Financial Transactions Using Machine Learning
Authors: Ali Kazemi
Abstract:
The digital evolution of financial services, while offering unprecedented convenience and accessibility, has also escalated the vulnerabilities to fraudulent activities. In this study, we introduce a distinct approach to real-time anomaly detection in financial transactions, aiming to fortify the defenses of banking and financial institutions against such threats. Utilizing unsupervised machine learning algorithms, specifically autoencoders and isolation forests, our research focuses on identifying irregular patterns indicative of fraud within transactional data, thus enabling immediate action to prevent financial loss. The data we used in this study included the monetary value of each transaction. This is a crucial feature as fraudulent transactions may have distributions of different amounts than legitimate ones, such as timestamps indicating when transactions occurred. Analyzing transactions' temporal patterns can reveal anomalies (e.g., unusual activity in the middle of the night). Also, the sector or category of the merchant where the transaction occurred, such as retail, groceries, online services, etc. Specific categories may be more prone to fraud. Moreover, the type of payment used (e.g., credit, debit, online payment systems). Different payment methods have varying risk levels associated with fraud. This dataset, anonymized to ensure privacy, reflects a wide array of transactions typical of a global banking institution, ranging from small-scale retail purchases to large wire transfers, embodying the diverse nature of potentially fraudulent activities. By engineering features that capture the essence of transactions, including normalized amounts and encoded categorical variables, we tailor our data to enhance model sensitivity to anomalies. The autoencoder model leverages its reconstruction error mechanism to flag transactions that deviate significantly from the learned normal pattern, while the isolation forest identifies anomalies based on their susceptibility to isolation from the dataset's majority. Our experimental results, validated through techniques such as k-fold cross-validation, are evaluated using precision, recall, and the F1 score alongside the area under the receiver operating characteristic (ROC) curve. Our models achieved an F1 score of 0.85 and a ROC AUC of 0.93, indicating high accuracy in detecting fraudulent transactions without excessive false positives. This study contributes to the academic discourse on financial fraud detection and provides a practical framework for banking institutions seeking to implement real-time anomaly detection systems. By demonstrating the effectiveness of unsupervised learning techniques in a real-world context, our research offers a pathway to significantly reduce the incidence of financial fraud, thereby enhancing the security and trustworthiness of digital financial services.Keywords: anomaly detection, financial fraud, machine learning, autoencoders, isolation forest, transactional data analysis
Procedia PDF Downloads 572530 Smart-Textile Containers for Urban Mobility
Authors: René Vieroth, Christian Dils, M. V. Krshiwoblozki, Christine Kallmayer, Martin Schneider-Ramelow, Klaus-Dieter Lang
Abstract:
Green urban mobility in commercial and private contexts is one of the great challenges for the continuously growing cities all over the world. Bicycle based solutions are already and since a long time the key to success. Modern developments like e-bikes and high-end cargo-bikes complement the portfolio. Weight, aerodynamic drag, and security for the transported goods are the key factors for working solutions. Recent achievements in the field of smart-textiles allowed the creation of a totally new generation of intelligent textile cargo containers, which fulfill those demands. The fusion of technical textiles, design and electrical engineering made it possible to create an ecological solution which is very near to become a product. This paper shows all the details of this solution that includes an especially developed sensor textile for cut detection, a protective textile layer for intrusion prevention, an universal-charging-unit for energy harvesting from diverse sources and a low-energy alarm system with GSM/GPRS connection, GPS location and RFID interface.Keywords: cargo-bike, cut-detection, e-bike, energy-harvesting, green urban mobility, logistics, smart-textiles, textile-integrity sensor
Procedia PDF Downloads 3152529 Antimicrobial Effect of Natamycin against Food Spoilage Fungi and Yeast Contaminated Fermented Foods
Authors: Pervin Basaran Akocak
Abstract:
Food antimicrobials are compounds that are incorporated into food matrixes in order to cause death or delay the growth of spoilage or pathogenic microorganisms. As a result, microbiological deterioration is prevented throughout storage and food distribution. In this study, the effect of natural antimycotic natamycin (C33H47NO13, with a molecular mass of 665.725), a GRAS (Generally Recognized As Safe) commercial compound produced by different strains of Streptomyces sp., was tested against various fermented food contamination fungi and yeast species. At the concentration of 100 µg/ml, natamycin exhibited stronger antifungal activity against fungi than yeast species tested. The exposure time of natamycin for complete inhibition of the species tested were found to be between 100-180 min at 300-750 µg/ml concentration. SEM observations of fungal species demonstrated that natamycin distorted and damaged the conidia and hyphae by inhibiting spore germination and mycelial growth. Natamycin can be considered as a potential candidate in hurdle food treatments for preventing fungal and yeast invasion and resulting deterioration of fermented products.Keywords: natamycin, antifungal, fermented food, food spoilage fungi
Procedia PDF Downloads 5152528 An Approach to Low Velocity Impact Damage Modelling of Variable Stiffness Curved Composite Plates
Authors: Buddhi Arachchige, Hessam Ghasemnejad
Abstract:
In this study, the post impact behavior of curved composite plates subjected to low velocity impact was studied analytically and numerically. Approaches to damage modelling are proposed through the degradation of stiffness in the damaged region by reduction of thickness in the damage region. Spring-mass models were used to model the impact response of the plate and impactor. The study involved designing two damage models to compare and contrast the model best fitted with the numerical results. The theoretical force-time responses were compared with the numerical results obtained through a detailed study carried out in LS-DYNA. The modified damage model established a good prediction with the analytical force-time response for different layups and geometry. This study provides a gateway in selecting the most effective layups for variable stiffness curved composite panels able to withstand a higher impact damage.Keywords: analytical modelling, composite damage, impact, variable stiffness
Procedia PDF Downloads 2772527 Resilience-Based Emergency Bridge Inspection Routing and Repair Scheduling under Uncertainty
Authors: Zhenyu Zhang, Hsi-Hsien Wei
Abstract:
Highway network systems play a vital role in disaster response for disaster-damaged areas. Damaged bridges in such network systems can impede disaster response by disrupting transportation of rescue teams or humanitarian supplies. Therefore, emergency inspection and repair of bridges to quickly collect damage information of bridges and recover the functionality of highway networks is of paramount importance to disaster response. A widely used measure of a network’s capability to recover from disasters is resilience. To enhance highway network resilience, plenty of studies have developed various repair scheduling methods for the prioritization of bridge-repair tasks. These methods assume that repair activities are performed after the damage to a highway network is fully understood via inspection, although inspecting all bridges in a regional highway network may take days, leading to the significant delay in repairing bridges. In reality, emergency repair activities can be commenced as soon as the damage data of some bridges that are crucial to emergency response are obtained. Given that emergency bridge inspection and repair (EBIR) activities are executed simultaneously in the response phase, the real-time interactions between these activities can occur – the blockage of highways due to repair activities can affect inspection routes which in turn have an impact on emergency repair scheduling by providing real-time information on bridge damages. However, the impact of such interactions on the optimal emergency inspection routes (EIR) and emergency repair schedules (ERS) has not been discussed in prior studies. To overcome the aforementioned deficiencies, this study develops a routing and scheduling model for EBIR while accounting for real-time inspection-repair interactions to maximize highway network resilience. A stochastic, time-dependent integer program is proposed for the complex and real-time interacting EBIR problem given multiple inspection and repair teams at locations as set post-disaster. A hybrid genetic algorithm that integrates a heuristic approach into a traditional genetic algorithm to accelerate the evolution process is developed. Computational tests are performed using data from the 2008 Wenchuan earthquake, based on a regional highway network in Sichuan, China, consisting of 168 highway bridges on 36 highways connecting 25 cities/towns. The results show that the simultaneous implementation of bridge inspection and repair activities can significantly improve the highway network resilience. Moreover, the deployment of inspection and repair teams should match each other, and the network resilience will not be improved once the unilateral increase in inspection teams or repair teams exceeds a certain level. This study contributes to both knowledge and practice. First, the developed mathematical model makes it possible for capturing the impact of real-time inspection-repair interactions on inspection routing and repair scheduling and efficiently deriving optimal EIR and ERS on a large and complex highway network. Moreover, this study contributes to the organizational dimension of highway network resilience by providing optimal strategies for highway bridge management. With the decision support tool, disaster managers are able to identify the most critical bridges for disaster management and make decisions on proper inspection and repair strategies to improve highway network resilience.Keywords: disaster management, emergency bridge inspection and repair, highway network, resilience, uncertainty
Procedia PDF Downloads 1092526 Image Processing-Based Maize Disease Detection Using Mobile Application
Authors: Nathenal Thomas
Abstract:
In the food chain and in many other agricultural products, corn, also known as maize, which goes by the scientific name Zea mays subsp, is a widely produced agricultural product. Corn has the highest adaptability. It comes in many different types, is employed in many different industrial processes, and is more adaptable to different agro-climatic situations. In Ethiopia, maize is among the most widely grown crop. Small-scale corn farming may be a household's only source of food in developing nations like Ethiopia. The aforementioned data demonstrates that the country's requirement for this crop is excessively high, and conversely, the crop's productivity is very low for a variety of reasons. The most damaging disease that greatly contributes to this imbalance between the crop's supply and demand is the corn disease. The failure to diagnose diseases in maize plant until they are too late is one of the most important factors influencing crop output in Ethiopia. This study will aid in the early detection of such diseases and support farmers during the cultivation process, directly affecting the amount of maize produced. The diseases in maize plants, such as northern leaf blight and cercospora leaf spot, have distinct symptoms that are visible. This study aims to detect the most frequent and degrading maize diseases using the most efficiently used subset of machine learning technology, deep learning so, called Image Processing. Deep learning uses networks that can be trained from unlabeled data without supervision (unsupervised). It is a feature that simulates the exercises the human brain goes through when digesting data. Its applications include speech recognition, language translation, object classification, and decision-making. Convolutional Neural Network (CNN) for Image Processing, also known as convent, is a deep learning class that is widely used for image classification, image detection, face recognition, and other problems. it will also use this algorithm as the state-of-the-art for my research to detect maize diseases by photographing maize leaves using a mobile phone.Keywords: CNN, zea mays subsp, leaf blight, cercospora leaf spot
Procedia PDF Downloads 742525 Covid Medical Imaging Trial: Utilising Artificial Intelligence to Identify Changes on Chest X-Ray of COVID
Authors: Leonard Tiong, Sonit Singh, Kevin Ho Shon, Sarah Lewis
Abstract:
Investigation into the use of artificial intelligence in radiology continues to develop at a rapid rate. During the coronavirus pandemic, the combination of an exponential increase in chest x-rays and unpredictable staff shortages resulted in a huge strain on the department's workload. There is a World Health Organisation estimate that two-thirds of the global population does not have access to diagnostic radiology. Therefore, there could be demand for a program that could detect acute changes in imaging compatible with infection to assist with screening. We generated a conventional neural network and tested its efficacy in recognizing changes compatible with coronavirus infection. Following ethics approval, a deidentified set of 77 normal and 77 abnormal chest x-rays in patients with confirmed coronavirus infection were used to generate an algorithm that could train, validate and then test itself. DICOM and PNG image formats were selected due to their lossless file format. The model was trained with 100 images (50 positive, 50 negative), validated against 28 samples (14 positive, 14 negative), and tested against 26 samples (13 positive, 13 negative). The initial training of the model involved training a conventional neural network in what constituted a normal study and changes on the x-rays compatible with coronavirus infection. The weightings were then modified, and the model was executed again. The training samples were in batch sizes of 8 and underwent 25 epochs of training. The results trended towards an 85.71% true positive/true negative detection rate and an area under the curve trending towards 0.95, indicating approximately 95% accuracy in detecting changes on chest X-rays compatible with coronavirus infection. Study limitations include access to only a small dataset and no specificity in the diagnosis. Following a discussion with our programmer, there are areas where modifications in the weighting of the algorithm can be made in order to improve the detection rates. Given the high detection rate of the program, and the potential ease of implementation, this would be effective in assisting staff that is not trained in radiology in detecting otherwise subtle changes that might not be appreciated on imaging. Limitations include the lack of a differential diagnosis and application of the appropriate clinical history, although this may be less of a problem in day-to-day clinical practice. It is nonetheless our belief that implementing this program and widening its scope to detecting multiple pathologies such as lung masses will greatly assist both the radiology department and our colleagues in increasing workflow and detection rate.Keywords: artificial intelligence, COVID, neural network, machine learning
Procedia PDF Downloads 932524 Molecular Detection and Characterization of Shiga Toxogenic Escherichia coli Associated with Dairy Product
Authors: Mohamed Al-Hazmi, Abdullah Al-Arfaj, Moussa Ihab
Abstract:
Raw, unpasteurized milk can carry dangerous bacteria such as Salmonella, E. coli, and Listeria, which are responsible for causing numerous foodborne illnesses. The objective of this study was molecular characterization of shiga toxogenic E. coli in raw milk collected from different Egyptian governorates by multiplex PCR. During the period of 25th May to 25th October 2012, a total of 320 bulk-tank milk samples were collected from 10 cow farms located in different Egyptian governorates. Bacteriological examination of milk samples revealed the presence of E. coli organisms in 65 samples (20.3%), serotyping of the E. coli isolates revealed, 35 strains (10.94%) O111, 15 strains (4.69%) O157: H7, 10 strains (3.13%) O128 and 5 strains (1.56%) O119. Multiplex PCR for detection of shiga toxin type 2 and intimin genes revealed positive amplification of 255 bp fragment of shiga toxin type 2 gene and 384 bp fragment of intimin gene from all E. coli serovar O157: H7, while from serovar O111 were 25 (71.43%), 20 (57.14%) and from serovar O128 were 6 (60%), 8 (80%), respectively. The results of multiplex PCR assay are useful for identification of STEC possessing the eaeA and stx2 genes.Keywords: raw milk, E. coli, multiplex PCR, Shiga toxin type 2, intimin gene
Procedia PDF Downloads 3062523 Change Detection of Vegetative Areas Using Land Use Land Cover Derived from NDVI of Desert Encroached Areas
Authors: T. Garba, T. O. Quddus, Y. Y. Babanyara, M. A. Modibbo
Abstract:
Desertification is define as the changing of productive land into a desert as the result of ruination of land by man-induced soil erosion, which forces famers in the affected areas to move migrate or encourage into reserved areas in search of a fertile land for their farming activities. This study therefore used remote sensing imageries to determine the level of changes in the vegetative areas. To achieve that Normalized Difference of the Vegetative Index (NDVI), classified imageries and image slicing derived from landsat TM 1986, land sat ETM 1999 and Nigeria sat 1 2007 were used to determine changes in vegetations. From the Classified imageries it was discovered that there a more natural vegetation in classified images of 1986 than that of 1999 and 2007. This finding is also future in the three NDVI imageries, it was discovered that there is increased in high positive pixel value from 0.04 in 1986 to 0.22 in 1999 and to 0.32 in 2007. The figures in the three histogram also indicted that there is increased in vegetative areas from 29.15 Km2 in 1986, to 60.58 Km2 in 1999 and then to 109 Km2 in 2007. The study recommends among other things that there is need to restore natural vegetation through discouraging of farming activities in and around the natural vegetation in the study area.Keywords: vegetative index, classified imageries, change detection, landsat, vegetation
Procedia PDF Downloads 3602522 Klotho Level as a Marker of Low Bone Mineral Density in Egyptian Sickle Cell Disease Patients
Authors: Mona Hamdy, Iman Shaheen, Hadeel Seif Eldin, Basma Ali, Omnia Abdeldayem
Abstract:
Summary: Bone involvement of sickle cell disease (SCD) patients varies from acute clinical manifestations of painful vaso-occlusive crises or osteomyelitis to more chronic affection of bone mineral density (BMD) and debilitating osteonecrosis and osteoporosis. Secreted klotho protein is involved in calcium (Ca) reabsorption in the kidney. This study aimed to measure serum klotho levels in children with SCD to determine the possibility of using it as a marker of low BMD in children with SCD in correlation with a dual-energy radiograph absorptiometry scan. This study included 60 sickle disease patients and 30 age-matched and sex-matched control participants without SCD. A highly statistically significant difference was found between patients with normal BMD and those with low BMD, with serum Ca and klotho levels being lower in the latter group. Klotho serum level correlated positively with both serum Ca and BMD. Serum klotho level showed 94.9% sensitivity and 95.2% specificity in the detection of low BMD. Both serum Ca and klotho serum levels may be useful markers for detection of low BMD related to SCD with high sensitivity and specificity; however, klotho may be a better indicator as it is less affected by the nutritional and endocrinal status of patients or by intake of Ca supplements.Keywords: sickle cell disease, BMD, osteoporosis, DEXA, klotho
Procedia PDF Downloads 1042521 Conductometric Methanol Microsensor Based on Electrospun PVC-Nickel Phthalocyanine Composite Nanofiber Technology
Authors: Ibrahim Musa, Guy Raffin, Marie Hangouet, Nadia Zine, Nicole Jaffrezic-Renault, Abdelhamid Errachid
Abstract:
Due to its application in different domains, such as fuel cell configuration and adulteration of alcoholic beverages, a miniaturized sensor for methanol detection is urgently required. A conductometric microsensor for measuring volatile organic compounds (VOC) was conceived, based on electrospun composite nanofibers of polyvinyl chloride (PVC) doped with nickel phthalocyanine(NiPc) deposited on interdigitated electrodes (IDEs) used transducers. The nanofiber's shape, structure, percent atomic content and thermal properties were studied using analytical techniques, including scanning electron microscopy (SEM), Fourier transform infrared spectroscopy (FTIR), and thermogravimetric analysis (TGA), respectively. The methanol sensor showed good sensitivity (505µS/cm(v/v) ⁻¹), low LOD (15 ppm), short response time (13 s), and short recovery time (15 s). The sensor was 4 times more sensitive to methanol than to ethanol and 19 times more sensitive to methanol than to acetone. Furthermore, the sensor response was unaffected by the interfering water vapor, making it more suitable for VOC sensing in the presence of humidity. The sensor was applied for conductometric detection of methanol in rubbing alcohol.Keywords: composite, methanol, conductometric sensor, electrospun, nanofiber, nickel phthalocyanine, PVC
Procedia PDF Downloads 222520 Rapid Atmospheric Pressure Photoionization-Mass Spectrometry (APPI-MS) Method for the Detection of Polychlorinated Dibenzo-P-Dioxins and Dibenzofurans in Real Environmental Samples Collected within the Vicinity of Industrial Incinerators
Authors: M. Amo, A. Alvaro, A. Astudillo, R. Mc Culloch, J. C. del Castillo, M. Gómez, J. M. Martín
Abstract:
Polychlorinated dibenzo-p-dioxins and dibenzofurans (PCDD/Fs) of course comprise a range of highly toxic compounds that may exist as particulates within the air or accumulate within water supplies, soil, or vegetation. They may be created either ubiquitously or naturally within the environment as a product of forest fires or volcanic eruptions. It is only since the industrial revolution, however, that it has become necessary to closely monitor their generation as a byproduct of manufacturing/combustion processes, in an effort to mitigate widespread contamination events. Of course, the environmental concentrations of these toxins are expected to be extremely low, therefore highly sensitive and accurate methods are required for their determination. Since ionization of non-polar compounds through electrospray and APCI is difficult and inefficient, we evaluate the performance of a novel low-flow Atmospheric Pressure Photoionization (APPI) source for the trace detection of various dioxins and furans using rapid Mass Spectrometry workflows. Air, soil and biota (vegetable matter) samples were collected monthly during one year from various locations within the vicinity of an industrial incinerator in Spain. Analytes were extracted and concentrated using soxhlet extraction in toluene and concentrated by rotavapor and nitrogen flow. Various ionization methods as electrospray (ES) and atmospheric pressure chemical ionization (APCI) were evaluated, however, only the low-flow APPI source was capable of providing the necessary performance, in terms of sensitivity, required for detecting all targeted analytes. In total, 10 analytes including 2,3,7,8-tetrachlorodibenzodioxin (TCDD) were detected and characterized using the APPI-MS method. Both PCDDs and PCFDs were detected most efficiently in negative ionization mode. The most abundant ion always corresponded to the loss of a chlorine and addition of an oxygen, yielding [M-Cl+O]- ions. MRM methods were created in order to provide selectivity for each analyte. No chromatographic separation was employed; however, matrix effects were determined to have a negligible impact on analyte signals. Triple Quadrupole Mass Spectrometry was chosen because of its unique potential for high sensitivity and selectivity. The mass spectrometer used was a Sciex´s Qtrap3200 working in negative Multi Reacting Monitoring Mode (MRM). Typically mass detection limits were determined to be near the 1-pg level. The APPI-MS2 technology applied to the detection of PCDD/Fs allows fast and reliable atmospheric analysis, minimizing considerably operational times and costs, with respect other technologies available. In addition, the limit of detection can be easily improved using a more sensitive mass spectrometer since the background in the analysis channel is very low. The APPI developed by SEADM allows polar and non-polar compounds ionization with high efficiency and repeatability.Keywords: atmospheric pressure photoionization-mass spectrometry (APPI-MS), dioxin, furan, incinerator
Procedia PDF Downloads 2082519 Performance Analysis of a Combined Ordered Successive and Interference Cancellation Using Zero-Forcing Detection over Rayleigh Fading Channels in Mimo Systems
Authors: Jamal R. Elbergali
Abstract:
Multiple Input Multiple Output (MIMO) systems are wireless systems with multiple antenna elements at both ends of the link. Wireless communication systems demand high data rate and spectral efficiency with increased reliability. MIMO systems have been popular techniques to achieve these goals because increased data rate is possible through spatial multiplexing scheme and diversity. Spatial Multiplexing (SM) is used to achieve higher possible throughput than diversity. In this paper, we propose a Zero-Forcing (ZF) detection using a combination of Ordered Successive Interference Cancellation (OSIC) and Zero Forcing using Interference Cancellation (ZF-IC). The proposed method used an OSIC based on Signal to Noise Ratio (SNR) ordering to get the estimation of last symbol (x ̃_(N_T )), then the estimated last symbol is considered to be an input to the ZF-IC. We analyze the Bit Error Rate (BER) performance of the proposed MIMO system over Rayleigh Fading Channel, using Binary Phase Shift Keying (BPSK) modulation scheme. The results show better performance than the previous methods.Keywords: SNR, BER, BPSK, MIMO, modulation, zero forcing (ZF), OSIC, ZF-IC, spatial multiplexing (SM)
Procedia PDF Downloads 4232518 A Simple Adaptive Atomic Decomposition Voice Activity Detector Implemented by Matching Pursuit
Authors: Thomas Bryan, Veton Kepuska, Ivica Kostanic
Abstract:
A simple adaptive voice activity detector (VAD) is implemented using Gabor and gammatone atomic decomposition of speech for high Gaussian noise environments. Matching pursuit is used for atomic decomposition, and is shown to achieve optimal speech detection capability at high data compression rates for low signal to noise ratios. The most active dictionary elements found by matching pursuit are used for the signal reconstruction so that the algorithm adapts to the individual speakers dominant time-frequency characteristics. Speech has a high peak to average ratio enabling matching pursuit greedy heuristic of highest inner products to isolate high energy speech components in high noise environments. Gabor and gammatone atoms are both investigated with identical logarithmically spaced center frequencies, and similar bandwidths. The algorithm performs equally well for both Gabor and gammatone atoms with no significant statistical differences. The algorithm achieves 70% accuracy at a 0 dB SNR, 90% accuracy at a 5 dB SNR and 98% accuracy at a 20dB SNR using 30dB SNR as a reference for voice activity.Keywords: atomic decomposition, gabor, gammatone, matching pursuit, voice activity detection
Procedia PDF Downloads 2902517 Discrimination of Bio-Analytes by Using Two-Dimensional Nano Sensor Array
Authors: P. Behera, K. K. Singh, D. K. Saini, M. De
Abstract:
Implementation of 2D materials in the detection of bio analytes is highly advantageous in the field of sensing because of its high surface to volume ratio. We have designed our sensor array with different cationic two-dimensional MoS₂, where surface modification was achieved by cationic thiol ligands with different functionality. Green fluorescent protein (GFP) was chosen as signal transducers for its biocompatibility and anionic nature, which can bind to the cationic MoS₂ surface easily, followed by fluorescence quenching. The addition of bio-analyte to the sensor can decomplex the cationic MoS₂ and GFP conjugates, followed by the regeneration of GFP fluorescence. The fluorescence response pattern belongs to various analytes collected and transformed to linear discriminant analysis (LDA) for classification. At first, 15 different proteins having wide range of molecular weight and isoelectric points were successfully discriminated at 50 nM with detection limit of 1 nM. The sensor system was also executed in biofluids such as serum, where 10 different proteins at 2.5 μM were well separated. After successful discrimination of protein analytes, the sensor array was implemented for bacteria sensing. Six different bacteria were successfully classified at OD = 0.05 with a detection limit corresponding to OD = 0.005. The optimized sensor array was able to classify uropathogens from non-uropathogens in urine medium. Further, the technique was applied for discrimination of bacteria possessing resistance to different types and amounts of drugs. We found out the mechanism of sensing through optical and electrodynamic studies, which indicates the interaction between bacteria with the sensor system was mainly due to electrostatic force of interactions, but the separation of native bacteria from their drug resistant variant was due to Van der Waals forces. There are two ways bacteria can be detected, i.e., through bacterial cells and lysates. The bacterial lysates contain intracellular information and also safe to analysis as it does not contain live cells. Lysates of different drug resistant bacteria were patterned effectively from the native strain. From unknown sample analysis, we found that discrimination of bacterial cells is more sensitive than that of lysates. But the analyst can prefer bacterial lysates over live cells for safer analysis.Keywords: array-based sensing, drug resistant bacteria, linear discriminant analysis, two-dimensional MoS₂
Procedia PDF Downloads 1432516 Identifying the Structural Components of Old Buildings from Floor Plans
Authors: Shi-Yu Xu
Abstract:
The top three risk factors that have contributed to building collapses during past earthquake events in Taiwan are: "irregular floor plans or elevations," "insufficient columns in single-bay buildings," and the "weak-story problem." Fortunately, these unsound structural characteristics can be directly identified from the floor plans. However, due to the vast number of old buildings, conducting manual inspections to identify these compromised structural features in all existing structures would be time-consuming and prone to human errors. This study aims to develop an algorithm that utilizes artificial intelligence techniques to automatically pinpoint the structural components within a building's floor plans. The obtained spatial information will be utilized to construct a digital structural model of the building. This information, particularly regarding the distribution of columns in the floor plan, can then be used to conduct preliminary seismic assessments of the building. The study employs various image processing and pattern recognition techniques to enhance detection efficiency and accuracy. The study enables a large-scale evaluation of structural vulnerability for numerous old buildings, providing ample time to arrange for structural retrofitting in those buildings that are at risk of significant damage or collapse during earthquakes.Keywords: structural vulnerability detection, object recognition, seismic capacity assessment, old buildings, artificial intelligence
Procedia PDF Downloads 892515 Characterization of Mechanical Properties of Graphene-Modified Epoxy Resin for Pipeline Repair
Authors: Siti Nur Afifah Azraai, Lim Kar Sing, Nordin Yahaya, Norhazilan Md Noor
Abstract:
This experimental study consists of a characterization of epoxy grout where an amount of 2% of graphene nanoplatelets particles were added to commercial epoxy resin to evaluate their behavior regarding neat epoxy resin. Compressive tests, tensile tests and flexural tests were conducted to study the effect of graphene nanoplatelets on neat epoxy resin. By comparing graphene-based and neat epoxy grout, there is no significant increase of strength due to weak interface in the graphene nanoplatelets/epoxy composites. From this experiment, the tension and flexural strength of graphene-based epoxy grouts is slightly lower than ones of neat epoxy grout. Nevertheless, the addition of graphene has produced more consistent results according to a smaller standard deviation of strength. Furthermore, the graphene has also improved the ductility of the grout, hence reducing its brittle behaviour. This shows that the performance of graphene-based grout is reliably predictable and able to minimize sudden rupture. This is important since repair design of damaged pipeline is of deterministic nature.Keywords: composite, epoxy resin, graphene nanoplatelets, pipeline
Procedia PDF Downloads 482