Search results for: probability-based damage detection (PBDD)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5714

Search results for: probability-based damage detection (PBDD)

4844 Role of Physical Properties of Maize Grains Towards Resistance to Sitotroga Cerealella (OLIV.) (Gelechiidae: Lepidoptera) in No Choice

Authors: Sohail Ahmed, Ahmad Raza

Abstract:

Physical properties of maize grains were correlated with levels of the life history of Sitotroga cerealella (Oliv.) (Gelechiidae: Lepidoptera) in no choice test to find out relative resistance in different varieties. Eight maize varieties /lines (EV-6089, Sahiwal-2002, Golden, 34N43, EV-1098, Sultan, China-1, EV-20) including seven yellow and one white were obtained from Maize and Millet Research Institute, Yousaf Wala, Sahiwal, Punjab, Pakistan. Freshly laid eggs (one day old) of S. cerealella were obtained and cultured on a susceptible maize variety for two generations for later on shifting to test varieties. Results showed that maximum moth emergence (10.33), fecundity (35.66), hatching (87.66%), moth weight (5.05 mg), development time (36.0 days) damage (93.35%) and grain weight loss (38.84%) was found in varieties, 34N43 and Golden, Sultan, Sahiwal 2002, 34N43, EV-6089, 34N43 and EV-1089, respectively. Varieties had significant difference with other varieties in these parameters (P<0.05). The varieties had positive as well as negative correlation between hardness index, grain weight and bulk density with the biological parameters of S. cerealella, percent grain damage and weight loss. Possible involvement of these grain properties in the resistance of maize grains towards S. cerealella is discussed.

Keywords: sitotroga cerealella, hardness index, grain damage, maize, varieties

Procedia PDF Downloads 384
4843 Dynamic Background Updating for Lightweight Moving Object Detection

Authors: Kelemewerk Destalem, Joongjae Cho, Jaeseong Lee, Ju H. Park, Joonhyuk Yoo

Abstract:

Background subtraction and temporal difference are often used for moving object detection in video. Both approaches are computationally simple and easy to be deployed in real-time image processing. However, while the background subtraction is highly sensitive to dynamic background and illumination changes, the temporal difference approach is poor at extracting relevant pixels of the moving object and at detecting the stopped or slowly moving objects in the scene. In this paper, we propose a moving object detection scheme based on adaptive background subtraction and temporal difference exploiting dynamic background updates. The proposed technique consists of a histogram equalization, a linear combination of background and temporal difference, followed by the novel frame-based and pixel-based background updating techniques. Finally, morphological operations are applied to the output images. Experimental results show that the proposed algorithm can solve the drawbacks of both background subtraction and temporal difference methods and can provide better performance than that of each method.

Keywords: background subtraction, background updating, real time, light weight algorithm, temporal difference

Procedia PDF Downloads 340
4842 Analysis of Conflict and Acceptance Factors on Water and Land Photovoltaic Facility

Authors: Taehyun Kim, Taehyun Kim, Hyunjoo Park

Abstract:

Photovoltaic facility occurs conflicts and disputes over environmental issues such as soil runoff, landscapes damage, and ecosystems damage. Because of these problems, huge social and economic cost occurred. The purpose of this study is to analyze resident‘s acceptability and conflict factors on the location of PV facilities, and suggest ways to promote resident’s acceptability and solutions for conflicts. Literature review, cases analysis, and expert interview on the acceptance and conflict factors related to the location of PV facilities are used to derive results. The results of this study are expected to contribute to the minimization of environmental impact and social conflict due to the development of renewable energy in the future.

Keywords: acceptance factor, conflict factor, factor analysis, photovoltaic facility

Procedia PDF Downloads 174
4841 Stubble and Senesced Leaves Are the Primary Sites of Ice Nucleation Activity in Wheat

Authors: Amanuel Bekuma, Rebecca Swift, Sarah Jackson, Ben Biddulph

Abstract:

Economic loss to frost damage is increasing over the past years in the Western Australian Wheatbelt. Agronomic, genetic, and climatic works have still found a weak correlation between temperature and frost damage. One possibility that has not been explored within the Australian cropping system is whether ice nucleation active bacteria (INB) either present in situ on crop residue or introduced by rainfall could be responsible for the increased sensitivity of cereal plants to frost at different stages of development. This study investigated upper and lower leaf canopy, stubble, and soil as a potential site of ice nucleation activity (INA) and tracked the changes in INA during the plant development. We found that older leaves of wheat are the primary sites of ice nucleation (-4.7 to -6.3°C) followed by stubble (-5.7 to -6.7°C) which increases the risk of frost damage during heading and flowering (the most susceptible stages). However, healthy and green upper canopy leaves (flag and flag-2) and the soil have lower INA (< -11°C) during the frost-sensitive stage of wheat. We anticipate the higher INA on the stubble and older leaves to be due to the presence of biologically active ice-nucleating bacteria (INB), known to cause frost injury to sensitive plants at -5°C. Stubble retained or applied during the growing season further exacerbates additional frost risk by potentially increasing the INB load. The implications of the result for stubble and frost risk management in a frost-prone landscape will be discussed.

Keywords: frost, ice-nucleation-activity, stubble, wheat

Procedia PDF Downloads 135
4840 Aspects on the Problems of Road Asset Management and Maintenance in Albania

Authors: Diana Bardhi

Abstract:

Road safety is an essential part of the economic and social development of any industrialized country. Decisions to maintain and improve the reliability, functionality of infrastructure structures can only be achieved through integrated road life cycle planning and management. There has always been a tendency to review road maintenance strategies, but there is still no serious and reliable administration due to not only insufficient funds but also problems in the proper reorganization of this system. The safety and performance of the road system depend on the ongoing activity of road maintenance management. For it to be effective, it is necessary to intervene before the degradation has caused irreparable damage or damage with a high economic cost of repairs. Investments in road infrastructure during 2006-2014 show that the life of these projects presents problems related to the maintenance and management of life cycle performance in a wide range of constituent elements. Maintenance planning includes various problems that depend on the degree of degradation of asphalt layers, the degree of damage to road structures (bridges, tunnels, culverts, and the economic planning of resources for their repair). The purpose of this study is first to provide a brief overview of the problems in the field of maintenance and life cycle management of road infrastructure investments, proposing ways to reorganize the sector of administration and maintenance of ongoing roads and secondly testing and evaluating the work and nature of standards of different types of road infrastructure projects, through a methodology consisting of a) development, b) data collection, and c) analysis.

Keywords: infrastructure, maintenance, depreciation, efficiency

Procedia PDF Downloads 156
4839 A New Mathematical Method for Heart Attack Forecasting

Authors: Razi Khalafi

Abstract:

Myocardial Infarction (MI) or acute Myocardial Infarction (AMI), commonly known as a heart attack, occurs when blood flow stops to part of the heart causing damage to the heart muscle. An ECG can often show evidence of a previous heart attack or one that's in progress. The patterns on the ECG may indicate which part of your heart has been damaged, as well as the extent of the damage. In chaos theory, the correlation dimension is a measure of the dimensionality of the space occupied by a set of random points, often referred to as a type of fractal dimension. In this research by considering ECG signal as a random walk we work on forecasting the oncoming heart attack by analysing the ECG signals using the correlation dimension. In order to test the model a set of ECG signals for patients before and after heart attack was used and the strength of model for forecasting the behaviour of these signals were checked. Results show this methodology can forecast the ECG and accordingly heart attack with high accuracy.

Keywords: heart attack, ECG, random walk, correlation dimension, forecasting

Procedia PDF Downloads 505
4838 Financial Statement Fraud: The Need for a Paradigm Shift to Forensic Accounting

Authors: Ifedapo Francis Awolowo

Abstract:

The unrelenting series of embarrassing audit failures should stimulate a paradigm shift in accounting. And in this age of information revolution, there is need for a constant improvement on the products or services one offers to the market in order to be relevant. This study explores the perceptions of external auditors, forensic accountants and accounting academics on whether a paradigm shift to forensic accounting can reduce financial statement frauds. Through Neo-empiricism/inductive analytical approach, findings reveal that a paradigm shift to forensic accounting might be the right step in the right direction in order to increase the chances of fraud prevention and detection in the financial statement. This research has implication on accounting education on the need to incorporate forensic accounting into present day accounting curriculum. Accounting professional bodies, accounting standard setters and accounting firms all have roles to play in incorporating forensic accounting education into accounting curriculum. Particularly, there is need to alter the ISA 240 to make the prevention and detection of frauds the responsibilities of bot those charged with the management and governance of companies and statutory auditors.

Keywords: financial statement fraud, forensic accounting, fraud prevention and detection, auditing, audit expectation gap, corporate governance

Procedia PDF Downloads 365
4837 Detection and Classification Strabismus Using Convolutional Neural Network and Spatial Image Processing

Authors: Anoop T. R., Otman Basir, Robert F. Hess, Eileen E. Birch, Brooke A. Koritala, Reed M. Jost, Becky Luu, David Stager, Ben Thompson

Abstract:

Strabismus refers to a misalignment of the eyes. Early detection and treatment of strabismus in childhood can prevent the development of permanent vision loss due to abnormal development of visual brain areas. We developed a two-stage method for strabismus detection and classification based on photographs of the face. The first stage detects the presence or absence of strabismus, and the second stage classifies the type of strabismus. The first stage comprises face detection using Haar cascade, facial landmark estimation, face alignment, aligned face landmark detection, segmentation of the eye region, and detection of strabismus using VGG 16 convolution neural networks. Face alignment transforms the face to a canonical pose to ensure consistency in subsequent analysis. Using facial landmarks, the eye region is segmented from the aligned face and fed into a VGG 16 CNN model, which has been trained to classify strabismus. The CNN determines whether strabismus is present and classifies the type of strabismus (exotropia, esotropia, and vertical deviation). If stage 1 detects strabismus, the eye region image is fed into stage 2, which starts with the estimation of pupil center coordinates using mask R-CNN deep neural networks. Then, the distance between the pupil coordinates and eye landmarks is calculated along with the angle that the pupil coordinates make with the horizontal and vertical axis. The distance and angle information is used to characterize the degree and direction of the strabismic eye misalignment. This model was tested on 100 clinically labeled images of children with (n = 50) and without (n = 50) strabismus. The True Positive Rate (TPR) and False Positive Rate (FPR) of the first stage were 94% and 6% respectively. The classification stage has produced a TPR of 94.73%, 94.44%, and 100% for esotropia, exotropia, and vertical deviations, respectively. This method also had an FPR of 5.26%, 5.55%, and 0% for esotropia, exotropia, and vertical deviation, respectively. The addition of one more feature related to the location of corneal light reflections may reduce the FPR, which was primarily due to children with pseudo-strabismus (the appearance of strabismus due to a wide nasal bridge or skin folds on the nasal side of the eyes).

Keywords: strabismus, deep neural networks, face detection, facial landmarks, face alignment, segmentation, VGG 16, mask R-CNN, pupil coordinates, angle deviation, horizontal and vertical deviation

Procedia PDF Downloads 92
4836 Modified Gold Screen Printed Electrode with Ruthenium Complex for Selective Detection of Porcine DNA

Authors: Siti Aishah Hasbullah

Abstract:

Studies on identification of pork content in food have grown rapidly to meet the Halal food standard in Malaysia. The used mitochondria DNA (mtDNA) approaches for the identification of pig species is thought to be the most precise marker due to the mtDNA genes are present in thousands of copies per cell, the large variability of mtDNA. The standard method commonly used for DNA detection is based on polymerase chain reaction (PCR) method combined with gel electrophoresis but has major drawback. Its major drawbacks are laborious, need longer time and toxic to handle. Therefore, the need for simplicity and fast assay of DNA is vital and has triggered us to develop DNA biosensors for porcine DNA detection. Therefore, the aim of this project is to develop electrochemical DNA biosensor based on ruthenium (II) complex, [Ru(bpy)2(p-PIP)]2+ as DNA hybridization label. The interaction of DNA and [Ru(bpy)2(p-HPIP)]2+ will be studied by electrochemical transduction using Gold Screen-Printed Electrode (GSPE) modified with gold nanoparticles (AuNPs) and succinimide acrylic microspheres. The electrochemical detection by redox active ruthenium (II) complex was measured by cyclic voltammetry (CV) and differential pulse voltammetry (DPV). The results indicate that the interaction of [Ru(bpy)2(PIP)]2+ with hybridization complementary DNA has higher response compared to single-stranded and mismatch complementary DNA. Under optimized condition, this porcine DNA biosensor incorporated modified GSPE shows good linear range towards porcine DNA.

Keywords: gold, screen printed electrode, ruthenium, porcine DNA

Procedia PDF Downloads 307
4835 Surface-Enhanced Raman Detection in Chip-Based Chromatography via a Droplet Interface

Authors: Renata Gerhardt, Detlev Belder

Abstract:

Raman spectroscopy has attracted much attention as a structurally descriptive and label-free detection method. It is particularly suited for chemical analysis given as it is non-destructive and molecules can be identified via the fingerprint region of the spectra. In this work possibilities are investigated how to integrate Raman spectroscopy as a detection method for chip-based chromatography, making use of a droplet interface. A demanding task in lab-on-a-chip applications is the specific and sensitive detection of low concentrated analytes in small volumes. Fluorescence detection is frequently utilized but restricted to fluorescent molecules. Furthermore, no structural information is provided. Another often applied technique is mass spectrometry which enables the identification of molecules based on their mass to charge ratio. Additionally, the obtained fragmentation pattern gives insight into the chemical structure. However, it is only applicable as an end-of-the-line detection because analytes are destroyed during measurements. In contrast to mass spectrometry, Raman spectroscopy can be applied on-chip and substances can be processed further downstream after detection. A major drawback of Raman spectroscopy is the inherent weakness of the Raman signal, which is due to the small cross-sections associated with the scattering process. Enhancement techniques, such as surface enhanced Raman spectroscopy (SERS), are employed to overcome the poor sensitivity even allowing detection on a single molecule level. In SERS measurements, Raman signal intensity is improved by several orders of magnitude if the analyte is in close proximity to nanostructured metal surfaces or nanoparticles. The main gain of lab-on-a-chip technology is the building block-like ability to seamlessly integrate different functionalities, such as synthesis, separation, derivatization and detection on a single device. We intend to utilize this powerful toolbox to realize Raman detection in chip-based chromatography. By interfacing on-chip separations with a droplet generator, the separated analytes are encapsulated into numerous discrete containers. These droplets can then be injected with a silver nanoparticle solution and investigated via Raman spectroscopy. Droplet microfluidics is a sub-discipline of microfluidics which instead of a continuous flow operates with the segmented flow. Segmented flow is created by merging two immiscible phases (usually an aqueous phase and oil) thus forming small discrete volumes of one phase in the carrier phase. The study surveys different chip designs to realize coupling of chip-based chromatography with droplet microfluidics. With regards to maintaining a sufficient flow rate for chromatographic separation and ensuring stable eluent flow over the column different flow rates of eluent and oil phase are tested. Furthermore, the detection of analytes in droplets with surface enhanced Raman spectroscopy is examined. The compartmentalization of separated compounds preserves the analytical resolution since the continuous phase restricts dispersion between the droplets. The droplets are ideal vessels for the insertion of silver colloids thus making use of the surface enhancement effect and improving the sensitivity of the detection. The long-term goal of this work is the first realization of coupling chip based chromatography with droplets microfluidics to employ surface enhanced Raman spectroscopy as means of detection.

Keywords: chip-based separation, chip LC, droplets, Raman spectroscopy, SERS

Procedia PDF Downloads 243
4834 Rapid and Sensitive Detection: Biosensors as an Innovative Analytical Tools

Authors: Sylwia Baluta, Joanna Cabaj, Karol Malecha

Abstract:

The evolution of biosensors was driven by the need for faster and more versatile analytical methods for application in important areas including clinical, diagnostics, food analysis or environmental monitoring, with minimum sample pretreatment. Rapid and sensitive neurotransmitters detection is extremely important in modern medicine. These compounds mainly occur in the brain and central nervous system of mammals. Any changes in the neurotransmitters concentration may lead to many diseases, such as Parkinson’s or schizophrenia. Classical techniques of chemical analysis, despite many advantages, do not permit to obtain immediate results or automatization of measurements.

Keywords: adrenaline, biosensor, dopamine, laccase, tyrosinase

Procedia PDF Downloads 140
4833 Modelling and Detecting the Demagnetization Fault in the Permanent Magnet Synchronous Machine Using the Current Signature Analysis

Authors: Yassa Nacera, Badji Abderrezak, Saidoune Abdelmalek, Houassine Hamza

Abstract:

Several kinds of faults can occur in a permanent magnet synchronous machine (PMSM) systems: bearing faults, electrically short/open faults, eccentricity faults, and demagnetization faults. Demagnetization fault means that the strengths of permanent magnets (PM) in PMSM decrease, and it causes low output torque, which is undesirable for EVs. The fault is caused by physical damage, high-temperature stress, inverse magnetic field, and aging. Motor current signature analysis (MCSA) is a conventional motor fault detection method based on the extraction of signal features from stator current. a simulation model of the PMSM under partial demagnetization and uniform demagnetization fault was established, and different degrees of demagnetization fault were simulated. The harmonic analyses using the Fast Fourier Transform (FFT) show that the fault diagnosis method based on the harmonic wave analysis is only suitable for partial demagnetization fault of the PMSM and does not apply to uniform demagnetization fault of the PMSM.

Keywords: permanent magnet, diagnosis, demagnetization, modelling

Procedia PDF Downloads 65
4832 Detecting Indigenous Languages: A System for Maya Text Profiling and Machine Learning Classification Techniques

Authors: Alejandro Molina-Villegas, Silvia Fernández-Sabido, Eduardo Mendoza-Vargas, Fátima Miranda-Pestaña

Abstract:

The automatic detection of indigenous languages ​​in digital texts is essential to promote their inclusion in digital media. Underrepresented languages, such as Maya, are often excluded from language detection tools like Google’s language-detection library, LANGDETECT. This study addresses these limitations by developing a hybrid language detection solution that accurately distinguishes Maya (YUA) from Spanish (ES). Two strategies are employed: the first focuses on creating a profile for the Maya language within the LANGDETECT library, while the second involves training a Naive Bayes classification model with two categories, YUA and ES. The process includes comprehensive data preprocessing steps, such as cleaning, normalization, tokenization, and n-gram counting, applied to text samples collected from various sources, including articles from La Jornada Maya, a major newspaper in Mexico and the only media outlet that includes a Maya section. After the training phase, a portion of the data is used to create the YUA profile within LANGDETECT, which achieves an accuracy rate above 95% in identifying the Maya language during testing. Additionally, the Naive Bayes classifier, trained and tested on the same database, achieves an accuracy close to 98% in distinguishing between Maya and Spanish, with further validation through F1 score, recall, and logarithmic scoring, without signs of overfitting. This strategy, which combines the LANGDETECT profile with a Naive Bayes model, highlights an adaptable framework that can be extended to other underrepresented languages in future research. This fills a gap in Natural Language Processing and supports the preservation and revitalization of these languages.

Keywords: indigenous languages, language detection, Maya language, Naive Bayes classifier, natural language processing, low-resource languages

Procedia PDF Downloads 15
4831 The Impact of Recurring Events in Fake News Detection

Authors: Ali Raza, Shafiq Ur Rehman Khan, Raja Sher Afgun Usmani, Asif Raza, Basit Umair

Abstract:

Detection of Fake news and missing information is gaining popularity, especially after the advancement in social media and online news platforms. Social media platforms are the main and speediest source of fake news propagation, whereas online news websites contribute to fake news dissipation. In this study, we propose a framework to detect fake news using the temporal features of text and consider user feedback to identify whether the news is fake or not. In recent studies, the temporal features in text documents gain valuable consideration from Natural Language Processing and user feedback and only try to classify the textual data as fake or true. This research article indicates the impact of recurring and non-recurring events on fake and true news. We use two models BERT and Bi-LSTM to investigate, and it is concluded from BERT we get better results and 70% of true news are recurring and rest of 30% are non-recurring.

Keywords: natural language processing, fake news detection, machine learning, Bi-LSTM

Procedia PDF Downloads 21
4830 Evaluating the Diagnostic Accuracy of the ctDNA Methylation for Liver Cancer

Authors: Maomao Cao

Abstract:

Objective: To test the performance of ctDNA methylation for the detection of liver cancer. Methods: A total of 1233 individuals have been recruited in 2017. 15 male and 15 female samples (including 10 cases of liver cancer) were randomly selected in the present study. CfDNA was extracted by MagPure Circulating DNA Maxi Kit. The concentration of cfDNA was obtained by Qubit™ dsDNA HS Assay Kit. A pre-constructed predictive model was used to analyze methylation data and to give a predictive score for each cfDNA sample. Individuals with a predictive score greater than or equal to 80 were classified as having liver cancer. CT tests were considered the gold standard. Sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for the diagnosis of liver cancer were calculated. Results: 9 patients were diagnosed with liver cancer according to the prediction model (with high sensitivity and threshold of 80 points), with scores of 99.2, 91.9, 96.6, 92.4, 91.3, 92.5, 96.8, 91.1, and 92.2, respectively. The sensitivity, specificity, positive predictive value, and negative predictive value of ctDNA methylation for the diagnosis of liver cancer were 0.70, 0.90, 0.78, and 0.86, respectively. Conclusions: ctDNA methylation could be an acceptable diagnostic modality for the detection of liver cancer.

Keywords: liver cancer, ctDNA methylation, detection, diagnostic performance

Procedia PDF Downloads 148
4829 A Comparison of Inverse Simulation-Based Fault Detection in a Simple Robotic Rover with a Traditional Model-Based Method

Authors: Murray L. Ireland, Kevin J. Worrall, Rebecca Mackenzie, Thaleia Flessa, Euan McGookin, Douglas Thomson

Abstract:

Robotic rovers which are designed to work in extra-terrestrial environments present a unique challenge in terms of the reliability and availability of systems throughout the mission. Should some fault occur, with the nearest human potentially millions of kilometres away, detection and identification of the fault must be performed solely by the robot and its subsystems. Faults in the system sensors are relatively straightforward to detect, through the residuals produced by comparison of the system output with that of a simple model. However, faults in the input, that is, the actuators of the system, are harder to detect. A step change in the input signal, caused potentially by the loss of an actuator, can propagate through the system, resulting in complex residuals in multiple outputs. These residuals can be difficult to isolate or distinguish from residuals caused by environmental disturbances. While a more complex fault detection method or additional sensors could be used to solve these issues, an alternative is presented here. Using inverse simulation (InvSim), the inputs and outputs of the mathematical model of the rover system are reversed. Thus, for a desired trajectory, the corresponding actuator inputs are obtained. A step fault near the input then manifests itself as a step change in the residual between the system inputs and the input trajectory obtained through inverse simulation. This approach avoids the need for additional hardware on a mass- and power-critical system such as the rover. The InvSim fault detection method is applied to a simple four-wheeled rover in simulation. Additive system faults and an external disturbance force and are applied to the vehicle in turn, such that the dynamic response and sensor output of the rover are impacted. Basic model-based fault detection is then employed to provide output residuals which may be analysed to provide information on the fault/disturbance. InvSim-based fault detection is then employed, similarly providing input residuals which provide further information on the fault/disturbance. The input residuals are shown to provide clearer information on the location and magnitude of an input fault than the output residuals. Additionally, they can allow faults to be more clearly discriminated from environmental disturbances.

Keywords: fault detection, ground robot, inverse simulation, rover

Procedia PDF Downloads 305
4828 Air Handling Units Power Consumption Using Generalized Additive Model for Anomaly Detection: A Case Study in a Singapore Campus

Authors: Ju Peng Poh, Jun Yu Charles Lee, Jonathan Chew Hoe Khoo

Abstract:

The emergence of digital twin technology, a digital replica of physical world, has improved the real-time access to data from sensors about the performance of buildings. This digital transformation has opened up many opportunities to improve the management of the building by using the data collected to help monitor consumption patterns and energy leakages. One example is the integration of predictive models for anomaly detection. In this paper, we use the GAM (Generalised Additive Model) for the anomaly detection of Air Handling Units (AHU) power consumption pattern. There is ample research work on the use of GAM for the prediction of power consumption at the office building and nation-wide level. However, there is limited illustration of its anomaly detection capabilities, prescriptive analytics case study, and its integration with the latest development of digital twin technology. In this paper, we applied the general GAM modelling framework on the historical data of the AHU power consumption and cooling load of the building between Jan 2018 to Aug 2019 from an education campus in Singapore to train prediction models that, in turn, yield predicted values and ranges. The historical data are seamlessly extracted from the digital twin for modelling purposes. We enhanced the utility of the GAM model by using it to power a real-time anomaly detection system based on the forward predicted ranges. The magnitude of deviation from the upper and lower bounds of the uncertainty intervals is used to inform and identify anomalous data points, all based on historical data, without explicit intervention from domain experts. Notwithstanding, the domain expert fits in through an optional feedback loop through which iterative data cleansing is performed. After an anomalously high or low level of power consumption detected, a set of rule-based conditions are evaluated in real-time to help determine the next course of action for the facilities manager. The performance of GAM is then compared with other approaches to evaluate its effectiveness. Lastly, we discuss the successfully deployment of this approach for the detection of anomalous power consumption pattern and illustrated with real-world use cases.

Keywords: anomaly detection, digital twin, generalised additive model, GAM, power consumption, supervised learning

Procedia PDF Downloads 152
4827 Proposition Model of Micromechanical Damage to Predict Reduction in Stiffness of a Fatigued A-SMC Composite

Authors: Houssem Ayari

Abstract:

Sheet molding compounds (SMC) are high strength thermoset moulding materials reinforced with glass treated with thermocompression. SMC composites combine fibreglass resins and polyester/phenolic/vinyl and unsaturated acrylic to produce a high strength moulding compound. These materials are usually formulated to meet the performance requirements of the moulding part. In addition, the vinyl ester resins used in the new advanced SMC systems (A-SMC) have many desirable features, including mechanical properties comparable to epoxy, excellent chemical resistance and tensile resistance, and cost competitiveness. In this paper, a proposed model is used to take into account the Young modulus evolutions of advanced SMC systems (A-SMC) composite under fatigue tests. The proposed model and the used approach are in good agreement with the experimental results.

Keywords: composites SFRC, damage, fatigue, Mori-Tanaka

Procedia PDF Downloads 115
4826 Detection of Helicobacter Pylori by PCR and ELISA Methods in Patients with Hyperlipidemia

Authors: Simin Khodabakhshi, Hossein Rassi

Abstract:

Hyperlipidemia refers to any of several acquired or genetic disorders that result in a high level of lipids circulating in the blood. Helicobacter pylori infection is a contributing factor in the progression of hyperlipidemia with serum lipid changes. The aim of this study was to detect of Helicobacter pylori by PCR and serological methods in patients with hyperlipidemia. In this case-control study, 174 patients with hyperlipidemia and 174 healthy controls were studied. Also, demographics, physical and biochemical parameters were performed in all samples. The DNA extracted from blood specimens was amplified by H pylori cagA specific primers. The results show that H. pylori cagA positivity was detected in 79% of the hyperlipidemia and in 56% of the control group by ELISA test and 49% of the hyperlipidemia and in 24% of the control group by PCR test. Prevalence of H. pylori infection was significantly higher in hyperlipidemia as compared to controls. In addition, patients with hyperlipidemia had significantly higher values for triglyceride, total cholesterol, LDL-C, waist to hip ratio, body mass index, diastolic and systolic blood pressure and lower levels of HDL-C than control participants (all p < 0.0001). Our result detected the ELISA was a rapid and cost-effective detection and considering the high prevalence of cytotoxigenic H. pylori strains, cag A is suggested as a promising target for PCR and ELISA tests for detection of infection with toxigenic strains. In general, it can be concluded that molecular analysis of H. pylori cagA and clinical parameters are important in early detection of hyperlipidemia and atherosclerosis with H. pylori infection by PCR and ELISA tests.

Keywords: Helicobacter pylori, hyperlipidemia, PCR, ELISA

Procedia PDF Downloads 198
4825 Performance Degradation for the GLR Test-Statistics for Spatial Signal Detection

Authors: Olesya Bolkhovskaya, Alexander Maltsev

Abstract:

Antenna arrays are widely used in modern radio systems in sonar and communications. The solving of the detection problems of a useful signal on the background of noise is based on the GLRT method. There is a large number of problem which depends on the known a priori information. In this work, in contrast to the majority of already solved problems, it is used only difference spatial properties of the signal and noise for detection. We are analyzing the influence of the degree of non-coherence of signal and noise unhomogeneity on the performance characteristics of different GLRT statistics. The description of the signal and noise is carried out by means of the spatial covariance matrices C in the cases of different number of known information. The partially coherent signal is simulated as a plane wave with a random angle of incidence of the wave concerning a normal. Background noise is simulated as random process with uniform distribution function in each element. The results of investigation of degradation of performance characteristics for different cases are represented in this work.

Keywords: GLRT, Neumann-Pearson’s criterion, Test-statistics, degradation, spatial processing, multielement antenna array

Procedia PDF Downloads 382
4824 Protein Remote Homology Detection by Using Profile-Based Matrix Transformation Approaches

Authors: Bin Liu

Abstract:

As one of the most important tasks in protein sequence analysis, protein remote homology detection has been studied for decades. Currently, the profile-based methods show state-of-the-art performance. Position-Specific Frequency Matrix (PSFM) is widely used profile. However, there exists noise information in the profiles introduced by the amino acids with low frequencies. In this study, we propose a method to remove the noise information in the PSFM by removing the amino acids with low frequencies called Top frequency profile (TFP). Three new matrix transformation methods, including Autocross covariance (ACC) transformation, Tri-gram, and K-separated bigram (KSB), are performed on these profiles to convert them into fixed length feature vectors. Combined with Support Vector Machines (SVMs), the predictors are constructed. Evaluated on two benchmark datasets, and experimental results show that these proposed methods outperform other state-of-the-art predictors.

Keywords: protein remote homology detection, protein fold recognition, top frequency profile, support vector machines

Procedia PDF Downloads 124
4823 Alternative Approach to the Machine Vision System Operating for Solving Industrial Control Issue

Authors: M. S. Nikitenko, S. A. Kizilov, D. Y. Khudonogov

Abstract:

The paper considers an approach to a machine vision operating system combined with using a grid of light markers. This approach is used to solve several scientific and technical problems, such as measuring the capability of an apron feeder delivering coal from a lining return port to a conveyor in the technology of mining high coal releasing to a conveyor and prototyping an autonomous vehicle obstacle detection system. Primary verification of a method of calculating bulk material volume using three-dimensional modeling and validation in laboratory conditions with relative errors calculation were carried out. A method of calculating the capability of an apron feeder based on a machine vision system and a simplifying technology of a three-dimensional modelled examined measuring area with machine vision was offered. The proposed method allows measuring the volume of rock mass moved by an apron feeder using machine vision. This approach solves the volume control issue of coal produced by a feeder while working off high coal by lava complexes with release to a conveyor with accuracy applied for practical application. The developed mathematical apparatus for measuring feeder productivity in kg/s uses only basic mathematical functions such as addition, subtraction, multiplication, and division. Thus, this fact simplifies software development, and this fact expands the variety of microcontrollers and microcomputers suitable for performing tasks of calculating feeder capability. A feature of an obstacle detection issue is to correct distortions of the laser grid, which simplifies their detection. The paper presents algorithms for video camera image processing and autonomous vehicle model control based on obstacle detection machine vision systems. A sample fragment of obstacle detection at the moment of distortion with the laser grid is demonstrated.

Keywords: machine vision, machine vision operating system, light markers, measuring capability, obstacle detection system, autonomous transport

Procedia PDF Downloads 113
4822 Local Boundary Analysis for Generative Theory of Tonal Music: From the Aspect of Classic Music Melody Analysis

Authors: Po-Chun Wang, Yan-Ru Lai, Sophia I. C. Lin, Alvin W. Y. Su

Abstract:

The Generative Theory of Tonal Music (GTTM) provides systematic approaches to recognizing local boundaries of music. The rules have been implemented in some automated melody segmentation algorithms. Besides, there are also deep learning methods with GTTM features applied to boundary detection tasks. However, these studies might face constraints such as a lack of or inconsistent label data. The GTTM database is currently the most widely used GTTM database, which includes manually labeled GTTM rules and local boundaries. Even so, we found some problems with these labels. They are sometimes discrepancies with GTTM rules. In addition, since it is labeled at different times by multiple musicians, they are not within the same scope in some cases. Therefore, in this paper, we examine this database with musicians from the aspect of classical music and relabel the scores. The relabeled database - GTTM Database v2.0 - will be released for academic research usage. Despite the experimental and statistical results showing that the relabeled database is more consistent, the improvement in boundary detection is not substantial. It seems that we need more clues than GTTM rules for boundary detection in the future.

Keywords: dataset, GTTM, local boundary, neural network

Procedia PDF Downloads 144
4821 Development of an Electrochemical Aptasensor for the Detection of Human Osteopontin Protein

Authors: Sofia G. Meirinho, Luis G. Dias, António M. Peres, Lígia R. Rodrigues

Abstract:

The emerging development of electrochemical aptasen sors has enabled the easy and fast detection of protein biomarkers in standard and real samples. Biomarkers are produced by body organs or tumours and provide a measure of antigens on cell surfaces. When detected in high amounts in blood, they can be suggestive of tumour activity. These biomarkers are more often used to evaluate treatment effects or to assess the potential for metastatic disease in patients with established disease. Osteopontin (OPN) is a protein found in all body fluids and constitutes a possible biomarker because its overexpression has been related with breast cancer evolution and metastasis. Currently, biomarkers are commonly used for the development of diagnostic methods, allowing the detection of the disease in its initial stages. A previously described RNA aptamer was used in the current work to develop a simple and sensitive electrochemical aptasensor with high affinity for human OPN. The RNA aptamer was biotinylated and immobilized on a gold electrode by avidin-biotin interaction. The electrochemical signal generated from the aptamer–target molecule interaction was monitored electrochemically using cyclic voltammetry in the presence of [Fe (CN) 6]−3/− as a redox probe. The signal observed showed a current decrease due to the binding of OPN. The preliminary results showed that this aptasensor enables the detection of OPN in standard solutions, showing good selectivity towards the target in the presence of others interfering proteins such as bovine OPN and bovine serum albumin. The results gathered in the current work suggest that the proposed electrochemical aptasensor is a simple and sensitive detection tool for human OPN and so, may have future applications in cancer disease monitoring.

Keywords: osteopontin, aptamer, aptasensor, screen-printed electrode, cyclic voltammetry

Procedia PDF Downloads 428
4820 Realistic Testing Procedure of Power Swing Blocking Function in Distance Relay

Authors: Farzad Razavi, Behrooz Taheri, Mohammad Parpaei, Mehdi Mohammadi Ghalesefidi, Siamak Zarei

Abstract:

As one of the major problems in protecting large-dimension power systems, power swing and its effect on distance have caused a lot of damages to energy transfer systems in many parts of the world. Therefore, power swing has gained attentions of many researchers, which has led to invention of different methods for power swing detection. Power swing detection algorithm is highly important in distance relay, but protection relays should have general requirements such as correct fault detection, response rate, and minimization of disturbances in a power system. To ensure meeting the requirements, protection relays need different tests during development, setup, maintenance, configuration, and troubleshooting steps. This paper covers power swing scheme of the modern numerical relay protection, 7sa522 to address the effect of the different fault types on the function of the power swing blocking. In this study, it was shown that the different fault types during power swing cause different time for unblocking distance relay.

Keywords: power swing, distance relay, power system protection, relay test, transient in power system

Procedia PDF Downloads 382
4819 Development of a Semiconductor Material Based on Functionalized Graphene: Application to the Detection of Nitrogen Oxides (NOₓ)

Authors: Djamil Guettiche, Ahmed Mekki, Tighilt Fatma-Zohra, Rachid Mahmoud

Abstract:

The aim of this study was to synthesize and characterize conducting polymer composites of polypyrrole and graphene, including pristine and surface-treated graphene (PPy/GO, PPy/rGO, and PPy/rGO-ArCOOH), for use as sensitive elements in a homemade chemiresistive module for on-line detection of nitrogen oxides vapors. The chemiresistive module was prepared, characterized, and evaluated for performance. Structural and morphological characterizations of the composite were carried out using FTIR, Raman spectroscopy, and XRD analyses. After exposure to NO and NO₂ gases in both static and dynamic modes, the sensitivity, selectivity, limit of detection, and response time of the sensor were determined at ambient temperature. The resulting sensor showed high sensitivity, selectivity, and reversibility, with a low limit of detection of 1 ppm. A composite of polypyrrole and graphene functionalized with aryl 4-carboxy benzene diazonium salt was synthesized and characterized using FTIR, scanning electron microscopy, transmission electron microscopy, UV-visible, and X-ray diffraction. The PPy-rGOArCOOH composite exhibited a good electrical resistance response to NO₂ at room temperature and showed enhanced NO₂-sensing properties compared to PPy-rGO thin films. The selectivity and stability of the NO₂ sensor based on the PPy/rGO-ArCOOH nanocomposite were also investigated.

Keywords: conducting polymers, surface treated graphene, diazonium salt, polypyrrole, Nitrogen oxide sensing

Procedia PDF Downloads 76
4818 Weed Classification Using a Two-Dimensional Deep Convolutional Neural Network

Authors: Muhammad Ali Sarwar, Muhammad Farooq, Nayab Hassan, Hammad Hassan

Abstract:

Pakistan is highly recognized for its agriculture and is well known for producing substantial amounts of wheat, cotton, and sugarcane. However, some factors contribute to a decline in crop quality and a reduction in overall output. One of the main factors contributing to this decline is the presence of weed and its late detection. This process of detection is manual and demands a detailed inspection to be done by the farmer itself. But by the time detection of weed, the farmer will be able to save its cost and can increase the overall production. The focus of this research is to identify and classify the four main types of weeds (Small-Flowered Cranesbill, Chick Weed, Prickly Acacia, and Black-Grass) that are prevalent in our region’s major crops. In this work, we implemented three different deep learning techniques: YOLO-v5, Inception-v3, and Deep CNN on the same Dataset, and have concluded that deep convolutions neural network performed better with an accuracy of 97.45% for such classification. In relative to the state of the art, our proposed approach yields 2% better results. We devised the architecture in an efficient way such that it can be used in real-time.

Keywords: deep convolution networks, Yolo, machine learning, agriculture

Procedia PDF Downloads 116
4817 Development of Digital Twin Concept to Detect Abnormal Changes in Structural Behaviour

Authors: Shady Adib, Vladimir Vinogradov, Peter Gosling

Abstract:

Digital Twin (DT) technology is a new technology that appeared in the early 21st century. The DT is defined as the digital representation of living and non-living physical assets. By connecting the physical and virtual assets, data are transmitted smoothly, allowing the virtual asset to fully represent the physical asset. Although there are lots of studies conducted on the DT concept, there is still limited information about the ability of the DT models for monitoring and detecting unexpected changes in structural behaviour in real time. This is due to the large computational efforts required for the analysis and an excessively large amount of data transferred from sensors. This paper aims to develop the DT concept to be able to detect the abnormal changes in structural behaviour in real time using advanced modelling techniques, deep learning algorithms, and data acquisition systems, taking into consideration model uncertainties. finite element (FE) models were first developed offline to be used with a reduced basis (RB) model order reduction technique for the construction of low-dimensional space to speed the analysis during the online stage. The RB model was validated against experimental test results for the establishment of a DT model of a two-dimensional truss. The established DT model and deep learning algorithms were used to identify the location of damage once it has appeared during the online stage. Finally, the RB model was used again to identify the damage severity. It was found that using the RB model, constructed offline, speeds the FE analysis during the online stage. The constructed RB model showed higher accuracy for predicting the damage severity, while deep learning algorithms were found to be useful for estimating the location of damage with small severity.

Keywords: data acquisition system, deep learning, digital twin, model uncertainties, reduced basis, reduced order model

Procedia PDF Downloads 97
4816 Automatic Detection and Update of Region of Interest in Vehicular Traffic Surveillance Videos

Authors: Naydelis Brito Suárez, Deni Librado Torres Román, Fernando Hermosillo Reynoso

Abstract:

Automatic detection and generation of a dynamic ROI (Region of Interest) in vehicle traffic surveillance videos based on a static camera in Intelligent Transportation Systems is challenging for computer vision-based systems. The dynamic ROI, being a changing ROI, should capture any other moving object located outside of a static ROI. In this work, the video is represented by a Tensor model composed of a Background and a Foreground Tensor, which contains all moving vehicles or objects. The values of each pixel over a time interval are represented by time series, and some pixel rows were selected. This paper proposes a pixel entropy-based algorithm for automatic detection and generation of a dynamic ROI in traffic videos under the assumption of two types of theoretical pixel entropy behaviors: (1) a pixel located at the road shows a high entropy value due to disturbances in this zone by vehicle traffic, (2) a pixel located outside the road shows a relatively low entropy value. To study the statistical behavior of the selected pixels, detecting the entropy changes and consequently moving objects, Shannon, Tsallis, and Approximate entropies were employed. Although Tsallis entropy achieved very high results in real-time, Approximate entropy showed results slightly better but in greater time.

Keywords: convex hull, dynamic ROI detection, pixel entropy, time series, moving objects

Procedia PDF Downloads 73
4815 Integrated Lateral Flow Electrochemical Strip for Leptospirosis Diagnosis

Authors: Wanwisa Deenin, Abdulhadee Yakoh, Chahya Kreangkaiwal, Orawon Chailapakul, Kanitha Patarakul, Sudkate Chaiyo

Abstract:

LipL32 is an outer membrane protein present only on pathogenic Leptospira species, which are the causative agent of leptospirosis. Leptospirosis symptoms are often misdiagnosed with other febrile illnesses as the clinical manifestations are non-specific. Therefore, an accurate diagnostic tool for leptospirosis is indeed critical for proper and prompt treatment. Typical diagnosis via serological assays is generally performed to assess the antibodies produced against Leptospira. However, their delayed antibody response and complicated procedure are undoubtedly limited the practical utilization especially in primary care setting. Here, we demonstrate for the first time an early-stage detection of LipL32 by an integrated lateral-flow immunoassay with electrochemical readout (eLFIA). A ferrocene trace tag was monitored via differential pulse voltammetry operated on a smartphone-based device, thus allowing for on-field testing. Superior performance in terms of the lowest detectable limit of detection (LOD) of 8.53 pg/mL and broad linear dynamic range (5 orders of magnitude) among other sensors available thus far was established. Additionally, the developed test strip provided a straightforward yet sensitive approach for diagnosis of leptospirosis using the collected human sera from patients, in which the results were comparable to the real-time polymerase chain reaction technique.

Keywords: leptospirosis, electrochemical detection, lateral flow immunosensor, point-of-care testing, early-stage detection

Procedia PDF Downloads 92