Search results for: edge detection method
19699 Automatic Detection Of Diabetic Retinopathy
Authors: Zaoui Ismahene, Bahri Sidi Mohamed, Abbassa Nadira
Abstract:
Diabetic Retinopathy (DR) is a leading cause of vision impairment and blindness among individuals with diabetes. Early diagnosis is crucial for effective treatment, yet current diagnostic methods rely heavily on manual analysis of retinal images, which can be time-consuming and prone to subjectivity. This research proposes an automated system for the detection of DR using Jacobi wavelet-based feature extraction combined with Support Vector Machines (SVM) for classification. The integration of wavelet analysis with machine learning techniques aims to improve the accuracy, efficiency, and reliability of DR diagnosis. In this study, retinal images are preprocessed through normalization, resizing, and noise reduction to enhance the quality of the images. The Jacobi wavelet transform is then applied to extract both global and local features, effectively capturing subtle variations in retinal images that are indicative of DR. These extracted features are fed into an SVM classifier, known for its robustness in handling high-dimensional data and its ability to achieve high classification accuracy. The SVM classifier is optimized using parameter tuning to improve performance. The proposed methodology is evaluated using a comprehensive dataset of retinal images, encompassing a range of DR severity levels. The results show that the proposed system outperforms traditional wavelet-based methods, demonstrating significantly higher accuracy, sensitivity, and specificity in detecting DR. By leveraging the discriminative power of Jacobi wavelet features and the robustness of SVM, the system provides a promising solution for the automatic detection of DR, which could assist ophthalmologists in early diagnosis and intervention, ultimately improving patient outcomes. This research highlights the potential of combining wavelet-based image processing with machine learning for advancing automated medical diagnostics.Keywords: iabetic retinopathy (DR), Jacobi wavelets, machine learning, feature extraction, classification
Procedia PDF Downloads 719698 Vibratinal Spectroscopic Identification of Beta-Carotene in Usnic Acid and PAHs as a Potential Martian Analogue
Authors: A. I. Alajtal, H. G. M. Edwards, M. A. Elbagermi
Abstract:
Raman spectroscopy is currently a part of the instrumentation suite of the ESA ExoMars mission for the remote detection of life signatures in the Martian surface and subsurface. Terrestrial analogues of Martian sites have been identified and the biogeological modifications incurred as a result of extremophilic activity have been studied. Analytical instrumentation protocols for the unequivocal detection of biomarkers in suitable geological matrices are critical for future unmanned explorations, including the forthcoming ESA ExoMars mission to search for life on Mars scheduled for 2018 and Raman spectroscopy is currently a part of the Pasteur instrumentation suite of this mission. Here, Raman spectroscopy using 785nm excitation was evaluated for determining various concentrations of beta-carotene in admixture with polyaromatic hydrocarbons and usnic acid have been investigated by Raman microspectrometry to determine the lowest levels detectable in simulation of their potential identification remotely in geobiological conditions in Martian scenarios. Information from this study will be important for the development of a miniaturized Raman instrument for targetting Martian sites where the biosignatures of relict or extant life could remain in the geological record.Keywords: raman spectroscopy, mars-analog, beta-carotene, PAHs
Procedia PDF Downloads 34119697 Detection of Adulterants in Milk Using IoT
Authors: Shaik Mohammad Samiullah Shariff, Siva Sreenath, Sai Haripriya, Prathyusha, M. Padma Lalitha
Abstract:
The Internet of Things (IoT) is the emerging technology that has been utilized to extend the possibilities for smart dairy farming (SDF). Milk consumption is continually increasing due to the world's growing population. As a result, some providers are prone to using dishonest measures to close the supply-demand imbalance, such as adding adulterants to milk. To identify the presence of adulterants in milk, traditional testing methods necessitate the use of particular chemicals and equipment. While efficient, this method has the disadvantage of yielding difficult and time-consuming qualitative results. Furthermore, same milk sample cannot be tested for other adulterants later. As a result, this study proposes an IoT-based approach for identifying adulterants in milk by measuring electrical conductivity (EC) or Total Dissolved Solids (TDS) and PH. In order to achieve this, an Arduino UNO microcontroller is used to assess the contaminants. When there is no adulteration, the pH and TDS values of milk range from 6.45 to 6.67 and 750 to 780ppm, respectively, according to this study. Finally, the data is uploaded to the cloud via an IoT device attached to the Ubidot web platform.Keywords: internet of things (IoT), pH sensor, TDS sensor, EC sensor, industry 4.0
Procedia PDF Downloads 8019696 A New Type Safety-Door for Earthquake Disaster Prevention: Part I
Authors: Daniel Y. Abebe, Jaehyouk Choi
Abstract:
From the past earthquake events, many people get hurt at the exit while they are trying to go out of the buildings because of the exit doors are unable to be opened. The door is not opened because it deviates from its the original position. The aim of this research is to develop and evaluate a new type safety door that keeps the door frame in its original position or keeps its edge angles perpendicular during and post-earthquake. The proposed door is composed of three components: outer frame joined to the wall, inner frame (door frame) and circular hollow section connected to the inner and outer frame which is used as seismic energy dissipating device.Keywords: safety-door, earthquake disaster, low yield point steel, passive energy dissipating device, FE analysis
Procedia PDF Downloads 52719695 Development of a Device for Detecting Fluids in the Esophagus
Authors: F. J. Puertas, M. Castro, A. Tebar, P. J. Fito, R. Gadea, J. M. Monzó, R. J. Colom
Abstract:
There is a great diversity of diseases that affect the integrity of the walls of the esophagus, generally of a digestive nature. Among them, gastroesophageal reflux is a common disease in the general population, affecting the patient's quality of life; however, there are still unmet diagnostic and therapeutic issues. The consequences of untreated or asymptomatic acid reflux on the esophageal mucosa are not only pain, heartburn, and acid regurgitation but also an increased risk of esophageal cancer. Currently, the diagnostic methods to detect problems in the esophageal tract are invasive and annoying, as 24-hour impedance-pH monitoring forces the patient to be uncomfortable for hours to be able to make a correct diagnosis. In this work, the development of a sensor able to measure in depth is proposed, allowing the detection of liquids circulating in the esophageal tract. The multisensor detection system is based on radiofrequency photospectrometry. At an experimental level, consumers representative of the population in terms of sex and age have been used, placing the sensors between the trachea and the diaphragm analyzing the measurements in vacuum, water, orange juice and saline medium. The results obtained have allowed us to detect the appearance of different liquid media in the esophagus, segregating them based on their ionic content.Keywords: bioimpedance, dielectric spectroscopy, gastroesophageal reflux, GERD
Procedia PDF Downloads 10419694 Static Light Scattering Method for the Analysis of Raw Cow's Milk
Authors: V. Villa-Cruz, H. Pérez-Ladron de Guevara, J. E. Diaz-Díaz
Abstract:
Static Light Scattering (SLS) was used as a method to analyse cow's milk raw, coming from the town of Lagos de Moreno, Jalisco, Mexico. This method is based on the analysis of the dispersion of light laser produced by a set of particles in solution. Based on the above, raw milk, which contains particles of fat globules, with a diameter of 2000 nm and particles of micelles of protein with 300 nm in diameter were analyzed. For this, dilutions of commercial milk were made (1.0%, 2.0% and 3.3%) to obtain a pattern of laser light scattering and also made measurements of raw cow's milk. Readings were taken in a sweep initial angle 10° to 170°, results were analyzed with the program OriginPro 7. The SLS method gives us an estimate of the percentage of fat content in milk samples. It can be concluded that the SLS method, is a quick method of analysis to detect adulteration in raw cow's milk.Keywords: light scattering, milk analysis, adulteration in milk, micelles, OriginPro
Procedia PDF Downloads 37719693 Software Engineering Inspired Cost Estimation for Process Modelling
Authors: Felix Baumann, Aleksandar Milutinovic, Dieter Roller
Abstract:
Up to this point business process management projects in general and business process modelling projects in particular could not rely on a practical and scientifically validated method to estimate cost and effort. Especially the model development phase is not covered by a cost estimation method or model. Further phases of business process modelling starting with implementation are covered by initial solutions which are discussed in the literature. This article proposes a method of filling this gap by deriving a cost estimation method from available methods in similar domains namely software development or software engineering. Software development is regarded as closely similar to process modelling as we show. After the proposition of this method different ideas for further analysis and validation of the method are proposed. We derive this method from COCOMO II and Function Point which are established methods of effort estimation in the domain of software development. For this we lay out similarities of the software development rocess and the process of process modelling which is a phase of the Business Process Management life-cycle.Keywords: COCOMO II, busines process modeling, cost estimation method, BPM COCOMO
Procedia PDF Downloads 44419692 Enhancing Financial Security: Real-Time Anomaly Detection in Financial Transactions Using Machine Learning
Authors: Ali Kazemi
Abstract:
The digital evolution of financial services, while offering unprecedented convenience and accessibility, has also escalated the vulnerabilities to fraudulent activities. In this study, we introduce a distinct approach to real-time anomaly detection in financial transactions, aiming to fortify the defenses of banking and financial institutions against such threats. Utilizing unsupervised machine learning algorithms, specifically autoencoders and isolation forests, our research focuses on identifying irregular patterns indicative of fraud within transactional data, thus enabling immediate action to prevent financial loss. The data we used in this study included the monetary value of each transaction. This is a crucial feature as fraudulent transactions may have distributions of different amounts than legitimate ones, such as timestamps indicating when transactions occurred. Analyzing transactions' temporal patterns can reveal anomalies (e.g., unusual activity in the middle of the night). Also, the sector or category of the merchant where the transaction occurred, such as retail, groceries, online services, etc. Specific categories may be more prone to fraud. Moreover, the type of payment used (e.g., credit, debit, online payment systems). Different payment methods have varying risk levels associated with fraud. This dataset, anonymized to ensure privacy, reflects a wide array of transactions typical of a global banking institution, ranging from small-scale retail purchases to large wire transfers, embodying the diverse nature of potentially fraudulent activities. By engineering features that capture the essence of transactions, including normalized amounts and encoded categorical variables, we tailor our data to enhance model sensitivity to anomalies. The autoencoder model leverages its reconstruction error mechanism to flag transactions that deviate significantly from the learned normal pattern, while the isolation forest identifies anomalies based on their susceptibility to isolation from the dataset's majority. Our experimental results, validated through techniques such as k-fold cross-validation, are evaluated using precision, recall, and the F1 score alongside the area under the receiver operating characteristic (ROC) curve. Our models achieved an F1 score of 0.85 and a ROC AUC of 0.93, indicating high accuracy in detecting fraudulent transactions without excessive false positives. This study contributes to the academic discourse on financial fraud detection and provides a practical framework for banking institutions seeking to implement real-time anomaly detection systems. By demonstrating the effectiveness of unsupervised learning techniques in a real-world context, our research offers a pathway to significantly reduce the incidence of financial fraud, thereby enhancing the security and trustworthiness of digital financial services.Keywords: anomaly detection, financial fraud, machine learning, autoencoders, isolation forest, transactional data analysis
Procedia PDF Downloads 6019691 Smart-Textile Containers for Urban Mobility
Authors: René Vieroth, Christian Dils, M. V. Krshiwoblozki, Christine Kallmayer, Martin Schneider-Ramelow, Klaus-Dieter Lang
Abstract:
Green urban mobility in commercial and private contexts is one of the great challenges for the continuously growing cities all over the world. Bicycle based solutions are already and since a long time the key to success. Modern developments like e-bikes and high-end cargo-bikes complement the portfolio. Weight, aerodynamic drag, and security for the transported goods are the key factors for working solutions. Recent achievements in the field of smart-textiles allowed the creation of a totally new generation of intelligent textile cargo containers, which fulfill those demands. The fusion of technical textiles, design and electrical engineering made it possible to create an ecological solution which is very near to become a product. This paper shows all the details of this solution that includes an especially developed sensor textile for cut detection, a protective textile layer for intrusion prevention, an universal-charging-unit for energy harvesting from diverse sources and a low-energy alarm system with GSM/GPRS connection, GPS location and RFID interface.Keywords: cargo-bike, cut-detection, e-bike, energy-harvesting, green urban mobility, logistics, smart-textiles, textile-integrity sensor
Procedia PDF Downloads 31719690 Selection the Most Suitable Method for DNA Extraction from Muscle of Iran's Canned Tuna by Comparison of Different DNA Extraction Methods
Authors: Marjan Heidarzadeh
Abstract:
High quality and purity of DNA isolated from canned tuna is essential for species identification. In this study, the efficiency of five different methods for DNA extraction was compared. Method of national standard in Iran, the CTAB precipitation method, Wizard DNA Clean Up system, Nucleospin and GenomicPrep were employed. DNA was extracted from two different canned tuna in brine and oil of the same tuna species. Three samples of each type of product were analyzed with the different methods. The quantity and quality of DNA extracted was evaluated using the 260 nm absorbance and ratio A260/A280 by spectrophotometer picodrop. Results showed that the DNA extraction from canned tuna preserved in different liquid media could be optimized by employing a specific DNA extraction method in each case. Best results were obtained with CTAB method for canned tuna in oil and with Wizard method for canned tuna in brine.Keywords: canned tuna PCR, DNA, DNA extraction methods, species identification
Procedia PDF Downloads 65919689 Classification of Digital Chest Radiographs Using Image Processing Techniques to Aid in Diagnosis of Pulmonary Tuberculosis
Authors: A. J. S. P. Nileema, S. Kulatunga , S. H. Palihawadana
Abstract:
Computer aided detection (CAD) system was developed for the diagnosis of pulmonary tuberculosis using digital chest X-rays with MATLAB image processing techniques using a statistical approach. The study comprised of 200 digital chest radiographs collected from the National Hospital for Respiratory Diseases - Welisara, Sri Lanka. Pre-processing was done to remove identification details. Lung fields were segmented and then divided into four quadrants; right upper quadrant, left upper quadrant, right lower quadrant, and left lower quadrant using the image processing techniques in MATLAB. Contrast, correlation, homogeneity, energy, entropy, and maximum probability texture features were extracted using the gray level co-occurrence matrix method. Descriptive statistics and normal distribution analysis were performed using SPSS. Depending on the radiologists’ interpretation, chest radiographs were classified manually into PTB - positive (PTBP) and PTB - negative (PTBN) classes. Features with standard normal distribution were analyzed using an independent sample T-test for PTBP and PTBN chest radiographs. Among the six features tested, contrast, correlation, energy, entropy, and maximum probability features showed a statistically significant difference between the two classes at 95% confidence interval; therefore, could be used in the classification of chest radiograph for PTB diagnosis. With the resulting value ranges of the five texture features with normal distribution, a classification algorithm was then defined to recognize and classify the quadrant images; if the texture feature values of the quadrant image being tested falls within the defined region, it will be identified as a PTBP – abnormal quadrant and will be labeled as ‘Abnormal’ in red color with its border being highlighted in red color whereas if the texture feature values of the quadrant image being tested falls outside of the defined value range, it will be identified as PTBN–normal and labeled as ‘Normal’ in blue color but there will be no changes to the image outline. The developed classification algorithm has shown a high sensitivity of 92% which makes it an efficient CAD system and with a modest specificity of 70%.Keywords: chest radiographs, computer aided detection, image processing, pulmonary tuberculosis
Procedia PDF Downloads 12819688 Critical Analysis of Heat Exchanger Cycle for its Maintainability Using Failure Modes and Effect Analysis and Pareto Analysis
Authors: Sayali Vyas, Atharva Desai, Shreyas Badave, Apurv Kulkarni, B. Rajiv
Abstract:
The Failure Modes and Effect Analysis (FMEA) is an efficient evaluation technique to identify potential failures in products, processes, and services. FMEA is designed to identify and prioritize failure modes. It proves to be a useful method for identifying and correcting possible failures at its earliest possible level so that one can avoid consequences of poor performance. In this paper, FMEA tool is used in detection of failures of various components of heat exchanger cycle and to identify critical failures of the components which may hamper the system’s performance. Further, a detailed Pareto analysis is done to find out the most critical components of the cycle, the causes of its failures, and possible recommended actions. This paper can be used as a checklist which will help in maintainability of the system.Keywords: FMEA, heat exchanger cycle, Ishikawa diagram, pareto analysis, RPN (Risk Priority Number)
Procedia PDF Downloads 40419687 Angular-Coordinate Driven Radial Tree Drawing
Authors: Farshad Ghassemi Toosi, Nikola S. Nikolov
Abstract:
We present a visualization technique for radial drawing of trees consisting of two slightly different algorithms. Both of them make use of node-link diagrams for visual encoding. This visualization creates clear drawings without edge crossing. One of the algorithms is suitable for real-time visualization of large trees, as it requires minimal recalculation of the layout if leaves are inserted or removed from the tree; while the other algorithm makes better utilization of the drawing space. The algorithms are very similar and follow almost the same procedure but with different parameters. Both algorithms assign angular coordinates for all nodes which are then converted into 2D Cartesian coordinates for visualization. We present both algorithms and discuss how they compare to each other.Keywords: Radial drawing, Visualization, Algorithm, Use of node-link diagrams
Procedia PDF Downloads 34319686 An Optimized Method for Calculating the Linear and Nonlinear Response of SDOF System Subjected to an Arbitrary Base Excitation
Authors: Hossein Kabir, Mojtaba Sadeghi
Abstract:
Finding the linear and nonlinear responses of a typical single-degree-of-freedom system (SDOF) is always being regarded as a time-consuming process. This study attempts to provide modifications in the renowned Newmark method in order to make it more time efficient than it used to be and make it more accurate by modifying the system in its own non-linear state. The efficacy of the presented method is demonstrated by assigning three base excitations such as Tabas 1978, El Centro 1940, and MEXICO CITY/SCT 1985 earthquakes to a SDOF system, that is, SDOF, to compute the strength reduction factor, yield pseudo acceleration, and ductility factor.Keywords: single-degree-of-freedom system (SDOF), linear acceleration method, nonlinear excited system, equivalent displacement method, equivalent energy method
Procedia PDF Downloads 32319685 Image Processing-Based Maize Disease Detection Using Mobile Application
Authors: Nathenal Thomas
Abstract:
In the food chain and in many other agricultural products, corn, also known as maize, which goes by the scientific name Zea mays subsp, is a widely produced agricultural product. Corn has the highest adaptability. It comes in many different types, is employed in many different industrial processes, and is more adaptable to different agro-climatic situations. In Ethiopia, maize is among the most widely grown crop. Small-scale corn farming may be a household's only source of food in developing nations like Ethiopia. The aforementioned data demonstrates that the country's requirement for this crop is excessively high, and conversely, the crop's productivity is very low for a variety of reasons. The most damaging disease that greatly contributes to this imbalance between the crop's supply and demand is the corn disease. The failure to diagnose diseases in maize plant until they are too late is one of the most important factors influencing crop output in Ethiopia. This study will aid in the early detection of such diseases and support farmers during the cultivation process, directly affecting the amount of maize produced. The diseases in maize plants, such as northern leaf blight and cercospora leaf spot, have distinct symptoms that are visible. This study aims to detect the most frequent and degrading maize diseases using the most efficiently used subset of machine learning technology, deep learning so, called Image Processing. Deep learning uses networks that can be trained from unlabeled data without supervision (unsupervised). It is a feature that simulates the exercises the human brain goes through when digesting data. Its applications include speech recognition, language translation, object classification, and decision-making. Convolutional Neural Network (CNN) for Image Processing, also known as convent, is a deep learning class that is widely used for image classification, image detection, face recognition, and other problems. it will also use this algorithm as the state-of-the-art for my research to detect maize diseases by photographing maize leaves using a mobile phone.Keywords: CNN, zea mays subsp, leaf blight, cercospora leaf spot
Procedia PDF Downloads 7519684 Low-Cost, Portable Optical Sensor with Regression Algorithm Models for Accurate Monitoring of Nitrites in Environments
Authors: David X. Dong, Qingming Zhang, Meng Lu
Abstract:
Nitrites enter waterways as runoff from croplands and are discharged from many industrial sites. Excessive nitrite inputs to water bodies lead to eutrophication. On-site rapid detection of nitrite is of increasing interest for managing fertilizer application and monitoring water source quality. Existing methods for detecting nitrites use spectrophotometry, ion chromatography, electrochemical sensors, ion-selective electrodes, chemiluminescence, and colorimetric methods. However, these methods either suffer from high cost or provide low measurement accuracy due to their poor selectivity to nitrites. Therefore, it is desired to develop an accurate and economical method to monitor nitrites in environments. We report a low-cost optical sensor, in conjunction with a machine learning (ML) approach to enable high-accuracy detection of nitrites in water sources. The sensor works under the principle of measuring molecular absorptions of nitrites at three narrowband wavelengths (295 nm, 310 nm, and 357 nm) in the ultraviolet (UV) region. These wavelengths are chosen because they have relatively high sensitivity to nitrites; low-cost light-emitting devices (LEDs) and photodetectors are also available at these wavelengths. A regression model is built, trained, and utilized to minimize cross-sensitivities of these wavelengths to the same analyte, thus achieving precise and reliable measurements with various interference ions. The measured absorbance data is input to the trained model that can provide nitrite concentration prediction for the sample. The sensor is built with i) a miniature quartz cuvette as the test cell that contains a liquid sample under test, ii) three low-cost UV LEDs placed on one side of the cell as light sources, with each LED providing a narrowband light, and iii) a photodetector with a built-in amplifier and an analog-to-digital converter placed on the other side of the test cell to measure the power of transmitted light. This simple optical design allows measuring the absorbance data of the sample at the three wavelengths. To train the regression model, absorbances of nitrite ions and their combination with various interference ions are first obtained at the three UV wavelengths using a conventional spectrophotometer. Then, the spectrophotometric data are inputs to different regression algorithm models for training and evaluating high-accuracy nitrite concentration prediction. Our experimental results show that the proposed approach enables instantaneous nitrite detection within several seconds. The sensor hardware costs about one hundred dollars, which is much cheaper than a commercial spectrophotometer. The ML algorithm helps to reduce the average relative errors to below 3.5% over a concentration range from 0.1 ppm to 100 ppm of nitrites. The sensor has been validated to measure nitrites at three sites in Ames, Iowa, USA. This work demonstrates an economical and effective approach to the rapid, reagent-free determination of nitrites with high accuracy. The integration of the low-cost optical sensor and ML data processing can find a wide range of applications in environmental monitoring and management.Keywords: optical sensor, regression model, nitrites, water quality
Procedia PDF Downloads 7419683 Data Mining in Healthcare for Predictive Analytics
Authors: Ruzanna Muradyan
Abstract:
Medical data mining is a crucial field in contemporary healthcare that offers cutting-edge tactics with enormous potential to transform patient care. This abstract examines how sophisticated data mining techniques could transform the healthcare industry, with a special focus on how they might improve patient outcomes. Healthcare data repositories have dynamically evolved, producing a rich tapestry of different, multi-dimensional information that includes genetic profiles, lifestyle markers, electronic health records, and more. By utilizing data mining techniques inside this vast library, a variety of prospects for precision medicine, predictive analytics, and insight production become visible. Predictive modeling for illness prediction, risk stratification, and therapy efficacy evaluations are important points of focus. Healthcare providers may use this abundance of data to tailor treatment plans, identify high-risk patient populations, and forecast disease trajectories by applying machine learning algorithms and predictive analytics. Better patient outcomes, more efficient use of resources, and early treatments are made possible by this proactive strategy. Furthermore, data mining techniques act as catalysts to reveal complex relationships between apparently unrelated data pieces, providing enhanced insights into the cause of disease, genetic susceptibilities, and environmental factors. Healthcare practitioners can get practical insights that guide disease prevention, customized patient counseling, and focused therapies by analyzing these associations. The abstract explores the problems and ethical issues that come with using data mining techniques in the healthcare industry. In order to properly use these approaches, it is essential to find a balance between data privacy, security issues, and the interpretability of complex models. Finally, this abstract demonstrates the revolutionary power of modern data mining methodologies in transforming the healthcare sector. Healthcare practitioners and researchers can uncover unique insights, enhance clinical decision-making, and ultimately elevate patient care to unprecedented levels of precision and efficacy by employing cutting-edge methodologies.Keywords: data mining, healthcare, patient care, predictive analytics, precision medicine, electronic health records, machine learning, predictive modeling, disease prognosis, risk stratification, treatment efficacy, genetic profiles, precision health
Procedia PDF Downloads 6519682 Assessing the Theoretical Suitability of Sentinel-2 and Worldview-3 Data for Hydrocarbon Mapping of Spill Events, Using Hydrocarbon Spectral Slope Model
Authors: K. Tunde Olagunju, C. Scott Allen, Freek Van Der Meer
Abstract:
Identification of hydrocarbon oil in remote sensing images is often the first step in monitoring oil during spill events. Most remote sensing methods adopt techniques for hydrocarbon identification to achieve detection in order to model an appropriate cleanup program. Identification on optical sensors does not only allow for detection but also for characterization and quantification. Until recently, in optical remote sensing, quantification and characterization are only potentially possible using high-resolution laboratory and airborne imaging spectrometers (hyperspectral data). Unlike multispectral, hyperspectral data are not freely available, as this data category is mainly obtained via airborne survey at present. In this research, two (2) operational high-resolution multispectral satellites (WorldView-3 and Sentinel-2) are theoretically assessed for their suitability for hydrocarbon characterization, using the hydrocarbon spectral slope model (HYSS). This method utilized the two most persistent hydrocarbon diagnostic/absorption features at 1.73 µm and 2.30 µm for hydrocarbon mapping on multispectral data. In this research, spectra measurement of seven (7) different hydrocarbon oils (crude and refined oil) taken on ten (10) different substrates with the use of laboratory ASD Fieldspec were convolved to Sentinel-2 and WorldView-3 resolution, using their full width half maximum (FWHM) parameter. The resulting hydrocarbon slope values obtained from the studied samples enable clear qualitative discrimination of most hydrocarbons, despite the presence of different background substrates, particularly on WorldView-3. Due to close conformity of central wavelengths and narrow bandwidths to key hydrocarbon bands used in HYSS, the statistical significance for qualitative analysis on WorldView-3 sensors for all studied hydrocarbon oil returned with 95% confidence level (P-value ˂ 0.01), except for Diesel. Using multifactor analysis of variance (MANOVA), the discriminating power of HYSS is statistically significant for most hydrocarbon-substrate combinations on Sentinel-2 and WorldView-3 FWHM, revealing the potential of these two operational multispectral sensors as rapid response tools for hydrocarbon mapping. One notable exception is highly transmissive hydrocarbons on Sentinel-2 data due to the non-conformity of spectral bands with key hydrocarbon absorptions and the relatively coarse bandwidth (> 100 nm).Keywords: hydrocarbon, oil spill, remote sensing, hyperspectral, multispectral, hydrocarbon-substrate combination, Sentinel-2, WorldView-3
Procedia PDF Downloads 21819681 Covid Medical Imaging Trial: Utilising Artificial Intelligence to Identify Changes on Chest X-Ray of COVID
Authors: Leonard Tiong, Sonit Singh, Kevin Ho Shon, Sarah Lewis
Abstract:
Investigation into the use of artificial intelligence in radiology continues to develop at a rapid rate. During the coronavirus pandemic, the combination of an exponential increase in chest x-rays and unpredictable staff shortages resulted in a huge strain on the department's workload. There is a World Health Organisation estimate that two-thirds of the global population does not have access to diagnostic radiology. Therefore, there could be demand for a program that could detect acute changes in imaging compatible with infection to assist with screening. We generated a conventional neural network and tested its efficacy in recognizing changes compatible with coronavirus infection. Following ethics approval, a deidentified set of 77 normal and 77 abnormal chest x-rays in patients with confirmed coronavirus infection were used to generate an algorithm that could train, validate and then test itself. DICOM and PNG image formats were selected due to their lossless file format. The model was trained with 100 images (50 positive, 50 negative), validated against 28 samples (14 positive, 14 negative), and tested against 26 samples (13 positive, 13 negative). The initial training of the model involved training a conventional neural network in what constituted a normal study and changes on the x-rays compatible with coronavirus infection. The weightings were then modified, and the model was executed again. The training samples were in batch sizes of 8 and underwent 25 epochs of training. The results trended towards an 85.71% true positive/true negative detection rate and an area under the curve trending towards 0.95, indicating approximately 95% accuracy in detecting changes on chest X-rays compatible with coronavirus infection. Study limitations include access to only a small dataset and no specificity in the diagnosis. Following a discussion with our programmer, there are areas where modifications in the weighting of the algorithm can be made in order to improve the detection rates. Given the high detection rate of the program, and the potential ease of implementation, this would be effective in assisting staff that is not trained in radiology in detecting otherwise subtle changes that might not be appreciated on imaging. Limitations include the lack of a differential diagnosis and application of the appropriate clinical history, although this may be less of a problem in day-to-day clinical practice. It is nonetheless our belief that implementing this program and widening its scope to detecting multiple pathologies such as lung masses will greatly assist both the radiology department and our colleagues in increasing workflow and detection rate.Keywords: artificial intelligence, COVID, neural network, machine learning
Procedia PDF Downloads 9719680 Molecular Detection and Characterization of Shiga Toxogenic Escherichia coli Associated with Dairy Product
Authors: Mohamed Al-Hazmi, Abdullah Al-Arfaj, Moussa Ihab
Abstract:
Raw, unpasteurized milk can carry dangerous bacteria such as Salmonella, E. coli, and Listeria, which are responsible for causing numerous foodborne illnesses. The objective of this study was molecular characterization of shiga toxogenic E. coli in raw milk collected from different Egyptian governorates by multiplex PCR. During the period of 25th May to 25th October 2012, a total of 320 bulk-tank milk samples were collected from 10 cow farms located in different Egyptian governorates. Bacteriological examination of milk samples revealed the presence of E. coli organisms in 65 samples (20.3%), serotyping of the E. coli isolates revealed, 35 strains (10.94%) O111, 15 strains (4.69%) O157: H7, 10 strains (3.13%) O128 and 5 strains (1.56%) O119. Multiplex PCR for detection of shiga toxin type 2 and intimin genes revealed positive amplification of 255 bp fragment of shiga toxin type 2 gene and 384 bp fragment of intimin gene from all E. coli serovar O157: H7, while from serovar O111 were 25 (71.43%), 20 (57.14%) and from serovar O128 were 6 (60%), 8 (80%), respectively. The results of multiplex PCR assay are useful for identification of STEC possessing the eaeA and stx2 genes.Keywords: raw milk, E. coli, multiplex PCR, Shiga toxin type 2, intimin gene
Procedia PDF Downloads 30919679 Gene Names Identity Recognition Using Siamese Network for Biomedical Publications
Authors: Micheal Olaolu Arowolo, Muhammad Azam, Fei He, Mihail Popescu, Dong Xu
Abstract:
As the quantity of biological articles rises, so does the number of biological route figures. Each route figure shows gene names and relationships. Annotating pathway diagrams manually is time-consuming. Advanced image understanding models could speed up curation, but they must be more precise. There is rich information in biological pathway figures. The first step to performing image understanding of these figures is to recognize gene names automatically. Classical optical character recognition methods have been employed for gene name recognition, but they are not optimized for literature mining data. This study devised a method to recognize an image bounding box of gene name as a photo using deep Siamese neural network models to outperform the existing methods using ResNet, DenseNet and Inception architectures, the results obtained about 84% accuracy.Keywords: biological pathway, gene identification, object detection, Siamese network
Procedia PDF Downloads 29619678 Change Detection of Vegetative Areas Using Land Use Land Cover Derived from NDVI of Desert Encroached Areas
Authors: T. Garba, T. O. Quddus, Y. Y. Babanyara, M. A. Modibbo
Abstract:
Desertification is define as the changing of productive land into a desert as the result of ruination of land by man-induced soil erosion, which forces famers in the affected areas to move migrate or encourage into reserved areas in search of a fertile land for their farming activities. This study therefore used remote sensing imageries to determine the level of changes in the vegetative areas. To achieve that Normalized Difference of the Vegetative Index (NDVI), classified imageries and image slicing derived from landsat TM 1986, land sat ETM 1999 and Nigeria sat 1 2007 were used to determine changes in vegetations. From the Classified imageries it was discovered that there a more natural vegetation in classified images of 1986 than that of 1999 and 2007. This finding is also future in the three NDVI imageries, it was discovered that there is increased in high positive pixel value from 0.04 in 1986 to 0.22 in 1999 and to 0.32 in 2007. The figures in the three histogram also indicted that there is increased in vegetative areas from 29.15 Km2 in 1986, to 60.58 Km2 in 1999 and then to 109 Km2 in 2007. The study recommends among other things that there is need to restore natural vegetation through discouraging of farming activities in and around the natural vegetation in the study area.Keywords: vegetative index, classified imageries, change detection, landsat, vegetation
Procedia PDF Downloads 36419677 Klotho Level as a Marker of Low Bone Mineral Density in Egyptian Sickle Cell Disease Patients
Authors: Mona Hamdy, Iman Shaheen, Hadeel Seif Eldin, Basma Ali, Omnia Abdeldayem
Abstract:
Summary: Bone involvement of sickle cell disease (SCD) patients varies from acute clinical manifestations of painful vaso-occlusive crises or osteomyelitis to more chronic affection of bone mineral density (BMD) and debilitating osteonecrosis and osteoporosis. Secreted klotho protein is involved in calcium (Ca) reabsorption in the kidney. This study aimed to measure serum klotho levels in children with SCD to determine the possibility of using it as a marker of low BMD in children with SCD in correlation with a dual-energy radiograph absorptiometry scan. This study included 60 sickle disease patients and 30 age-matched and sex-matched control participants without SCD. A highly statistically significant difference was found between patients with normal BMD and those with low BMD, with serum Ca and klotho levels being lower in the latter group. Klotho serum level correlated positively with both serum Ca and BMD. Serum klotho level showed 94.9% sensitivity and 95.2% specificity in the detection of low BMD. Both serum Ca and klotho serum levels may be useful markers for detection of low BMD related to SCD with high sensitivity and specificity; however, klotho may be a better indicator as it is less affected by the nutritional and endocrinal status of patients or by intake of Ca supplements.Keywords: sickle cell disease, BMD, osteoporosis, DEXA, klotho
Procedia PDF Downloads 10719676 A Semi-Implicit Phase Field Model for Droplet Evolution
Authors: M. H. Kazemi, D. Salac
Abstract:
A semi-implicit phase field method for droplet evolution is proposed. Using the phase field Cahn-Hilliard equation, we are able to track the interface in multiphase flow. The idea of a semi-implicit finite difference scheme is reviewed and employed to solve two nonlinear equations, including the Navier-Stokes and the Cahn-Hilliard equations. The use of a semi-implicit method allows us to have larger time steps compared to explicit schemes. The governing equations are coupled and then solved by a GMRES solver (generalized minimal residual method) using modified Gram-Schmidt orthogonalization. To show the validity of the method, we apply the method to the simulation of a rising droplet, a leaky dielectric drop and the coalescence of drops. The numerical solutions to the phase field model match well with existing solutions over a defined range of variables.Keywords: coalescence, leaky dielectric, numerical method, phase field, rising droplet, semi-implicit method
Procedia PDF Downloads 48619675 Conductometric Methanol Microsensor Based on Electrospun PVC-Nickel Phthalocyanine Composite Nanofiber Technology
Authors: Ibrahim Musa, Guy Raffin, Marie Hangouet, Nadia Zine, Nicole Jaffrezic-Renault, Abdelhamid Errachid
Abstract:
Due to its application in different domains, such as fuel cell configuration and adulteration of alcoholic beverages, a miniaturized sensor for methanol detection is urgently required. A conductometric microsensor for measuring volatile organic compounds (VOC) was conceived, based on electrospun composite nanofibers of polyvinyl chloride (PVC) doped with nickel phthalocyanine(NiPc) deposited on interdigitated electrodes (IDEs) used transducers. The nanofiber's shape, structure, percent atomic content and thermal properties were studied using analytical techniques, including scanning electron microscopy (SEM), Fourier transform infrared spectroscopy (FTIR), and thermogravimetric analysis (TGA), respectively. The methanol sensor showed good sensitivity (505µS/cm(v/v) ⁻¹), low LOD (15 ppm), short response time (13 s), and short recovery time (15 s). The sensor was 4 times more sensitive to methanol than to ethanol and 19 times more sensitive to methanol than to acetone. Furthermore, the sensor response was unaffected by the interfering water vapor, making it more suitable for VOC sensing in the presence of humidity. The sensor was applied for conductometric detection of methanol in rubbing alcohol.Keywords: composite, methanol, conductometric sensor, electrospun, nanofiber, nickel phthalocyanine, PVC
Procedia PDF Downloads 2619674 In vitro Evaluation of Capsaicin Patches for Transdermal Drug Delivery
Authors: Alija Uzunovic, Sasa Pilipovic, Aida Sapcanin, Zahida Ademovic, Berina Pilipović
Abstract:
Capsaicin is a naturally occurring alkaloid extracted from capsicum fruit extracts of different of Capsicum species. It has been employed topically to treat many diseases such as rheumatoid arthritis, osteoarthritis, cancer pain and nerve pain in diabetes. The high degree of pre-systemic metabolism of intragastrical capsaicin and the short half-life of capsaicin by intravenous administration made topical application of capsaicin advantageous. In this study, we have evaluated differences in the dissolution characteristics of capsaicin patch 11 mg (purchased from market) at different dissolution rotation speed. The proposed patch area is 308 cm2 (22 cm x 14 cm; it contains 36 µg of capsaicin per square centimeter of adhesive). USP Apparatus 5 (Paddle Over Disc) is used for transdermal patch testing. The dissolution study was conducted using USP apparatus 5 (n=6), ERWEKA DT800 dissolution tester (paddle-type) with addition of a disc. The fabricated patch of 308 cm2 is to be cut into 9 cm2 was placed against a disc (delivery side up) retained with the stainless-steel screen and exposed to 500 mL of phosphate buffer solution pH 7.4. All dissolution studies were carried out at 32 ± 0.5 °C and different rotation speed (50± 5; 100± 5 and 150± 5 rpm). 5 ml aliquots of samples were withdrawn at various time intervals (1, 4, 8 and 12 hours) and replaced with 5 ml of dissolution medium. Withdrawn were appropriately diluted and analyzed by reversed-phase liquid chromatography (RP-LC). A Reversed Phase Liquid Chromatography (RP-LC) method has been developed, optimized and validated for the separation and quantitation of capsaicin in a transdermal patch. The method uses a ProntoSIL 120-3-C18 AQ 125 x 4,0 mm (3 μm) column maintained at 600C. The mobile phase consisted of acetonitrile: water (50:50 v/v), the flow rate of 0.9 mL/min, the injection volume 10 μL and the detection wavelength 222 nm. The used RP-LC method is simple, sensitive and accurate and can be applied for fast (total chromatographic run time was 4.0 minutes) and simultaneous analysis of capsaicin and dihydrocapsaicin in a transdermal patch. According to the results obtained in this study, we can conclude that the relative difference of dissolution rate of capsaicin after 12 hours was elevated by increase of dissolution rotation speed (100 rpm vs 50 rpm: 84.9± 11.3% and 150 rpm vs 100 rpm: 39.8± 8.3%). Although several apparatus and procedures (USP apparatus 5, 6, 7 and a paddle over extraction cell method) have been used to study in vitro release characteristics of transdermal patches, USP Apparatus 5 (Paddle Over Disc) could be considered as a discriminatory test. would be able to point out the differences in the dissolution rate of capsaicin at different rotation speed.Keywords: capsaicin, in vitro, patch, RP-LC, transdermal
Procedia PDF Downloads 23119673 A Simple Adaptive Atomic Decomposition Voice Activity Detector Implemented by Matching Pursuit
Authors: Thomas Bryan, Veton Kepuska, Ivica Kostanic
Abstract:
A simple adaptive voice activity detector (VAD) is implemented using Gabor and gammatone atomic decomposition of speech for high Gaussian noise environments. Matching pursuit is used for atomic decomposition, and is shown to achieve optimal speech detection capability at high data compression rates for low signal to noise ratios. The most active dictionary elements found by matching pursuit are used for the signal reconstruction so that the algorithm adapts to the individual speakers dominant time-frequency characteristics. Speech has a high peak to average ratio enabling matching pursuit greedy heuristic of highest inner products to isolate high energy speech components in high noise environments. Gabor and gammatone atoms are both investigated with identical logarithmically spaced center frequencies, and similar bandwidths. The algorithm performs equally well for both Gabor and gammatone atoms with no significant statistical differences. The algorithm achieves 70% accuracy at a 0 dB SNR, 90% accuracy at a 5 dB SNR and 98% accuracy at a 20dB SNR using 30dB SNR as a reference for voice activity.Keywords: atomic decomposition, gabor, gammatone, matching pursuit, voice activity detection
Procedia PDF Downloads 29619672 Teacher’s Perception of Dalcroze Method Course as Teacher’s Enhancement Course: A Case Study in Hong Kong
Authors: Ka Lei Au
Abstract:
The Dalcroze method has been emerging in music classrooms, and music teachers are encouraged to integrate music and movement in their teaching. Music programs in colleges in Hong Kong have been introducing method courses such as Orff and Dalcroze method in music teaching as teacher’s education program. Since the targeted students of the course are music teachers who are making the decision of what approach to use in their classroom, their perception is significantly valued to identify how this approach is applicable in their teaching in regards to the teaching and learning culture and environment. This qualitative study aims to explore how the Dalcroze method as a teacher’s education course is perceived by music teachers from three aspects: 1) application in music teaching, 2) self-enhancement, 3) expectation. Through the lens of music teachers, data were collected from 30 music teachers who are taking the Dalcroze method course in music teaching in Hong Kong by the survey. The findings reveal the value and their intention of the Dalcroze method in Hong Kong. It also provides a significant reference for better development of such courses in the future in adaption to the culture, teaching and learning environment and teacher’s, student’s and parent’s perception of this approach.Keywords: Dalcroze method, music teaching, perception, self-enhancement, teacher’s education
Procedia PDF Downloads 40719671 Discrimination of Bio-Analytes by Using Two-Dimensional Nano Sensor Array
Authors: P. Behera, K. K. Singh, D. K. Saini, M. De
Abstract:
Implementation of 2D materials in the detection of bio analytes is highly advantageous in the field of sensing because of its high surface to volume ratio. We have designed our sensor array with different cationic two-dimensional MoS₂, where surface modification was achieved by cationic thiol ligands with different functionality. Green fluorescent protein (GFP) was chosen as signal transducers for its biocompatibility and anionic nature, which can bind to the cationic MoS₂ surface easily, followed by fluorescence quenching. The addition of bio-analyte to the sensor can decomplex the cationic MoS₂ and GFP conjugates, followed by the regeneration of GFP fluorescence. The fluorescence response pattern belongs to various analytes collected and transformed to linear discriminant analysis (LDA) for classification. At first, 15 different proteins having wide range of molecular weight and isoelectric points were successfully discriminated at 50 nM with detection limit of 1 nM. The sensor system was also executed in biofluids such as serum, where 10 different proteins at 2.5 μM were well separated. After successful discrimination of protein analytes, the sensor array was implemented for bacteria sensing. Six different bacteria were successfully classified at OD = 0.05 with a detection limit corresponding to OD = 0.005. The optimized sensor array was able to classify uropathogens from non-uropathogens in urine medium. Further, the technique was applied for discrimination of bacteria possessing resistance to different types and amounts of drugs. We found out the mechanism of sensing through optical and electrodynamic studies, which indicates the interaction between bacteria with the sensor system was mainly due to electrostatic force of interactions, but the separation of native bacteria from their drug resistant variant was due to Van der Waals forces. There are two ways bacteria can be detected, i.e., through bacterial cells and lysates. The bacterial lysates contain intracellular information and also safe to analysis as it does not contain live cells. Lysates of different drug resistant bacteria were patterned effectively from the native strain. From unknown sample analysis, we found that discrimination of bacterial cells is more sensitive than that of lysates. But the analyst can prefer bacterial lysates over live cells for safer analysis.Keywords: array-based sensing, drug resistant bacteria, linear discriminant analysis, two-dimensional MoS₂
Procedia PDF Downloads 14619670 3D-Mesh Robust Watermarking Technique for Ownership Protection and Authentication
Authors: Farhan A. Alenizi
Abstract:
Digital watermarking has evolved in the past years as an important means for data authentication and ownership protection. The images and video watermarking was well known in the field of multimedia processing; however, 3D objects' watermarking techniques have emerged as an important means for the same purposes, as 3D mesh models are in increasing use in different areas of scientific, industrial, and medical applications. Like the image watermarking techniques, 3D watermarking can take place in either space or transform domains. Unlike images and video watermarking, where the frames have regular structures in both space and temporal domains, 3D objects are represented in different ways as meshes that are basically irregular samplings of surfaces; moreover, meshes can undergo a large variety of alterations which may be hard to tackle. This makes the watermarking process more challenging. While the transform domain watermarking is preferable in images and videos, they are still difficult to implement in 3d meshes due to the huge number of vertices involved and the complicated topology and geometry, and hence the difficulty to perform the spectral decomposition, even though significant work was done in the field. Spatial domain watermarking has attracted significant attention in the past years; they can either act on the topology or on the geometry of the model. Exploiting the statistical characteristics in the 3D mesh models from both geometrical and topological aspects was useful in hiding data. However, doing that with minimal surface distortions to the mesh attracted significant research in the field. A 3D mesh blind watermarking technique is proposed in this research. The watermarking method depends on modifying the vertices' positions with respect to the center of the object. An optimal method will be developed to reduce the errors, minimizing the distortions that the 3d object may experience due to the watermarking process, and reducing the computational complexity due to the iterations and other factors. The technique relies on the displacement process of the vertices' locations depending on the modification of the variances of the vertices’ norms. Statistical analyses were performed to establish the proper distributions that best fit each mesh, and hence establishing the bins sizes. Several optimizing approaches were introduced in the realms of mesh local roughness, the statistical distributions of the norms, and the displacements in the mesh centers. To evaluate the algorithm's robustness against other common geometry and connectivity attacks, the watermarked objects were subjected to uniform noise, Laplacian smoothing, vertices quantization, simplification, and cropping. Experimental results showed that the approach is robust in terms of both perceptual and quantitative qualities. It was also robust against both geometry and connectivity attacks. Moreover, the probability of true positive detection versus the probability of false-positive detection was evaluated. To validate the accuracy of the test cases, the receiver operating characteristics (ROC) curves were drawn, and they’ve shown robustness from this aspect. 3D watermarking is still a new field but still a promising one.Keywords: watermarking, mesh objects, local roughness, Laplacian Smoothing
Procedia PDF Downloads 162