Search results for: fault detection and recovery
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5625

Search results for: fault detection and recovery

4485 Detection of Pharmaceutical Personal Protective Equipment in Video Stream

Authors: Michael Leontiev, Danil Zhilikov, Dmitry Lobanov, Lenar Klimov, Vyacheslav Chertan, Daniel Bobrov, Vladislav Maslov, Vasilii Vologdin, Ksenia Balabaeva

Abstract:

Pharmaceutical manufacturing is a complex process, where each stage requires a high level of safety and sterility. Personal Protective Equipment (PPE) is used for this purpose. Despite all the measures of control, the human factor (improper PPE wearing) causes numerous losses to human health and material property. This research proposes a solid computer vision system for ensuring safety in pharmaceutical laboratories. For this, we have tested a wide range of state-of-the-art object detection methods. Composing previously obtained results in this sphere with our own approach to this problem, we have reached a high accuracy ([email protected]) ranging from 0.77 up to 0.98 in detecting all the elements of a common set of PPE used in pharmaceutical laboratories. Our system is a step towards safe medicine production.

Keywords: sterility and safety in pharmaceutical development, personal protective equipment, computer vision, object detection, monitoring in pharmaceutical development, PPE

Procedia PDF Downloads 92
4484 Computer-Aided Classification of Liver Lesions Using Contrasting Features Difference

Authors: Hussein Alahmer, Amr Ahmed

Abstract:

Liver cancer is one of the common diseases that cause the death. Early detection is important to diagnose and reduce the incidence of death. Improvements in medical imaging and image processing techniques have significantly enhanced interpretation of medical images. Computer-Aided Diagnosis (CAD) systems based on these techniques play a vital role in the early detection of liver disease and hence reduce liver cancer death rate.  This paper presents an automated CAD system consists of three stages; firstly, automatic liver segmentation and lesion’s detection. Secondly, extracting features. Finally, classifying liver lesions into benign and malignant by using the novel contrasting feature-difference approach. Several types of intensity, texture features are extracted from both; the lesion area and its surrounding normal liver tissue. The difference between the features of both areas is then used as the new lesion descriptors. Machine learning classifiers are then trained on the new descriptors to automatically classify liver lesions into benign or malignant. The experimental results show promising improvements. Moreover, the proposed approach can overcome the problems of varying ranges of intensity and textures between patients, demographics, and imaging devices and settings.

Keywords: CAD system, difference of feature, fuzzy c means, lesion detection, liver segmentation

Procedia PDF Downloads 328
4483 Numerical Simulation of the Flow Channel in the Curved Plane Oil Skimmer

Authors: Xing Feng, Yuanbin Li

Abstract:

Oil spills at sea can cause severe marine environmental damage, including bringing huge hazards to living resources and human beings. In situ burning or chemical dispersant methods can be used to handle the oil spills sometimes, but these approaches will bring secondary pollution and fail in some situations. Oil recovery techniques have also been developed to recover oil using oil skimmer equipment installed on ships, while the hydrodynamic process of the oil flowing through the oil skimmer is very complicated and important for evaluating the recovery efficiency. Based on this, a two-dimensional numerical simulation platform for simulating the hydrodynamic process of the oil flowing through the oil skimmer is established based on the Navier-Stokes equations for viscous, incompressible fluid. Finally, the influence of the design of the flow channel in the curved plane oil skimmer on the hydrodynamic process of the oil flowing through the oil skimmer is investigated based on the established simulation platform.

Keywords: curved plane oil skimmer, flow channel, CFD, VOF

Procedia PDF Downloads 297
4482 Molecular Dynamics Simulation of the Effect of the Solid Gas Interface Nanolayer on Enhanced Thermal Conductivity of Copper-CO2 Nanofluid

Authors: Zeeshan Ahmed, Ajinkya Sarode, Pratik Basarkar, Atul Bhargav, Debjyoti Banerjee

Abstract:

The use of CO2 in oil recovery and in CO2 capture and storage is gaining traction in recent years. These applications involve heat transfer between CO2 and the base fluid, and hence, there arises a need to improve the thermal conductivity of CO2 to increase the process efficiency and reduce cost. One way to improve the thermal conductivity is through nanoparticle addition in the base fluid. The nanofluid model in this study consisted of copper (Cu) nanoparticles in varying concentrations with CO2 as a base fluid. No experimental data are available on thermal conductivity of CO2 based nanofluid. Molecular dynamics (MD) simulations are an increasingly adopted tool to perform preliminary assessments of nanoparticle (NP) fluid interactions. In this study, the effect of the formation of a nanolayer (or molecular layering) at the gas-solid interface on thermal conductivity is investigated using equilibrium MD simulations by varying NP diameter and keeping the volume fraction (1.413%) of nanofluid constant to check the diameter effect of NP on the nanolayer and thermal conductivity. A dense semi-solid fluid layer was seen to be formed at the NP-gas interface, and the thickness increases with increase in particle diameter, which also moves with the NP Brownian motion. Density distribution has been done to see the effect of nanolayer, and its thickness around the NP. These findings are extremely beneficial, especially to industries employed in oil recovery as increased thermal conductivity of CO2 will lead to enhanced oil recovery and thermal energy storage.

Keywords: copper-CO2 nanofluid, molecular dynamics simulation, molecular interfacial layer, thermal conductivity

Procedia PDF Downloads 340
4481 Detection of Safety Goggles on Humans in Industrial Environment Using Faster-Region Based on Convolutional Neural Network with Rotated Bounding Box

Authors: Ankit Kamboj, Shikha Talwar, Nilesh Powar

Abstract:

To successfully deliver our products in the market, the employees need to be in a safe environment, especially in an industrial and manufacturing environment. The consequences of delinquency in wearing safety glasses while working in industrial plants could be high risk to employees, hence the need to develop a real-time automatic detection system which detects the persons (violators) not wearing safety glasses. In this study a convolutional neural network (CNN) algorithm called faster region based CNN (Faster RCNN) with rotated bounding box has been used for detecting safety glasses on persons; the algorithm has an advantage of detecting safety glasses with different orientation angles on the persons. The proposed method of rotational bounding boxes with a convolutional neural network first detects a person from the images, and then the method detects whether the person is wearing safety glasses or not. The video data is captured at the entrance of restricted zones of the industrial environment (manufacturing plant), which is further converted into images at 2 frames per second. In the first step, the CNN with pre-trained weights on COCO dataset is used for person detection where the detections are cropped as images. Then the safety goggles are labelled on the cropped images using the image labelling tool called roLabelImg, which is used to annotate the ground truth values of rotated objects more accurately, and the annotations obtained are further modified to depict four coordinates of the rectangular bounding box. Next, the faster RCNN with rotated bounding box is used to detect safety goggles, which is then compared with traditional bounding box faster RCNN in terms of detection accuracy (average precision), which shows the effectiveness of the proposed method for detection of rotatory objects. The deep learning benchmarking is done on a Dell workstation with a 16GB Nvidia GPU.

Keywords: CNN, deep learning, faster RCNN, roLabelImg rotated bounding box, safety goggle detection

Procedia PDF Downloads 134
4480 Textual Analysis of Media Coverage on Women’s Employment during Covid-19 Recovery: Personal Choice versus Systemic Insufficiencies

Authors: Rania Al Namara

Abstract:

During the Covid-19 pandemic, women disproportionately left the workforce compared to men, and many remained outside of the labor market during the Covid-19 recovery period—a phenomenon referred to as the “she-recession” or “shecession.” While the number of women returning to work has increased, long-standing systemic inequalities interfere with women's equal participation in the workforce. Previous research on media framing has explored the importance of news coverage of women’s issues in print and magazines to shaping the public’s views on an issue and the national response. This study adopts textual analysis to examine how 50 news stories published on CNN and CBS in March 2023 frame women’s employment challenges as a matter of choice or as a matter of insufficient systems and analyzes the narratives portrayed to understand how this discourse affects national policies regarding women’s equality in the workforce. Findings suggest that media coverage centers on four themes: unequal wages at work, work-life integration, experiences of minority women, and the struggle to acquire leadership positions. Media coverage gives space to women to tell personal stories about facing these four societal challenges. However, little coverage is devoted to the political figures and institutions that either reinforce gender inequalities or advance women’s rights in these areas. These findings highlight the need for media stories that discuss policies and reforms that broaden the choices available to women in the first place.

Keywords: Covid-19 recovery, media coverage, shecession, women’s employment

Procedia PDF Downloads 98
4479 Bleaching Liquor Recovery of Batch-Wise and Continuous Method

Authors: Sidra Saleemi, Arsalan Khan, Urooj Baig, Tahir Jamil

Abstract:

In this research, it was examined that some residual amount of bleaching chemicals left in the liquor, this amount is more in Batch-wise process as compared to continuous process. These chemicals can be recovered and reused for bleaching by adding more quantity of fresh bleaching chemicals and water, this quantity will be required to balance the recipe for fabric. This liquor is recovered and samples were bleached with different modified recipe of liquor for both processes i.e. Batch-wise and continuous process. Every time good results were achieved with negligible variation in the quality parameter between the fabric bleached with fresh liquor and the fabric bleached with Recovered Liquor. Additionally, samples were dyed, and found that dyeing can be done easily on samples bleached with recover liquor.

Keywords: bleaching process, hydrogen peroxide, sodium hydroxide, liquor recovery

Procedia PDF Downloads 371
4478 Localization of Radioactive Sources with a Mobile Radiation Detection System using Profit Functions

Authors: Luís Miguel Cabeça Marques, Alberto Manuel Martinho Vale, José Pedro Miragaia Trancoso Vaz, Ana Sofia Baptista Fernandes, Rui Alexandre de Barros Coito, Tiago Miguel Prates da Costa

Abstract:

The detection and localization of hidden radioactive sources are of significant importance in countering the illicit traffic of Special Nuclear Materials and other radioactive sources and materials. Radiation portal monitors are commonly used at airports, seaports, and international land borders for inspecting cargo and vehicles. However, these equipment can be expensive and are not available at all checkpoints. Consequently, the localization of SNM and other radioactive sources often relies on handheld equipment, which can be time-consuming. The current study presents the advantages of real-time analysis of gamma-ray count rate data from a mobile radiation detection system based on simulated data and field tests. The incorporation of profit functions and decision criteria to optimize the detection system's path significantly enhances the radiation field information and reduces survey time during cargo inspection. For source position estimation, a maximum likelihood estimation algorithm is employed, and confidence intervals are derived using the Fisher information. The study also explores the impact of uncertainties, baselines, and thresholds on the performance of the profit function. The proposed detection system, utilizing a plastic scintillator with silicon photomultiplier sensors, boasts several benefits, including cost-effectiveness, high geometric efficiency, compactness, and lightweight design. This versatility allows for seamless integration into any mobile platform, be it air, land, maritime, or hybrid, and it can also serve as a handheld device. Furthermore, integration of the detection system into drones, particularly multirotors, and its affordability enable the automation of source search and substantial reduction in survey time, particularly when deploying a fleet of drones. While the primary focus is on inspecting maritime container cargo, the methodologies explored in this research can be applied to the inspection of other infrastructures, such as nuclear facilities or vehicles.

Keywords: plastic scintillators, profit functions, path planning, gamma-ray detection, source localization, mobile radiation detection system, security scenario

Procedia PDF Downloads 123
4477 Positive Effects of Aerobic Exercise after Bone Marrow Stem Cell Transplantation on Recovery of Dopaminergic Neurons and Promotion of Angiogenesis Markers in the Striatum of Parkinsonian Rats

Authors: S. A. Hashemvarzi, A. Heidarianpour, Z. Fallahmohammadi, M. Pourghasem, M. Kaviani

Abstract:

Introduction: Parkinson’s disease (PD) is a progressive neurodegenerative in the central nervous system characterized by the loss of dopaminergic neurons in the substantia nigra resulting in loss of dopamine release in the striatum. Non-drug treatment options such as Stem cell transplantation and exercise have been considered for treatment of Parkinson's disease. Purpose: The purpose of this study was to evaluate the effect of aerobic exercise after bone marrow stem cells transplantation on recovery of dopaminergic neurons and promotion of angiogenesis markers in the striatum of parkinsonian rats. Materials and Methods: 42 male Wistar rats were divided randomly into six groups: Normal (N), Sham (S), Parkinson’s (P), Stem cells transplanted Parkinson’s (SP), Exercised Parkinson’s (EP) and Stem cells transplanted + Exercised Parkinson’s (SEP). To create a model of Parkinson's, the striatum was destroyed by injection of 6-hydroxy-dopamine into the striatum through stereotaxic apparatus. Stem cells were derived from the bone marrow of femur and tibia of male rats with 6-8 weeks old. After cultivation, approximately 5×105 cells in 5 microliter of medium were injected into the striatum of rats through the channel. Aerobic exercise was included 8 weeks of running on the treadmill with a speed of 15 meters per minute. At the end, all subjects were decapitated and striatum tissues were separately isolated for measurement of vascular endothelial growth factor (VEGF), dopamine (DA) and tyrosine hydroxylase (TH) levels. Results: VEGF, DA and TH levels in the striatum of parkinsonian rats significantly increased in treatment groups (SP, EP and SEP), especially in SEP group compared to P group after treatment (P<0.05). Conclusion: The findings implicate that the BMSCs transplantation in combination with exercise would have synergistic effects leading to functional recovery, dopaminergic neurons recovery and promotion of angiogenesis marker in the striatum of parkinsonian rats.

Keywords: stem cells, treadmill training, neurotrophic factors, Parkinson

Procedia PDF Downloads 345
4476 Interdisciplinary Approach for Economic Production of Oil and Gas Reserves: Application of Geothermal Energy for Enhanced Oil Recovery

Authors: Dharmit Viroja, Prerakkumar Shah, Rajanikant Gajera, Ruchit Shah

Abstract:

With present scenario of aging oil and gas fields with high water cuts, volatile oil prices and increasing greenhouse gas emission, the need for alleviating such issues has necessitated for oil and gas industry to make the maximum out of available assets, infrastructure and reserves in mother Earth. Study undertaken emphasizes on utilizing Geothermal Energy under specific reservoir conditions for Enhanced oil Recovery (EOR) to boost up production. Allied benefits of this process include mitigation of electricity problem in remote fields and controlled CO-emission. Utilization of this energy for EOR and increasing economic life of field could surely be rewarding. A new way to value oil lands would be considered if geothermal co-production is integrated in the field development program. Temperature profile of co-produced fluid across its journey is a pivotal issue which has been studied. Geo pressured reservoirs resulting from trapped brine under an impermeable bed is also a frontier for exploitation. Hot geothermal fluid is a by-product of large number of oil and gas wells, historically this hot water has been seen as an inconvenience; however, it can be looked at as a useful resource. The production of hot fluids from abandoned and co-production of hot fluids from producing wells has potential to prolong life of oil and gas fields. The study encompasses various factors which are required for use of this technology and application of this process across various phases of oil and gas value chain. Interdisciplinary approach in oil and gas value chain has shown potential for economic production of estimated oil and gas reserves.

Keywords: enhanced oil recovery, geo-pressured reservoirs, geothermal energy, oil and gas value chain

Procedia PDF Downloads 347
4475 Exergy Analysis of Reverse Osmosis for Potable Water and Land Irrigation

Authors: M. Sarai Atab, A. Smallbone, A. P. Roskilly

Abstract:

A thermodynamic study is performed on the Reverse Osmosis (RO) desalination process for brackish water. The detailed RO model of thermodynamics properties with and without an energy recovery device was built in Simulink/MATLAB and validated against reported measurement data. The efficiency of desalination plants can be estimated by both the first and second laws of thermodynamics. While the first law focuses on the quantity of energy, the second law analysis (i.e. exergy analysis) introduces quality. This paper used the Main Outfall Drain in Iraq as a case study to conduct energy and exergy analysis of RO process. The result shows that it is feasible to use energy recovery method for reverse osmosis with salinity less than 15000 ppm as the exergy efficiency increases twice. Moreover, this analysis shows that the highest exergy destruction occurs in the rejected water and lowest occurs in the permeate flow rate accounting 37% for 4.3% respectively.

Keywords: brackish water, exergy, irrigation, reverse osmosis (RO)

Procedia PDF Downloads 178
4474 Phytoremediation of Arsenic-Contaminated Soil and Recovery of Valuable Arsenic Products

Authors: Valentine C. Eze, Adam P. Harvey

Abstract:

Contamination of groundwater and soil by heavy metals and metalloids through anthropogenic activities and natural occurrence poses serious environmental challenges globally. A possible solution to this problem is through phytoremediation of the contaminants using hyper-accumulating plants. Conventional phytoremediation treats the contaminated hyper-accumulator biomass as a waste stream which adds no value to the heavy metal(loid)s decontamination process. This study investigates strategies for remediation of soil contaminated with arsenic and the extractive chemical routes for recovery of arsenic and phosphorus from the hyper-accumulator biomass. Pteris cretica ferns species were investigated for their uptake of arsenic from soil containing 200 ± 3ppm of arsenic. The Pteris cretica ferns were shown to be capable of hyper-accumulation of arsenic, with maximum accumulations of about 4427 ± 79mg to 4875 ± 96mg of As per kg of the dry ferns. The arsenic in the Pteris cretica fronds was extracted into various solvents, with extraction efficiencies of 94.3 ± 2.1% for ethanol-water (1:1 v/v), 81.5 ± 3.2% for 1:1(v/v) methanol-water, and 70.8 ± 2.9% for water alone. The recovery efficiency of arsenic from the molybdic acid complex process 90.8 ± 5.3%. Phosphorus was also recovered from the molybdic acid complex process at 95.1 ± 4.6% efficiency. Quantitative precipitation of Mg₃(AsO₄)₂ and Mg₃(PO₄)₂ occurred in the treatment of the aqueous solutions of arsenic and phosphorus after stripping at pH of 8 – 10. The amounts of Mg₃(AsO₄)₂ and Mg₃(PO₄)₂ obtained were 96 ± 7.2% for arsenic and 94 ± 3.4% for phosphorus. The arsenic nanoparticles produced from the Mg₃(AsO₄)₂ recovered from the biomass have the average particles diameter of 45.5 ± 11.3nm. A two-stage reduction process – a first step pre-reduction of As(V) to As(III) with L-cysteine, followed by NaBH₄ reduction of the As(III) to As(0), was required to produced arsenic nanoparticles from the Mg₃(AsO₄)₂. The arsenic nanoparticles obtained are potentially valuable for medical applications, while the Mg₃(AsO₄)₂ could be used as an insecticide. The phosphorus contents of the Pteris cretica biomass was recovered as phosphomolybdic acid complex and converted to Mg₃(PO₄)₂, which could be useful in productions of fertilizer. Recovery of these valuable products from phytoremediation biomass would incentivize and drive commercial industries’ participation in remediation of contaminated lands.

Keywords: phytoremediation, Pteris cretica, hyper-accumulator, solvent extraction, molybdic acid process, arsenic nanoparticles

Procedia PDF Downloads 321
4473 Open-Source YOLO CV For Detection of Dust on Solar PV Surface

Authors: Jeewan Rai, Kinzang, Yeshi Jigme Choden

Abstract:

Accumulation of dust on solar panels impacts the overall efficiency and the amount of energy they produce. While various techniques exist for detecting dust to schedule cleaning, many of these methods use MATLAB image processing tools and other licensed software, which can be financially burdensome. This study will investigate the efficiency of a free open-source computer vision library using the YOLO algorithm. The proposed approach has been tested on images of solar panels with varying dust levels through an experiment setup. The experimental findings illustrated the effectiveness of using the YOLO-based image classification method and the overall dust detection approach with an accuracy of 90% in distinguishing between clean and dusty panels. This open-source solution provides a cost effective and accessible alternative to commercial image processing tools, offering solutions for optimizing solar panel maintenance and enhancing energy production.

Keywords: YOLO, openCV, dust detection, solar panels, computer vision, image processing

Procedia PDF Downloads 40
4472 Evaluation of Double Displacement Process via Gas Dumpflood from Multiple Gas Reservoirs

Authors: B. Rakjarit, S. Athichanagorn

Abstract:

Double displacement process is a method in which gas is injected at an updip well to displace the oil bypassed by waterflooding operation from downdip water injector. As gas injection is costly and a large amount of gas is needed, gas dump-flood from multiple gas reservoirs is an attractive alternative. The objective of this paper is to demonstrate the benefits of the novel approach of double displacement process via gas dump-flood from multiple gas reservoirs. A reservoir simulation model consisting of a dipping oil reservoir and several underlying layered gas reservoirs was constructed in order to investigate the performance of the proposed method. Initially, water was injected via the downdip well to displace oil towards the producer located updip. When the water cut at the producer became high, the updip well was shut in and perforated in the gas zones in order to dump gas into the oil reservoir. At this point, the downdip well was open for production. In order to optimize oil recovery, oil production and water injection rates and perforation strategy on the gas reservoirs were investigated for different numbers of gas reservoirs having various depths and thicknesses. Gas dump-flood from multiple gas reservoirs can help increase the oil recovery after implementation of waterflooding upto 10%. Although the amount of additional oil recovery is slightly lower than the one obtained in conventional double displacement process, the proposed process requires a small completion cost of the gas zones and no operating cost while the conventional method incurs high capital investment in gas compression facility and high-pressure gas pipeline and additional operating cost. From the simulation study, oil recovery can be optimized by producing oil at a suitable rate and perforating the gas zones with the right strategy which depends on depths, thicknesses and number of the gas reservoirs. Conventional double displacement process has been studied and successfully implemented in many fields around the world. However, the method of dumping gas into the oil reservoir instead of injecting it from surface during the second displacement process has never been studied. The study of this novel approach will help a practicing engineer to understand the benefits of such method and can implement it with minimum cost.

Keywords: gas dump-flood, multi-gas layers, double displacement process, reservoir simulation

Procedia PDF Downloads 411
4471 TiO₂ Nanotube Array Based Selective Vapor Sensors for Breath Analysis

Authors: Arnab Hazra

Abstract:

Breath analysis is a quick, noninvasive and inexpensive technique for disease diagnosis can be used on people of all ages without any risk. Only a limited number of volatile organic compounds (VOCs) can be associated with the occurrence of specific diseases. These VOCs can be considered as disease markers or breath markers. Selective detection with specific concentration of breath marker in exhaled human breath is required to detect a particular disease. For example, acetone (C₃H₆O), ethanol (C₂H₅OH), ethane (C₂H₆) etc. are the breath markers and abnormal concentrations of these VOCs in exhaled human breath indicates the diseases like diabetes mellitus, renal failure, breast cancer respectively. Nanomaterial-based vapor sensors are inexpensive, small and potential candidate for the detection of breath markers. In practical measurement, selectivity is the most crucial issue where trace detection of breath marker is needed to identify accurately in the presence of several interfering vapors and gases. Current article concerns a novel technique for selective and lower ppb level detection of breath markers at very low temperature based on TiO₂ nanotube array based vapor sensor devices. Highly ordered and oriented TiO₂ nanotube array was synthesized by electrochemical anodization of high purity tatinium (Ti) foil. 0.5 wt% NH₄F, ethylene glycol and 10 vol% H₂O was used as the electrolyte and anodization was carried out for 90 min with 40 V DC potential. Au/TiO₂ Nanotube/Ti, sandwich type sensor device was fabricated for the selective detection of VOCs in low concentration range. Initially, sensor was characterized where resistive and capacitive change of the sensor was recorded within the valid concentration range for individual breath markers (or organic vapors). Sensor resistance was decreased and sensor capacitance was increased with the increase of vapor concentration. Now, the ratio of resistive slope (mR) and capacitive slope (mC) provided a concentration independent constant term (M) for a particular vapor. For the detection of unknown vapor, ratio of resistive change and capacitive change at any concentration was same to the previously calculated constant term (M). After successful identification of the target vapor, concentration was calculated from the straight line behavior of resistance as a function of concentration. Current technique is suitable for the detection of particular vapor from a mixture of other interfering vapors.

Keywords: breath marker, vapor sensors, selective detection, TiO₂ nanotube array

Procedia PDF Downloads 159
4470 Design of an Ensemble Learning Behavior Anomaly Detection Framework

Authors: Abdoulaye Diop, Nahid Emad, Thierry Winter, Mohamed Hilia

Abstract:

Data assets protection is a crucial issue in the cybersecurity field. Companies use logical access control tools to vault their information assets and protect them against external threats, but they lack solutions to counter insider threats. Nowadays, insider threats are the most significant concern of security analysts. They are mainly individuals with legitimate access to companies information systems, which use their rights with malicious intents. In several fields, behavior anomaly detection is the method used by cyber specialists to counter the threats of user malicious activities effectively. In this paper, we present the step toward the construction of a user and entity behavior analysis framework by proposing a behavior anomaly detection model. This model combines machine learning classification techniques and graph-based methods, relying on linear algebra and parallel computing techniques. We show the utility of an ensemble learning approach in this context. We present some detection methods tests results on an representative access control dataset. The use of some explored classifiers gives results up to 99% of accuracy.

Keywords: cybersecurity, data protection, access control, insider threat, user behavior analysis, ensemble learning, high performance computing

Procedia PDF Downloads 132
4469 Optimization of Hate Speech and Abusive Language Detection on Indonesian-language Twitter using Genetic Algorithms

Authors: Rikson Gultom

Abstract:

Hate Speech and Abusive language on social media is difficult to detect, usually, it is detected after it becomes viral in cyberspace, of course, it is too late for prevention. An early detection system that has a fairly good accuracy is needed so that it can reduce conflicts that occur in society caused by postings on social media that attack individuals, groups, and governments in Indonesia. The purpose of this study is to find an early detection model on Twitter social media using machine learning that has high accuracy from several machine learning methods studied. In this study, the support vector machine (SVM), Naïve Bayes (NB), and Random Forest Decision Tree (RFDT) methods were compared with the Support Vector machine with genetic algorithm (SVM-GA), Nave Bayes with genetic algorithm (NB-GA), and Random Forest Decision Tree with Genetic Algorithm (RFDT-GA). The study produced a comparison table for the accuracy of the hate speech and abusive language detection model, and presented it in the form of a graph of the accuracy of the six algorithms developed based on the Indonesian-language Twitter dataset, and concluded the best model with the highest accuracy.

Keywords: abusive language, hate speech, machine learning, optimization, social media

Procedia PDF Downloads 131
4468 The Development of the First Inter-Agency Residential Rehabilitation Service for Gambling Disorder with Complex Clinical Needs

Authors: Dragos Dragomir-Stanciu, Leon Marsh

Abstract:

Background As a response to the gaps identified in recent research in the provision of residential care to address co-occurring health needs, including mental health problems and complexities Gamble Aware has facilitated the possibility to provide a new service which would extend the NGTS provision of residential rehabilitation for gambling disorder with complex and co-morbid presentation. Gordon Moody, together with Adferiad have been successful in securing the tender for this service and this presentation aims to introduce FOLD, the resulting model of treatment developed for the delivery of the service. Setting As a partnership, we have come together to coproduce a model which allows us to share our clinical and industry knowledge and build on our reputations as trusted treatment providers. The presentation will outline our expertise share in development of a unified approach to recovery-oriented models of care, clinical governance, risk assessment and management and aftercare and continuous recovery. We will also introduce our innovative specialist referral portal which will offer referring partners the ability to include the service user in planning their own recovery journey. Outcomes Our collaboration has resulted in the development of the FOLD model which includes three agile and flexible treatment packages aimed at offering the most enhanced and comprehensive treatment in UK, to date, for those most affected by gambling harm. The paper will offer insight into each treatment package and all recovery model stages involved, as well as into the partnership work with NGST providers, local mental health and social care providers and lived experience organisation that will enable us to offer support to more 100 people a year who would otherwise get “lost in the system”. Conclusion FOLD offers a great opportunity to develop, implement and evaluate a new, much needed, whole-person and whole-system approach to counter gambling related harms.

Keywords: gambling treatment, partnership working, integrated care pathways, NGTS, complex needs

Procedia PDF Downloads 137
4467 The Development of the Geological Structure of the Bengkulu Fore Arc Basin, Western Edge of Sundaland, Sumatra, and Its Relationship to Hydrocarbon Trapping Mechanism

Authors: Lauti Dwita Santy, Hermes Panggabean, Syahrir Andi Mangga

Abstract:

The Bengkulu Basin is part of the Sunda Arc system, which is a classic convergent type margin that occur around the southern rim of the Eurasian continental (Sundaland) plate. The basin is located between deep sea trench (Mentawai Outer Arc high) and the volvanic/ magmatic Arc of the Barisan Mountains Range. To the northwest it is bounded by Padang High, to the northest by Barisan Mountains (Sumatra Fault Zone) to the southwest by Mentawai Fault Zone and to the southeast by Semangko High/ Sunda Strait. The stratigraphic succession and tectonic development can be broadly divided into four stage/ periods, i.e Late Jurassic- Early Cretaceous, Late Eocene-Early Oligocene, Late Oligocene-Early Miocene, Middle Miocene-Late Miocene and Pliocene-Plistocene, which are mainly controlled by the development of subduction activities. The Pre Tertiary Basement consist of sedimentary and shallow water limestone, calcareous mudstone, cherts and tholeiitic volcanic rocks, with Late Jurassic to Early Cretaceous in age. The sedimentation in this basin is depend on the relief of the Pre Tertiary Basement (Woyla Terrane) and occured into two stages, i.e. transgressive stage during the Latest Oligocene-Early Middle Miocene Seblat Formation, and the regressive stage during the Latest Middle Miocene-Pleistocene (Lemau, Simpangaur and Bintunan Formations). The Pre-Tertiary Faults were more intensive than the overlying cover, The Tertiary Rocks. There are two main fault trends can be distinguished, Northwest–Southwest Faults and Northeast-Southwest Faults. The NW-SE fault (Ketaun) are commonly laterally persistent, are interpreted to the part of Sumatran Fault Systems. They commonly form the boundaries to the Pre Tertiary basement highs and therefore are one of the faults elements controlling the geometry and development of the Tertiary sedimentary basins.The Northeast-Southwest faults was formed a conjugate set to the Northwest–Southeast Faults. In the earliest Tertiary and reactivated during the Plio-Pleistocene in a compressive mode with subsequent dextral displacement. The Block Faulting accross these two sets of faults related to approximate North–South compression in Paleogene time and produced a series of elongate basins separated by basement highs in the backarc and forearc region. The Bengkulu basin is interpreted having evolved from pull apart feature in the area southwest of the main Sumatra Fault System related to NW-SE trending in dextral shear.Based on Pyrolysis Yield (PY) vs Total Organic Carbon (TOC) diagram show that Seblat and Lemau Formation belongs to oil and Gas Prone with the quality of the source rocks includes into excellent and good (Lemau Formation), Fair and Poor (Seblat Formation). The fine-grained carbonaceous sediment of the Seblat dan Lemau Formations as source rocks, the coarse grained and carbonate sediments of the Seblat and Lemau Formations as reservoir rocks, claystone bed in Seblat and Lemau Formation as caprock. The source rocks maturation are late immature to early mature, with kerogen type II and III (Seblat Formation), and late immature to post mature with kerogen type I and III (Lemau Formation). The burial history show to 2500 m in depthh with paleo temperature reached 80oC. Trapping mechanism occur during Oligo–Miocene and Middle Miocene, mainly in block faulting system.

Keywords: fore arc, bengkulu, sumatra, sundaland, hydrocarbon, trapping mechanism

Procedia PDF Downloads 560
4466 The Journey of a Malicious HTTP Request

Authors: M. Mansouri, P. Jaklitsch, E. Teiniker

Abstract:

SQL injection on web applications is a very popular kind of attack. There are mechanisms such as intrusion detection systems in order to detect this attack. These strategies often rely on techniques implemented at high layers of the application but do not consider the low level of system calls. The problem of only considering the high level perspective is that an attacker can circumvent the detection tools using certain techniques such as URL encoding. One technique currently used for detecting low-level attacks on privileged processes is the tracing of system calls. System calls act as a single gate to the Operating System (OS) kernel; they allow catching the critical data at an appropriate level of detail. Our basic assumption is that any type of application, be it a system service, utility program or Web application, “speaks” the language of system calls when having a conversation with the OS kernel. At this level we can see the actual attack while it is happening. We conduct an experiment in order to demonstrate the suitability of system call analysis for detecting SQL injection. We are able to detect the attack. Therefore we conclude that system calls are not only powerful in detecting low-level attacks but that they also enable us to detect high-level attacks such as SQL injection.

Keywords: Linux system calls, web attack detection, interception, SQL

Procedia PDF Downloads 360
4465 An Alternative Institutional Design for Efficient Management of Nepalese Irrigation Systems

Authors: Tirtha Raj Dhakal, Brian Davidson, Bob Farquharson

Abstract:

Institutional design is important if water resources are to be managed efficiently. In Nepal, the supply of water in both farmer- and agency-managed irrigation systems is inefficient because of the weak institutional frameworks. This type of inefficiency is linked with collective problems such as non-excludability of irrigation water, inadequate recognition of property rights and externalities. Irrigation scheme surveys from Nepal as well as existing literature revealed that the Nepalese irrigation sector is facing many issues such as low cost recovery, inadequate maintenance of the schemes and inefficient allocation and utilization of irrigation water. The institutional practices currently in place also fail to create/force any incentives for farmers to use water efficiently and to pay for its use. This, thus, compels the need of refined institutional framework that can address the collective problems and improve irrigation efficiency.

Keywords: agency-managed, cost recovery, farmer-managed, institutional design

Procedia PDF Downloads 430
4464 Power Recovery in Egyptian Natural Gas Pressure Reduction Stations Using Turboexpander Systems

Authors: Kamel A. Elshorbagy, Mohamed A. Hussein, Rola S. Afify

Abstract:

Natural gas pressure reduction is typically achieved using pressure reducing valves, where isenthalpic expansion takes place with considerable amount of wasted energy in an irreversible throttling process of the gas. Replacing gas-throttling process by an expansion process in a turbo expander (TE) converts the pressure of natural gas into mechanical energy transmitted to a loading device (i.e. an electric generator). This paper investigates the performance of a turboexpander system for power recovery at natural gas pressure reduction stations. There is a considerable temperature drop associated with the turboexpander process. Essential preheating is required, using gas fired boilers, to avoid undesirable effects of a low outlet temperature. Various system configurations were simulated by the general flow sheet simulator HYSYS and factors affecting the overall performance of the systems were investigated. Power outputs and fuel requirements were found using typical gas flow variation data. The simulation was performed for two case studies in which real input data are used. These case studies involve a domestic (commercial) and an industrial natural gas pressure reduction stations in Egypt. Economic studies of using the turboexpander system in both of the two natural gas pressure reduction stations are conducted using precise data obtained through communication with several companies working in this field. The results of economic analysis, for the two case studies, prove that using turboexpander systems in Egyptian natural gas reduction stations can be a successful project for energy conservation.

Keywords: natural gas, power recovery, reduction stations, turboexpander systems

Procedia PDF Downloads 330
4463 Credit Card Fraud Detection with Ensemble Model: A Meta-Heuristic Approach

Authors: Gong Zhilin, Jing Yang, Jian Yin

Abstract:

The purpose of this paper is to develop a novel system for credit card fraud detection based on sequential modeling of data using hybrid deep learning models. The projected model encapsulates five major phases are pre-processing, imbalance-data handling, feature extraction, optimal feature selection, and fraud detection with an ensemble classifier. The collected raw data (input) is pre-processed to enhance the quality of the data through alleviation of the missing data, noisy data as well as null values. The pre-processed data are class imbalanced in nature, and therefore they are handled effectively with the K-means clustering-based SMOTE model. From the balanced class data, the most relevant features like improved Principal Component Analysis (PCA), statistical features (mean, median, standard deviation) and higher-order statistical features (skewness and kurtosis). Among the extracted features, the most optimal features are selected with the Self-improved Arithmetic Optimization Algorithm (SI-AOA). This SI-AOA model is the conceptual improvement of the standard Arithmetic Optimization Algorithm. The deep learning models like Long Short-Term Memory (LSTM), Convolutional Neural Network (CNN), and optimized Quantum Deep Neural Network (QDNN). The LSTM and CNN are trained with the extracted optimal features. The outcomes from LSTM and CNN will enter as input to optimized QDNN that provides the final detection outcome. Since the QDNN is the ultimate detector, its weight function is fine-tuned with the Self-improved Arithmetic Optimization Algorithm (SI-AOA).

Keywords: credit card, data mining, fraud detection, money transactions

Procedia PDF Downloads 136
4462 Enabling Oral Communication and Accelerating Recovery: The Creation of a Novel Low-Cost Electroencephalography-Based Brain-Computer Interface for the Differently Abled

Authors: Rishabh Ambavanekar

Abstract:

Expressive Aphasia (EA) is an oral disability, common among stroke victims, in which the Broca’s area of the brain is damaged, interfering with verbal communication abilities. EA currently has no technological solutions and its only current viable solutions are inefficient or only available to the affluent. This prompts the need for an affordable, innovative solution to facilitate recovery and assist in speech generation. This project proposes a novel concept: using a wearable low-cost electroencephalography (EEG) device-based brain-computer interface (BCI) to translate a user’s inner dialogue into words. A low-cost EEG device was developed and found to be 10 to 100 times less expensive than any current EEG device on the market. As part of the BCI, a machine learning (ML) model was developed and trained using the EEG data. Two stages of testing were conducted to analyze the effectiveness of the device: a proof-of-concept and a final solution test. The proof-of-concept test demonstrated an average accuracy of above 90% and the final solution test demonstrated an average accuracy of above 75%. These two successful tests were used as a basis to demonstrate the viability of BCI research in developing lower-cost verbal communication devices. Additionally, the device proved to not only enable users to verbally communicate but has the potential to also assist in accelerated recovery from the disorder.

Keywords: neurotechnology, brain-computer interface, neuroscience, human-machine interface, BCI, HMI, aphasia, verbal disability, stroke, low-cost, machine learning, ML, image recognition, EEG, signal analysis

Procedia PDF Downloads 122
4461 Capability of Marine Macroalgae Chaetomorpha linum for Wastewater Phytoremediation and Biofuel Recovery

Authors: Zhipeng Chen, Lingfeng Wang, Shuang Qiu, Shijian Ge

Abstract:

Macroalgae are larger in size compared with microalgae; hence, they imposed lower separation and drying costs. To explore the potential for enhancing cultivation conditions in macroalgae Chaetomorpha linum (C. linum)-based bioreactor for nutrient recovery from municipal wastewaters and examine the biochemical composition of the macroalgae for the potential downstream production of biofuels, screening experiments were performed. This study suggested that C. linum grew well on primary (PW), secondary (SW), and centrate wastewater (CW). A step feeding approach was shown to significantly enhance biomass productivity when grown on 10% CW; meanwhile, nitrogen and phosphorus removal efficiencies increased to 86.8 ± 1.1% and 92.6 ± 0.2%, respectively. The CO₂-supplemented SW cultures were 1.20 times more productive than the corresponding controls without CO₂ supplementation. These findings demonstrate that C. linum could represent a promising and efficient wastewater treatment alternative which could also provide a feedstock for downstream processing to biofuels.

Keywords: biofuel production, macroalgae, nutrient removal, wastewater

Procedia PDF Downloads 168
4460 Silicon-Photonic-Sensor System for Botulinum Toxin Detection in Water

Authors: Binh T. T. Nguyen, Zhenyu Li, Eric Yap, Yi Zhang, Ai-Qun Liu

Abstract:

Silicon-photonic-sensor system is an emerging class of analytical technologies that use evanescent field wave to sensitively measure the slight difference in the surrounding environment. The wavelength shift induced by local refractive index change is used as an indicator in the system. These devices can be served as sensors for a wide variety of chemical or biomolecular detection in clinical and environmental fields. In our study, a system including a silicon-based micro-ring resonator, microfluidic channel, and optical processing is designed, fabricated for biomolecule detection. The system is demonstrated to detect Clostridium botulinum type A neurotoxin (BoNT) in different water sources. BoNT is one of the most toxic substances known and relatively easily obtained from a cultured bacteria source. The toxin is extremely lethal with LD50 of about 0.1µg/70kg intravenously, 1µg/ 70 kg by inhalation, and 70µg/kg orally. These factors make botulinum neurotoxins primary candidates as bioterrorism or biothreat agents. It is required to have a sensing system which can detect BoNT in a short time, high sensitive and automatic. For BoNT detection, silicon-based micro-ring resonator is modified with a linker for the immobilization of the anti-botulinum capture antibody. The enzymatic reaction is employed to increase the signal hence gains sensitivity. As a result, a detection limit to 30 pg/mL is achieved by our silicon-photonic sensor within a short period of 80 min. The sensor also shows high specificity versus the other type of botulinum. In the future, by designing the multifunctional waveguide array with fully automatic control system, it is simple to simultaneously detect multi-biomaterials at a low concentration within a short period. The system has a great potential to apply for online, real-time and high sensitivity for the label-free bimolecular rapid detection.

Keywords: biotoxin, photonic, ring resonator, sensor

Procedia PDF Downloads 119
4459 Diversity Indices as a Tool for Evaluating Quality of Water Ways

Authors: Khadra Ahmed, Khaled Kheireldin

Abstract:

In this paper, we present a pedestrian detection descriptor called Fused Structure and Texture (FST) features based on the combination of the local phase information with the texture features. Since the phase of the signal conveys more structural information than the magnitude, the phase congruency concept is used to capture the structural features. On the other hand, the Center-Symmetric Local Binary Pattern (CSLBP) approach is used to capture the texture information of the image. The dimension less quantity of the phase congruency and the robustness of the CSLBP operator on the flat images, as well as the blur and illumination changes, lead the proposed descriptor to be more robust and less sensitive to the light variations. The proposed descriptor can be formed by extracting the phase congruency and the CSLBP values of each pixel of the image with respect to its neighborhood. The histogram of the oriented phase and the histogram of the CSLBP values for the local regions in the image are computed and concatenated to construct the FST descriptor. Several experiments were conducted on INRIA and the low resolution DaimlerChrysler datasets to evaluate the detection performance of the pedestrian detection system that is based on the FST descriptor. A linear Support Vector Machine (SVM) is used to train the pedestrian classifier. These experiments showed that the proposed FST descriptor has better detection performance over a set of state of the art feature extraction methodologies.

Keywords: planktons, diversity indices, water quality index, water ways

Procedia PDF Downloads 520
4458 Sensor Registration in Multi-Static Sonar Fusion Detection

Authors: Longxiang Guo, Haoyan Hao, Xueli Sheng, Hanjun Yu, Jingwei Yin

Abstract:

In order to prevent target splitting and ensure the accuracy of fusion, system error registration is an important step in multi-static sonar fusion detection system. To eliminate the inherent system errors including distance error and angle error of each sonar in detection, this paper uses offline estimation method for error registration. Suppose several sonars from different platforms work together to detect a target. The target position detected by each sonar is based on each sonar’s own reference coordinate system. Based on the two-dimensional stereo projection method, this paper uses real-time quality control (RTQC) method and least squares (LS) method to estimate sensor biases. The RTQC method takes the average value of each sonar’s data as the observation value and the LS method makes the least square processing of each sonar’s data to get the observation value. In the underwater acoustic environment, matlab simulation is carried out and the simulation results show that both algorithms can estimate the distance and angle error of sonar system. The performance of the two algorithms is also compared through the root mean square error and the influence of measurement noise on registration accuracy is explored by simulation. The system error convergence of RTQC method is rapid, but the distribution of targets has a serious impact on its performance. LS method can not be affected by target distribution, but the increase of random noise will slow down the convergence rate. LS method is an improvement of RTQC method, which is widely used in two-dimensional registration. The improved method can be used for underwater multi-target detection registration.

Keywords: data fusion, multi-static sonar detection, offline estimation, sensor registration problem

Procedia PDF Downloads 174
4457 Vehicular Speed Detection Camera System Using Video Stream

Authors: C. A. Anser Pasha

Abstract:

In this paper, a new Vehicular Speed Detection Camera System that is applicable as an alternative to traditional radars with the same accuracy or even better is presented. The real-time measurement and analysis of various traffic parameters such as speed and number of vehicles are increasingly required in traffic control and management. Image processing techniques are now considered as an attractive and flexible method for automatic analysis and data collections in traffic engineering. Various algorithms based on image processing techniques have been applied to detect multiple vehicles and track them. The SDCS processes can be divided into three successive phases; the first phase is Objects detection phase, which uses a hybrid algorithm based on combining an adaptive background subtraction technique with a three-frame differencing algorithm which ratifies the major drawback of using only adaptive background subtraction. The second phase is Objects tracking, which consists of three successive operations - object segmentation, object labeling, and object center extraction. Objects tracking operation takes into consideration the different possible scenarios of the moving object like simple tracking, the object has left the scene, the object has entered the scene, object crossed by another object, and object leaves and another one enters the scene. The third phase is speed calculation phase, which is calculated from the number of frames consumed by the object to pass by the scene.

Keywords: radar, image processing, detection, tracking, segmentation

Procedia PDF Downloads 471
4456 Gaussian Probability Density for Forest Fire Detection Using Satellite Imagery

Authors: S. Benkraouda, Z. Djelloul-Khedda, B. Yagoubi

Abstract:

we present a method for early detection of forest fires from a thermal infrared satellite image, using the image matrix of the probability of belonging. The principle of the method is to compare a theoretical mathematical model to an experimental model. We considered that each line of the image matrix, as an embodiment of a non-stationary random process. Since the distribution of pixels in the satellite image is statistically dependent, we divided these lines into small stationary and ergodic intervals to characterize the image by an adequate mathematical model. A standard deviation was chosen to generate random variables, so each interval behaves naturally like white Gaussian noise. The latter has been selected as the mathematical model that represents a set of very majority pixels, which we can be considered as the image background. Before modeling the image, we made a few pretreatments, then the parameters of the theoretical Gaussian model were extracted from the modeled image, these settings will be used to calculate the probability of each interval of the modeled image to belong to the theoretical Gaussian model. The high intensities pixels are regarded as foreign elements to it, so they will have a low probability, and the pixels that belong to the background image will have a high probability. Finally, we did present the reverse of the matrix of probabilities of these intervals for a better fire detection.

Keywords: forest fire, forest fire detection, satellite image, normal distribution, theoretical gaussian model, thermal infrared matrix image

Procedia PDF Downloads 148